pith. machine review for the scientific record. sign in

arxiv: 2604.08810 · v1 · submitted 2026-04-09 · 💻 cs.CV · cs.LG

Recognition: unknown

R2G: A Multi-View Circuit Graph Benchmark Suite from RTL to GDSII

Zewei Zhou , Jiajun Zou , Jiajia Zhang , Ao Yang , Ruichao He , Haozheng Zhou , Ao Liu , Jiawei Liu , Leilei Jin , Shan Shen , Daying Sun

Authors on Pith no claims yet

Pith reviewed 2026-05-10 16:57 UTC · model grok-4.3

classification 💻 cs.CV cs.LG
keywords circuit graphsgraph neural networksphysical designbenchmark suiteRTL to GDSIImulti-view representationsEDA automation
0
0 comments X

The pith

View choice in circuit graph representations affects GNN accuracy more than the choice of GNN model.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper introduces R2G, a benchmark suite that converts the same 30 open-source circuit designs into five different graph views spanning synthesis, placement, and routing. Each view carries identical attribute information but attaches those features to nodes, edges, or subgraphs in stage-specific ways. Experiments with three GNN architectures show that swapping the graph view changes test R² by more than 0.3 for any fixed model, while swapping the model inside one view produces smaller gains. Node-centric views turn out to generalize best from placement to routing tasks, and a three- or four-layer decoder head is enough to drive predictions to R² above 0.99. The suite supplies loaders, splits, and metrics so future work can test new models under controlled representation conditions.

Core claim

R2G supplies five stage-aware circuit-graph views that encode the same attributes yet differ only in feature attachment points. Systematic evaluation on GINE, GAT, and ResGatedGCN models reveals that representation choice produces larger accuracy swings than architecture choice, node-centric attachments generalize across placement and routing, and decoder depth of three to four layers converts divergent runs into near-perfect predictions with R² exceeding 0.99.

What carries the argument

Five stage-aware graph views with information parity, each attaching the same attribute set to different structural elements of the circuit netlist.

If this is right

  • Selecting the graph view alone can raise prediction R² by more than 0.3 without any change to the underlying GNN.
  • Node-centric views provide the strongest cross-stage generalization for both placement and routing tasks.
  • Increasing decoder-head depth to three or four layers is sufficient to reach R² greater than 0.99 on these graphs.
  • Future physical-design models can be compared fairly only after fixing the graph view, because representation effects otherwise confound architecture comparisons.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • EDA tool flows could prioritize graph-construction pipelines over repeated architecture search, because representation gains appear larger.
  • The same multi-view discipline might apply to other netlist-derived tasks such as timing or power estimation where stage-specific feature attachment has not yet been isolated.
  • If new open-source designs are added with different technology nodes, the benchmark can test whether the observed dominance of node-centric views persists beyond the current 30 cores.

Load-bearing premise

The five views truly preserve identical information content and the synthesis-to-routing pipeline introduces no hidden biases from tool choices or IP core selection.

What would settle it

A controlled experiment in which the same GNN is retrained on each view while forcing identical node and edge feature sets; if the R² gap across views shrinks below 0.1, the claim that view choice dominates would be falsified.

Figures

Figures reproduced from arXiv: 2604.08810 by Ao Liu, Ao Yang, Daying Sun, Haozheng Zhou, Jiajia Zhang, Jiajun Zou, Jiawei Liu, Leilei Jin, Ruichao He, Shan Shen, Zewei Zhou.

Figure 1
Figure 1. Figure 1: Benchmark and dataset evolution across graph ML and [PITH_FULL_IMAGE:figures/full_fig_p001_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: OpenROAD RTL-to-GDSII post-end flow including [PITH_FULL_IMAGE:figures/full_fig_p003_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Original circuit schematic (a) and five complementary circuit graph views (b–f) used in R2G: (b) all-elements-as-nodes, (c) [PITH_FULL_IMAGE:figures/full_fig_p004_3.png] view at source ↗
read the original abstract

Graph neural networks (GNNs) are increasingly applied to physical design tasks such as congestion prediction and wirelength estimation, yet progress is hindered by inconsistent circuit representations and the absence of controlled evaluation protocols. We present R2G (RTL-to-GDSII), a multi-view circuit-graph benchmark suite that standardizes five stage-aware views with information parity (every view encodes the same attribute set, differing only in where features attach) over 30 open-source IP cores (up to $10^6$ nodes/edges). R2G provides an end-to-end DEF-to-graph pipeline spanning synthesis, placement, and routing stages, together with loaders, unified splits, domain metrics, and reproducible baselines. By decoupling representation choice from model choice, R2G isolates a confound that prior EDA and graph-ML benchmarks leave uncontrolled. In systematic studies with GINE, GAT, and ResGatedGCN, we find: (i) view choice dominates model choice, with Test R$^2$ varying by more than 0.3 across representations for a fixed GNN; (ii) node-centric views generalize best across both placement and routing; and (iii) decoder-head depth (3--4 layers) is the primary accuracy driver, turning divergent training into near-perfect predictions (R$^2$$>$0.99). Code and datasets are available at https://github.com/ShenShan123/R2G.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

1 major / 3 minor

Summary. The paper introduces the R2G benchmark suite, which standardizes five stage-aware circuit-graph views (with claimed information parity) derived from 30 open-source IP cores via an end-to-end DEF-to-graph pipeline spanning synthesis, placement, and routing. It supplies loaders, unified splits, domain-specific metrics, and reproducible baselines using GINE, GAT, and ResGatedGCN. Experiments show that view choice dominates model choice (Test R² varies by >0.3 for fixed GNN), node-centric views generalize best across stages, and decoder-head depth (3-4 layers) drives near-perfect predictions (R² > 0.99).

Significance. If the central claims hold, R2G supplies a much-needed controlled benchmark for GNNs in physical design, isolating representation effects that prior EDA and graph-ML work leaves confounded. The open release of data, code, and baselines, together with systematic multi-architecture comparisons and concrete R² metrics, are concrete strengths that lower the barrier for reproducible research on congestion, wirelength, and related tasks.

major comments (1)
  1. [Abstract and §3] Abstract and §3 (Views and Pipeline): The central claim that performance gaps are attributable to representation (rather than unequal attribute content) rests on 'information parity' (every view encodes the same attribute set, differing only in attachment point). The manuscript asserts this via the shared DEF-to-graph pipeline but provides no explicit verification—e.g., no table or statistics confirming that gate types, netlist connectivity, timing, and congestion signals are identically present across all five views. Without this check, the reported R² differences (>0.3) could reflect information loss in some views rather than the intended confound.
minor comments (3)
  1. [§5] §5 (Experiments): The description of unified splits and any data exclusion criteria for the 30 IP cores is brief; adding explicit details on train/validation/test construction and size-based filtering would improve defensibility of the generalization claims.
  2. [Figure 3] Figure 3 (or equivalent results table): The R² values are reported to two decimals; showing standard deviation across seeds or runs would better convey stability of the 'near-perfect' (R² > 0.99) regime.
  3. [§3] Notation: The distinction between 'node-centric' and other views is clear in text but could be reinforced with a small summary table listing attachment points and feature dimensions for each view.

Simulated Author's Rebuttal

1 responses · 0 unresolved

We thank the referee for the positive evaluation and the constructive comment on the information-parity claim. We address the concern point by point below.

read point-by-point responses
  1. Referee: [Abstract and §3] Abstract and §3 (Views and Pipeline): The central claim that performance gaps are attributable to representation (rather than unequal attribute content) rests on 'information parity' (every view encodes the same attribute set, differing only in attachment point). The manuscript asserts this via the shared DEF-to-graph pipeline but provides no explicit verification—e.g., no table or statistics confirming that gate types, netlist connectivity, timing, and congestion signals are identically present across all five views. Without this check, the reported R² differences (>0.3) could reflect information loss in some views rather than the intended confound.

    Authors: We agree that an explicit verification would strengthen the presentation. The information parity follows directly from the shared DEF-to-graph pipeline described in §3: all five views are generated from the identical DEF file and extract the same attribute set (gate types, netlist connectivity, timing arcs, congestion signals, etc.), differing solely in attachment point (node-centric vs. edge-centric) and stage-specific structural encoding. To address the referee’s request, we will insert a new table in the revised §3 that enumerates the attribute schema for each view together with summary statistics (e.g., attribute counts and value distributions) confirming equivalence. This addition will be purely expository and will not alter any experimental results or conclusions. revision: yes

Circularity Check

0 steps flagged

No circularity: benchmark supplies independent data, loaders, and empirical baselines

full rationale

The paper introduces a new multi-view circuit-graph dataset and pipeline over 30 IP cores, then reports direct empirical measurements (Test R² differences across views and models, decoder depth effects) on that data. No derivation, equation, or central claim reduces by construction to a parameter fitted inside the paper or to a self-citation chain; the information-parity assertion is a design statement about the DEF-to-graph conversion rather than a fitted or renamed result. The reported findings are falsifiable measurements on the released splits and therefore self-contained against external benchmarks.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

The work rests on standard assumptions about graph construction from EDA netlists and the applicability of message-passing GNNs; no new entities or heavily fitted parameters are introduced as load-bearing elements.

axioms (1)
  • domain assumption GNN message passing can capture relevant circuit properties for prediction tasks when graphs are constructed from DEF files
    Invoked when applying GINE, GAT, and ResGatedGCN to the generated views for congestion and wirelength tasks.

pith-pipeline@v0.9.0 · 5593 in / 1244 out tokens · 81628 ms · 2026-05-10T16:57:45.604901+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

37 extracted references · 10 canonical work pages · 1 internal anchor

  1. [1]

    readthedocs

    Openroad documentation.https://openroad. readthedocs . io / en / latest / main / README.html, . Accessed: 2025-11-06. 1

  2. [2]

    https : / / github

    Openroad: Unified application for rtl-to-gdsii flow. https : / / github . com / The - OpenROAD - Project/OpenROAD, . Accessed: 2025-11-06. 1, 3

  3. [3]

    GNN-RE: Graph neural networks for reverse engineering of gate- level netlists.IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, 41(8): 2435–2448, 2022

    Lilas Alrahis, Abhrajit Sengupta, Johann Knechtel, Satwik Patnaik, Hani Saleh, Baker Mohammad, Mah- moud Al-Qutayri, and Ozgur Sinanoglu. GNN-RE: Graph neural networks for reverse engineering of gate- level netlists.IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, 41(8): 2435–2448, 2022. 1, 2

  4. [4]

    Sdplib 1.2, a library of semidefinite program- ming test problems.Optimization Methods and Software, 11(1-4):683–690, 1999

    Xavier Bresson and Thomas Laurent. Residual gated graph convnets.arXiv preprint arXiv:1711.07553,

  5. [5]

    Circuitnet: An open- source dataset for machine learning applications in electronic design automation (eda).arXiv preprint arXiv:2208.01040, 2022

    Zhuomin Chai, Yuxiang Zhao, Yibo Lin, Wei Liu, Runsheng Wang, and Ru Huang. Circuitnet: An open- source dataset for machine learning applications in electronic design automation (eda).arXiv preprint arXiv:2208.01040, 2022. 1, 2, 4

  6. [6]

    CoRR abs/2110.11292

    Animesh Basak Chowdhury, Benjamin Tan, Ramesh Karri, and Siddharth Garg. Openabc-d: A large-scale dataset for machine learning guided integrated circuit synthesis.arXiv preprint arXiv:2110.11292, 2021. 1, 2, 4, 5

  7. [7]

    The cora dataset.https : / / graphsandnetworks

    Graph Consulting. The cora dataset.https : / / graphsandnetworks . com / the - cora - dataset/. Accessed: 2025-11-06. 2

  8. [8]

    Cktgnn: Circuit graph neural network for electronic design automa- tion

    Zehao Dong, Weidong Cao, Muhan Zhang, Dacheng Tao, Yixin Chen, and Xuan Zhang. Cktgnn: Circuit graph neural network for electronic design automa- tion. InThe Eleventh International Conference on Learning Representations, 2023. 1, 2

  9. [9]

    Benchmarking graph neural networks,

    Vijay Prakash Dwivedi, Chaitanya K Joshi, Thomas Laurent, Yoshua Bengio, and Xavier Bresson. Bench- marking graph neural networks.arXiv preprint arXiv:2003.00982, 2020. 1, 2, 7

  10. [10]

    arXiv preprint arXiv:2503.00205

    Jian Gao, Weidong Cao, Junyi Yang, and Xuan Zhang. Analoggenie: A generative engine for automatic dis- covery of analog circuit topologies.arXiv preprint arXiv:2503.00205, 2025. 5

  11. [11]

    Citeseer: An automatic citation indexing system

    C Lee Giles, Kurt D Bollacker, and Steve Lawrence. Citeseer: An automatic citation indexing system. In Proceedings of the third ACM conference on Digital libraries, pages 89–98, 1998. 1, 2

  12. [12]

    A timing engine inspired graph neural network model for pre-routing slack pre- diction

    Zizheng Guo, Mingjie Liu, Jiaqi Gu, Shuhan Zhang, David Z Pan, and Yibo Lin. A timing engine inspired graph neural network model for pre-routing slack pre- diction. InProceedings of the 59th ACM/IEEE Design Automation Conference, pages 1207–1212, 2022. 1, 2

  13. [13]

    Open graph benchmark: Datasets for machine learning on graphs.Advances in neural infor- mation processing systems, 33:22118–22133, 2020

    Weihua Hu, Matthias Fey, Marinka Zitnik, Yuxiao Dong, Hongyu Ren, Bowen Liu, Michele Catasta, and Jure Leskovec. Open graph benchmark: Datasets for machine learning on graphs.Advances in neural infor- mation processing systems, 33:22118–22133, 2020. 1, 2, 7, 8

  14. [14]

    Michael D Hutton, Jonathan Rose, Jerry P Grossman, and Derek G Corneil. Characterization and parameter- ized generation of synthetic combinational benchmark circuits.IEEE Transactions on Computer-Aided De- sign of Integrated Circuits and Systems, 17(10):985– 996, 2002. 2

  15. [15]

    Accelerating routability and timing optimization with open-source ai4eda dataset circuitnet and heteroge- neous platforms

    Xun Jiang, Zizheng Guo, Zhuomin Chai, Yuxiang Zhao, Yibo Lin, Runsheng Wang, and Ru Huang. Accelerating routability and timing optimization with open-source ai4eda dataset circuitnet and heteroge- neous platforms. In2023 IEEE/ACM International Conference on Computer Aided Design (ICCAD), pages 1–9. IEEE, 2023. 1

  16. [16]

    Circuitnet 2.0: An advanced dataset for promoting machine learning innovations in realis- tic chip design environment

    Xun Jiang, Yuxiang Zhao, Yibo Lin, Runsheng Wang, Ru Huang, et al. Circuitnet 2.0: An advanced dataset for promoting machine learning innovations in realis- tic chip design environment. InThe Twelfth Interna- tional Conference on Learning Representations, 2024. 1, 2, 4, 5

  17. [17]

    SNAP Datasets: Stanford large network dataset collection.http: //snap.stanford.edu/data, 2014

    Jure Leskovec and Andrej Krevl. SNAP Datasets: Stanford large network dataset collection.http: //snap.stanford.edu/data, 2014. 1, 2

  18. [18]

    Deeper insights into graph convolutional networks for semi- supervised learning

    Qimai Li, Zhichao Han, and Xiao-Ming Wu. Deeper insights into graph convolutional networks for semi- supervised learning. InThirty-Second AAAI confer- ence on artificial intelligence, 2018. 7, 8

  19. [19]

    Peiyu Liao, Dawei Guo, Zizheng Guo, Siting Liu, Yibo Lin, and Bei Yu. Dreamplace 4.0: Timing- driven placement with momentum-based net weight- ing and lagrangian-based refinement.IEEE Trans- actions on Computer-Aided Design of Integrated Cir- cuits and Systems, 42(10):3374–3387, 2023. 1

  20. [20]

    DE-HNN: An effective neural model for circuit netlist represen- tation

    Zhishang Luo, Truong Son Hy, Puoya Tabaghi, Micha¨el Defferrard, Elahe Rezaei, Ryan M Carey, Rhett Davis, Rajeev Jain, and Yusu Wang. DE-HNN: An effective neural model for circuit netlist represen- tation. InProceedings of The 27th International Con- ference on Artificial Intelligence and Statistics, pages 4258–4266, 2024. 1, 2

  21. [21]

    TUDataset: A collection of benchmark datasets for learning with graphs.arXiv preprint arXiv:2007.08663,

    Christopher Morris, Nils M Kriege, Franka Bause, Kristian Kersting, Petra Mutzel, and Marion Neu- mann. Tudataset: A collection of benchmark datasets for learning with graphs.arXiv preprint arXiv:2007.08663, 2020. 1, 2

  22. [22]

    Openls-dgf: An adaptive open- source dataset generation framework for machine learning tasks in logic synthesis.arXiv preprint arXiv:2411.09422, 2024

    Liwei Ni, Rui Wang, Miao Liu, Xingyu Meng, Xi- aoze Lin, Junfeng Liu, Guojie Luo, Zhufei Chu, Weikang Qian, Xiaoyan Yang, Biwei Xie, Xingquan Li, and Huawei Li. Openls-dgf: An adaptive open- source dataset generation framework for machine learning tasks in logic synthesis.arXiv preprint arXiv:2411.09422, 2024. 1, 2

  23. [23]

    Jingyu Pan, Chen-Chia Chang, Zhiyao Xie, Yiran Chen, and Hai Helen Li. Edalearn: A comprehen- sive rtl-to-signoff eda benchmark for democratized and reproducible ml for eda research.Proceedings of the 43rd IEEE/ACM International Conference on Computer-Aided Design, 2023. 2, 4

  24. [24]

    Dral, Matthias Rupp, and O

    Raghunathan Ramakrishnan, Pavlo O. Dral, Matthias Rupp, and O. Anatole von Lilienfeld. Quantum chem- istry structures and properties of 134k molecules.Sci- entific Data, 1:140022, 2014. 1, 2

  25. [25]

    Deep-learning-based pre-layout parasitic capacitance prediction on sram designs

    Shan Shen, Dingcheng Yang, Yuyang Xie, Chunyan Pei, Bei Yu, and Wenjian Yu. Deep-learning-based pre-layout parasitic capacitance prediction on sram designs. InProceedings of the Great Lakes Sympo- sium on VLSI 2024, pages 440–445, 2024. 1, 4

  26. [26]

    Few-shot learning on ams circuits and its application to parasitic capacitance pre- diction

    Shan Shen, Yibin Zhang, Hector Rodriguez Ro- driguez, and Wenjian Yu. Few-shot learning on ams circuits and its application to parasitic capacitance pre- diction. In2025 62nd ACM/IEEE Design Automation Conference (DAC), pages 1–7. IEEE, 2025. 4

  27. [27]

    Deepgate2: Functionality-aware circuit representation learning

    Zhengyuan Shi, Hongyang Pan, Sadaf Khan, Min Li, Yi Liu, Junhua Huang, Hui-Ling Zhen, Mingx- uan Yuan, Zhufei Chu, and Qiang Xu. Deepgate2: Functionality-aware circuit representation learning. In 2023 IEEE/ACM International Conference on Com- puter Aided Design (ICCAD), pages 1–9. IEEE, 2023. 5

  28. [28]

    Deepgate3: Towards scalable circuit representation learning

    Zhengyuan Shi, Ziyang Zheng, Sadaf Khan, Jianyuan Zhong, Min Li, and Qiang Xu. Deepgate3: Towards scalable circuit representation learning. InProceed- ings of the 43rd IEEE/ACM International Conference on Computer-Aided Design, pages 1–9, 2024. 5

  29. [29]

    Forgeeda: A comprehensive multimodal dataset for advancing eda

    Zhengyuan Shi, Zeju Li, Chengyu Ma, Yunhao Zhou, Ziyang Zheng, Jiawei Liu, Hongyang Pan, Lingfeng Zhou, Kezhi Li, Jiaying Zhu, et al. Forgeeda: A comprehensive multimodal dataset for advancing eda. arXiv preprint arXiv:2505.02016, 2025. 2, 5

  30. [30]

    Eda-schema: A graph datamodel schema and open dataset for digital design automa- tion

    Pratik Shrestha, Alec Aversa, Saran Phatharodom, and Ioannis Savidis. Eda-schema: A graph datamodel schema and open dataset for digital design automa- tion. InProceedings of the Great Lakes Symposium on VLSI 2024, pages 69–77, 2024. 2

  31. [31]

    Teague Sterling and John J. Irwin. Zinc 15—ligand discovery for everyone.Journal of Chemical Informa- tion and Modeling, 55(11):2324–2337, 2015. 1, 2

  32. [32]

    Amsnet: Netlist dataset for ams circuits

    Zhuofu Tao, Yichen Shi, Yiru Huo, Rui Ye, Zonghang Li, Li Huang, Chen Wu, Na Bai, Zhiping Yu, Ting- Jung Lin, and Lei He. Amsnet: Netlist dataset for ams circuits. In2024 IEEE LLM Aided Design Workshop (LAD), pages 1–5. IEEE, 2024. 1, 2

  33. [33]

    Graph attention networks

    Petar Velickovic, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Lio, and Yoshua Bengio. Graph attention networks. InInternational Conference on Learning Representations, 2018. 6

  34. [34]

    How Powerful are Graph Neural Networks?

    Keyulu Xu, Weihua Hu, Jure Leskovec, and Stefanie Jegelka. How powerful are graph neural networks? arXiv preprint arXiv:1810.00826, 2018. 6, 7, 8

  35. [35]

    Dingcheng Yang, Haoyuan Li, Wenjian Yu, Yuanbo Guo, and Wenjie Liang. Cnn-cap: Effective convo- lutional neural network-based capacitance models for interconnect capacitance extraction.ACM Transac- tions on Design Automation of Electronic Systems, 28 (4):1–22, 2023. 1

  36. [36]

    Versatile multi-stage graph neural network for circuit representation.Advances in Neural Informa- tion Processing Systems, 35:20313–20324, 2022

    Shuwen Yang, Zhihao Yang, Dong Li, Yingxueff Zhang, Zhanguang Zhang, Guojie Song, and Jianye Hao. Versatile multi-stage graph neural network for circuit representation.Advances in Neural Informa- tion Processing Systems, 35:20313–20324, 2022. 4

  37. [37]

    geomet- ric stage

    Marinka Zitnik and Jure Leskovec. Predicting multi- cellular function through multi-layer tissue networks. Bioinformatics, 2017. 1, 2 R2 : A Multi-View Circuit Graph Benchmark Suite from RTL to GDSII Supplementary Material This supplementary material provides comprehensive details, implementation specifics, and additional experi- mental analyses to suppor...