Recognition: unknown
R2G: A Multi-View Circuit Graph Benchmark Suite from RTL to GDSII
Pith reviewed 2026-05-10 16:57 UTC · model grok-4.3
The pith
View choice in circuit graph representations affects GNN accuracy more than the choice of GNN model.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
R2G supplies five stage-aware circuit-graph views that encode the same attributes yet differ only in feature attachment points. Systematic evaluation on GINE, GAT, and ResGatedGCN models reveals that representation choice produces larger accuracy swings than architecture choice, node-centric attachments generalize across placement and routing, and decoder depth of three to four layers converts divergent runs into near-perfect predictions with R² exceeding 0.99.
What carries the argument
Five stage-aware graph views with information parity, each attaching the same attribute set to different structural elements of the circuit netlist.
If this is right
- Selecting the graph view alone can raise prediction R² by more than 0.3 without any change to the underlying GNN.
- Node-centric views provide the strongest cross-stage generalization for both placement and routing tasks.
- Increasing decoder-head depth to three or four layers is sufficient to reach R² greater than 0.99 on these graphs.
- Future physical-design models can be compared fairly only after fixing the graph view, because representation effects otherwise confound architecture comparisons.
Where Pith is reading between the lines
- EDA tool flows could prioritize graph-construction pipelines over repeated architecture search, because representation gains appear larger.
- The same multi-view discipline might apply to other netlist-derived tasks such as timing or power estimation where stage-specific feature attachment has not yet been isolated.
- If new open-source designs are added with different technology nodes, the benchmark can test whether the observed dominance of node-centric views persists beyond the current 30 cores.
Load-bearing premise
The five views truly preserve identical information content and the synthesis-to-routing pipeline introduces no hidden biases from tool choices or IP core selection.
What would settle it
A controlled experiment in which the same GNN is retrained on each view while forcing identical node and edge feature sets; if the R² gap across views shrinks below 0.1, the claim that view choice dominates would be falsified.
Figures
read the original abstract
Graph neural networks (GNNs) are increasingly applied to physical design tasks such as congestion prediction and wirelength estimation, yet progress is hindered by inconsistent circuit representations and the absence of controlled evaluation protocols. We present R2G (RTL-to-GDSII), a multi-view circuit-graph benchmark suite that standardizes five stage-aware views with information parity (every view encodes the same attribute set, differing only in where features attach) over 30 open-source IP cores (up to $10^6$ nodes/edges). R2G provides an end-to-end DEF-to-graph pipeline spanning synthesis, placement, and routing stages, together with loaders, unified splits, domain metrics, and reproducible baselines. By decoupling representation choice from model choice, R2G isolates a confound that prior EDA and graph-ML benchmarks leave uncontrolled. In systematic studies with GINE, GAT, and ResGatedGCN, we find: (i) view choice dominates model choice, with Test R$^2$ varying by more than 0.3 across representations for a fixed GNN; (ii) node-centric views generalize best across both placement and routing; and (iii) decoder-head depth (3--4 layers) is the primary accuracy driver, turning divergent training into near-perfect predictions (R$^2$$>$0.99). Code and datasets are available at https://github.com/ShenShan123/R2G.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The paper introduces the R2G benchmark suite, which standardizes five stage-aware circuit-graph views (with claimed information parity) derived from 30 open-source IP cores via an end-to-end DEF-to-graph pipeline spanning synthesis, placement, and routing. It supplies loaders, unified splits, domain-specific metrics, and reproducible baselines using GINE, GAT, and ResGatedGCN. Experiments show that view choice dominates model choice (Test R² varies by >0.3 for fixed GNN), node-centric views generalize best across stages, and decoder-head depth (3-4 layers) drives near-perfect predictions (R² > 0.99).
Significance. If the central claims hold, R2G supplies a much-needed controlled benchmark for GNNs in physical design, isolating representation effects that prior EDA and graph-ML work leaves confounded. The open release of data, code, and baselines, together with systematic multi-architecture comparisons and concrete R² metrics, are concrete strengths that lower the barrier for reproducible research on congestion, wirelength, and related tasks.
major comments (1)
- [Abstract and §3] Abstract and §3 (Views and Pipeline): The central claim that performance gaps are attributable to representation (rather than unequal attribute content) rests on 'information parity' (every view encodes the same attribute set, differing only in attachment point). The manuscript asserts this via the shared DEF-to-graph pipeline but provides no explicit verification—e.g., no table or statistics confirming that gate types, netlist connectivity, timing, and congestion signals are identically present across all five views. Without this check, the reported R² differences (>0.3) could reflect information loss in some views rather than the intended confound.
minor comments (3)
- [§5] §5 (Experiments): The description of unified splits and any data exclusion criteria for the 30 IP cores is brief; adding explicit details on train/validation/test construction and size-based filtering would improve defensibility of the generalization claims.
- [Figure 3] Figure 3 (or equivalent results table): The R² values are reported to two decimals; showing standard deviation across seeds or runs would better convey stability of the 'near-perfect' (R² > 0.99) regime.
- [§3] Notation: The distinction between 'node-centric' and other views is clear in text but could be reinforced with a small summary table listing attachment points and feature dimensions for each view.
Simulated Author's Rebuttal
We thank the referee for the positive evaluation and the constructive comment on the information-parity claim. We address the concern point by point below.
read point-by-point responses
-
Referee: [Abstract and §3] Abstract and §3 (Views and Pipeline): The central claim that performance gaps are attributable to representation (rather than unequal attribute content) rests on 'information parity' (every view encodes the same attribute set, differing only in attachment point). The manuscript asserts this via the shared DEF-to-graph pipeline but provides no explicit verification—e.g., no table or statistics confirming that gate types, netlist connectivity, timing, and congestion signals are identically present across all five views. Without this check, the reported R² differences (>0.3) could reflect information loss in some views rather than the intended confound.
Authors: We agree that an explicit verification would strengthen the presentation. The information parity follows directly from the shared DEF-to-graph pipeline described in §3: all five views are generated from the identical DEF file and extract the same attribute set (gate types, netlist connectivity, timing arcs, congestion signals, etc.), differing solely in attachment point (node-centric vs. edge-centric) and stage-specific structural encoding. To address the referee’s request, we will insert a new table in the revised §3 that enumerates the attribute schema for each view together with summary statistics (e.g., attribute counts and value distributions) confirming equivalence. This addition will be purely expository and will not alter any experimental results or conclusions. revision: yes
Circularity Check
No circularity: benchmark supplies independent data, loaders, and empirical baselines
full rationale
The paper introduces a new multi-view circuit-graph dataset and pipeline over 30 IP cores, then reports direct empirical measurements (Test R² differences across views and models, decoder depth effects) on that data. No derivation, equation, or central claim reduces by construction to a parameter fitted inside the paper or to a self-citation chain; the information-parity assertion is a design statement about the DEF-to-graph conversion rather than a fitted or renamed result. The reported findings are falsifiable measurements on the released splits and therefore self-contained against external benchmarks.
Axiom & Free-Parameter Ledger
axioms (1)
- domain assumption GNN message passing can capture relevant circuit properties for prediction tasks when graphs are constructed from DEF files
Reference graph
Works this paper leans on
-
[1]
readthedocs
Openroad documentation.https://openroad. readthedocs . io / en / latest / main / README.html, . Accessed: 2025-11-06. 1
2025
-
[2]
https : / / github
Openroad: Unified application for rtl-to-gdsii flow. https : / / github . com / The - OpenROAD - Project/OpenROAD, . Accessed: 2025-11-06. 1, 3
2025
-
[3]
GNN-RE: Graph neural networks for reverse engineering of gate- level netlists.IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, 41(8): 2435–2448, 2022
Lilas Alrahis, Abhrajit Sengupta, Johann Knechtel, Satwik Patnaik, Hani Saleh, Baker Mohammad, Mah- moud Al-Qutayri, and Ozgur Sinanoglu. GNN-RE: Graph neural networks for reverse engineering of gate- level netlists.IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, 41(8): 2435–2448, 2022. 1, 2
2022
-
[4]
Xavier Bresson and Thomas Laurent. Residual gated graph convnets.arXiv preprint arXiv:1711.07553,
-
[5]
Zhuomin Chai, Yuxiang Zhao, Yibo Lin, Wei Liu, Runsheng Wang, and Ru Huang. Circuitnet: An open- source dataset for machine learning applications in electronic design automation (eda).arXiv preprint arXiv:2208.01040, 2022. 1, 2, 4
-
[6]
Animesh Basak Chowdhury, Benjamin Tan, Ramesh Karri, and Siddharth Garg. Openabc-d: A large-scale dataset for machine learning guided integrated circuit synthesis.arXiv preprint arXiv:2110.11292, 2021. 1, 2, 4, 5
-
[7]
The cora dataset.https : / / graphsandnetworks
Graph Consulting. The cora dataset.https : / / graphsandnetworks . com / the - cora - dataset/. Accessed: 2025-11-06. 2
2025
-
[8]
Cktgnn: Circuit graph neural network for electronic design automa- tion
Zehao Dong, Weidong Cao, Muhan Zhang, Dacheng Tao, Yixin Chen, and Xuan Zhang. Cktgnn: Circuit graph neural network for electronic design automa- tion. InThe Eleventh International Conference on Learning Representations, 2023. 1, 2
2023
-
[9]
Benchmarking graph neural networks,
Vijay Prakash Dwivedi, Chaitanya K Joshi, Thomas Laurent, Yoshua Bengio, and Xavier Bresson. Bench- marking graph neural networks.arXiv preprint arXiv:2003.00982, 2020. 1, 2, 7
-
[10]
arXiv preprint arXiv:2503.00205
Jian Gao, Weidong Cao, Junyi Yang, and Xuan Zhang. Analoggenie: A generative engine for automatic dis- covery of analog circuit topologies.arXiv preprint arXiv:2503.00205, 2025. 5
-
[11]
Citeseer: An automatic citation indexing system
C Lee Giles, Kurt D Bollacker, and Steve Lawrence. Citeseer: An automatic citation indexing system. In Proceedings of the third ACM conference on Digital libraries, pages 89–98, 1998. 1, 2
1998
-
[12]
A timing engine inspired graph neural network model for pre-routing slack pre- diction
Zizheng Guo, Mingjie Liu, Jiaqi Gu, Shuhan Zhang, David Z Pan, and Yibo Lin. A timing engine inspired graph neural network model for pre-routing slack pre- diction. InProceedings of the 59th ACM/IEEE Design Automation Conference, pages 1207–1212, 2022. 1, 2
2022
-
[13]
Open graph benchmark: Datasets for machine learning on graphs.Advances in neural infor- mation processing systems, 33:22118–22133, 2020
Weihua Hu, Matthias Fey, Marinka Zitnik, Yuxiao Dong, Hongyu Ren, Bowen Liu, Michele Catasta, and Jure Leskovec. Open graph benchmark: Datasets for machine learning on graphs.Advances in neural infor- mation processing systems, 33:22118–22133, 2020. 1, 2, 7, 8
2020
-
[14]
Michael D Hutton, Jonathan Rose, Jerry P Grossman, and Derek G Corneil. Characterization and parameter- ized generation of synthetic combinational benchmark circuits.IEEE Transactions on Computer-Aided De- sign of Integrated Circuits and Systems, 17(10):985– 996, 2002. 2
2002
-
[15]
Accelerating routability and timing optimization with open-source ai4eda dataset circuitnet and heteroge- neous platforms
Xun Jiang, Zizheng Guo, Zhuomin Chai, Yuxiang Zhao, Yibo Lin, Runsheng Wang, and Ru Huang. Accelerating routability and timing optimization with open-source ai4eda dataset circuitnet and heteroge- neous platforms. In2023 IEEE/ACM International Conference on Computer Aided Design (ICCAD), pages 1–9. IEEE, 2023. 1
2023
-
[16]
Circuitnet 2.0: An advanced dataset for promoting machine learning innovations in realis- tic chip design environment
Xun Jiang, Yuxiang Zhao, Yibo Lin, Runsheng Wang, Ru Huang, et al. Circuitnet 2.0: An advanced dataset for promoting machine learning innovations in realis- tic chip design environment. InThe Twelfth Interna- tional Conference on Learning Representations, 2024. 1, 2, 4, 5
2024
-
[17]
SNAP Datasets: Stanford large network dataset collection.http: //snap.stanford.edu/data, 2014
Jure Leskovec and Andrej Krevl. SNAP Datasets: Stanford large network dataset collection.http: //snap.stanford.edu/data, 2014. 1, 2
2014
-
[18]
Deeper insights into graph convolutional networks for semi- supervised learning
Qimai Li, Zhichao Han, and Xiao-Ming Wu. Deeper insights into graph convolutional networks for semi- supervised learning. InThirty-Second AAAI confer- ence on artificial intelligence, 2018. 7, 8
2018
-
[19]
Peiyu Liao, Dawei Guo, Zizheng Guo, Siting Liu, Yibo Lin, and Bei Yu. Dreamplace 4.0: Timing- driven placement with momentum-based net weight- ing and lagrangian-based refinement.IEEE Trans- actions on Computer-Aided Design of Integrated Cir- cuits and Systems, 42(10):3374–3387, 2023. 1
2023
-
[20]
DE-HNN: An effective neural model for circuit netlist represen- tation
Zhishang Luo, Truong Son Hy, Puoya Tabaghi, Micha¨el Defferrard, Elahe Rezaei, Ryan M Carey, Rhett Davis, Rajeev Jain, and Yusu Wang. DE-HNN: An effective neural model for circuit netlist represen- tation. InProceedings of The 27th International Con- ference on Artificial Intelligence and Statistics, pages 4258–4266, 2024. 1, 2
2024
-
[21]
Christopher Morris, Nils M Kriege, Franka Bause, Kristian Kersting, Petra Mutzel, and Marion Neu- mann. Tudataset: A collection of benchmark datasets for learning with graphs.arXiv preprint arXiv:2007.08663, 2020. 1, 2
-
[22]
Liwei Ni, Rui Wang, Miao Liu, Xingyu Meng, Xi- aoze Lin, Junfeng Liu, Guojie Luo, Zhufei Chu, Weikang Qian, Xiaoyan Yang, Biwei Xie, Xingquan Li, and Huawei Li. Openls-dgf: An adaptive open- source dataset generation framework for machine learning tasks in logic synthesis.arXiv preprint arXiv:2411.09422, 2024. 1, 2
-
[23]
Jingyu Pan, Chen-Chia Chang, Zhiyao Xie, Yiran Chen, and Hai Helen Li. Edalearn: A comprehen- sive rtl-to-signoff eda benchmark for democratized and reproducible ml for eda research.Proceedings of the 43rd IEEE/ACM International Conference on Computer-Aided Design, 2023. 2, 4
2023
-
[24]
Dral, Matthias Rupp, and O
Raghunathan Ramakrishnan, Pavlo O. Dral, Matthias Rupp, and O. Anatole von Lilienfeld. Quantum chem- istry structures and properties of 134k molecules.Sci- entific Data, 1:140022, 2014. 1, 2
2014
-
[25]
Deep-learning-based pre-layout parasitic capacitance prediction on sram designs
Shan Shen, Dingcheng Yang, Yuyang Xie, Chunyan Pei, Bei Yu, and Wenjian Yu. Deep-learning-based pre-layout parasitic capacitance prediction on sram designs. InProceedings of the Great Lakes Sympo- sium on VLSI 2024, pages 440–445, 2024. 1, 4
2024
-
[26]
Few-shot learning on ams circuits and its application to parasitic capacitance pre- diction
Shan Shen, Yibin Zhang, Hector Rodriguez Ro- driguez, and Wenjian Yu. Few-shot learning on ams circuits and its application to parasitic capacitance pre- diction. In2025 62nd ACM/IEEE Design Automation Conference (DAC), pages 1–7. IEEE, 2025. 4
2025
-
[27]
Deepgate2: Functionality-aware circuit representation learning
Zhengyuan Shi, Hongyang Pan, Sadaf Khan, Min Li, Yi Liu, Junhua Huang, Hui-Ling Zhen, Mingx- uan Yuan, Zhufei Chu, and Qiang Xu. Deepgate2: Functionality-aware circuit representation learning. In 2023 IEEE/ACM International Conference on Com- puter Aided Design (ICCAD), pages 1–9. IEEE, 2023. 5
2023
-
[28]
Deepgate3: Towards scalable circuit representation learning
Zhengyuan Shi, Ziyang Zheng, Sadaf Khan, Jianyuan Zhong, Min Li, and Qiang Xu. Deepgate3: Towards scalable circuit representation learning. InProceed- ings of the 43rd IEEE/ACM International Conference on Computer-Aided Design, pages 1–9, 2024. 5
2024
-
[29]
Forgeeda: A comprehensive multimodal dataset for advancing eda
Zhengyuan Shi, Zeju Li, Chengyu Ma, Yunhao Zhou, Ziyang Zheng, Jiawei Liu, Hongyang Pan, Lingfeng Zhou, Kezhi Li, Jiaying Zhu, et al. Forgeeda: A comprehensive multimodal dataset for advancing eda. arXiv preprint arXiv:2505.02016, 2025. 2, 5
-
[30]
Eda-schema: A graph datamodel schema and open dataset for digital design automa- tion
Pratik Shrestha, Alec Aversa, Saran Phatharodom, and Ioannis Savidis. Eda-schema: A graph datamodel schema and open dataset for digital design automa- tion. InProceedings of the Great Lakes Symposium on VLSI 2024, pages 69–77, 2024. 2
2024
-
[31]
Teague Sterling and John J. Irwin. Zinc 15—ligand discovery for everyone.Journal of Chemical Informa- tion and Modeling, 55(11):2324–2337, 2015. 1, 2
2015
-
[32]
Amsnet: Netlist dataset for ams circuits
Zhuofu Tao, Yichen Shi, Yiru Huo, Rui Ye, Zonghang Li, Li Huang, Chen Wu, Na Bai, Zhiping Yu, Ting- Jung Lin, and Lei He. Amsnet: Netlist dataset for ams circuits. In2024 IEEE LLM Aided Design Workshop (LAD), pages 1–5. IEEE, 2024. 1, 2
2024
-
[33]
Graph attention networks
Petar Velickovic, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Lio, and Yoshua Bengio. Graph attention networks. InInternational Conference on Learning Representations, 2018. 6
2018
-
[34]
How Powerful are Graph Neural Networks?
Keyulu Xu, Weihua Hu, Jure Leskovec, and Stefanie Jegelka. How powerful are graph neural networks? arXiv preprint arXiv:1810.00826, 2018. 6, 7, 8
work page internal anchor Pith review arXiv 2018
-
[35]
Dingcheng Yang, Haoyuan Li, Wenjian Yu, Yuanbo Guo, and Wenjie Liang. Cnn-cap: Effective convo- lutional neural network-based capacitance models for interconnect capacitance extraction.ACM Transac- tions on Design Automation of Electronic Systems, 28 (4):1–22, 2023. 1
2023
-
[36]
Versatile multi-stage graph neural network for circuit representation.Advances in Neural Informa- tion Processing Systems, 35:20313–20324, 2022
Shuwen Yang, Zhihao Yang, Dong Li, Yingxueff Zhang, Zhanguang Zhang, Guojie Song, and Jianye Hao. Versatile multi-stage graph neural network for circuit representation.Advances in Neural Informa- tion Processing Systems, 35:20313–20324, 2022. 4
2022
-
[37]
Marinka Zitnik and Jure Leskovec. Predicting multi- cellular function through multi-layer tissue networks. Bioinformatics, 2017. 1, 2 R2 : A Multi-View Circuit Graph Benchmark Suite from RTL to GDSII Supplementary Material This supplementary material provides comprehensive details, implementation specifics, and additional experi- mental analyses to suppor...
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.