pith. machine review for the scientific record. sign in

arxiv: 1903.02428 · v3 · submitted 2019-03-06 · 💻 cs.LG · stat.ML

Recognition: 2 theorem links

· Lean Theorem

Fast Graph Representation Learning with PyTorch Geometric

Authors on Pith no claims yet

Pith reviewed 2026-05-13 19:56 UTC · model grok-4.3

classification 💻 cs.LG stat.ML
keywords graph neural networksPyTorchdeep learning on graphsCUDA kernelsmini-batch processingpoint cloudsrelational learning
0
0 comments X

The pith

PyTorch Geometric speeds graph learning on GPUs via sparse acceleration, custom kernels, and variable-size batching.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper introduces PyTorch Geometric, a PyTorch-based library for deep learning on graphs, point clouds, and manifolds. It establishes that the library reaches high data throughput by combining sparse GPU acceleration, dedicated CUDA kernels for graph operations, and efficient mini-batch handling that processes inputs of differing sizes without padding. A sympathetic reader would care because these features make training graph neural networks and related models practical on large, irregular datasets that previously required heavy custom engineering. The authors support the claim with a comparative study of multiple methods run under uniform evaluation conditions.

Core claim

PyTorch Geometric is a library for deep learning on irregularly structured input data such as graphs, point clouds and manifolds, built upon PyTorch. In addition to general graph data structures and processing methods, it contains a variety of recently published methods from the domains of relational learning and 3D data processing. PyTorch Geometric achieves high data throughput by leveraging sparse GPU acceleration, by providing dedicated CUDA kernels and by introducing efficient mini-batch handling for input examples of different size.

What carries the argument

Sparse GPU tensor representations together with custom CUDA kernels and dynamic mini-batch collation that accommodates graphs and point clouds of varying sizes.

If this is right

  • Training graph neural networks on large collections of variable-sized graphs becomes feasible without custom data-loading optimizations.
  • A single consistent code base allows direct comparison and reproduction of multiple relational learning and 3D-processing methods.
  • Researchers can scale experiments to larger point-cloud or manifold datasets while keeping GPU utilization high.
  • Mini-batch training on heterogeneous input sizes no longer requires manual padding or grouping steps.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Widespread use of the library could create de-facto standard implementations and benchmarks for graph representation learning.
  • The same sparse-acceleration pattern may transfer to other frameworks or to domains with variable-length sequences such as text or audio.
  • Extensions to multi-GPU or distributed settings would follow naturally from the existing mini-batch design.

Load-bearing premise

The dedicated CUDA kernels and mini-batch routines are implemented correctly and the performance comparisons use identical, reproducible evaluation settings for every method.

What would settle it

Re-running the throughput benchmarks on the same hardware and observing that another library or implementation processes an equal number of examples per second or faster would falsify the performance claim.

read the original abstract

We introduce PyTorch Geometric, a library for deep learning on irregularly structured input data such as graphs, point clouds and manifolds, built upon PyTorch. In addition to general graph data structures and processing methods, it contains a variety of recently published methods from the domains of relational learning and 3D data processing. PyTorch Geometric achieves high data throughput by leveraging sparse GPU acceleration, by providing dedicated CUDA kernels and by introducing efficient mini-batch handling for input examples of different size. In this work, we present the library in detail and perform a comprehensive comparative study of the implemented methods in homogeneous evaluation scenarios.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 1 minor

Summary. The manuscript introduces PyTorch Geometric, a library for deep learning on irregularly structured data such as graphs, point clouds, and manifolds, built on top of PyTorch. It provides general graph data structures, processing methods, and implementations of various methods from relational learning and 3D data processing. The library claims to achieve high data throughput through sparse GPU acceleration, dedicated CUDA kernels, and efficient mini-batch handling for inputs of varying sizes. A comprehensive comparative study of the implemented methods is presented in homogeneous evaluation scenarios.

Significance. If the performance claims hold, this work offers a significant contribution by delivering an open-source, high-performance framework that facilitates research and development in graph representation learning. The engineering focus on throughput and scalability addresses key practical challenges in applying deep learning to irregular data, potentially enabling larger-scale experiments and broader adoption of these techniques.

major comments (2)
  1. [Experimental Evaluation] The comparative study is central to validating the throughput claims, but the manuscript should provide more details on the hardware configuration, dataset sizes, and exact baseline implementations to allow independent verification of the reported speedups.
  2. [Library Design] While the use of dedicated CUDA kernels is highlighted as key to efficiency, the paper would benefit from including complexity analysis or pseudocode for the mini-batch handling routine to demonstrate how it achieves better performance than standard PyTorch operations for variable-sized graphs.
minor comments (1)
  1. [Abstract] Consider adding a note on the open-source availability and GitHub repository link for the library to enhance accessibility.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the positive assessment and constructive comments on our manuscript. We address each major comment below and have incorporated the requested details into the revised version.

read point-by-point responses
  1. Referee: [Experimental Evaluation] The comparative study is central to validating the throughput claims, but the manuscript should provide more details on the hardware configuration, dataset sizes, and exact baseline implementations to allow independent verification of the reported speedups.

    Authors: We agree that additional experimental details will improve reproducibility. In the revised manuscript we have added a new subsection in the experimental evaluation that specifies the hardware (NVIDIA Tesla V100 GPUs, 32 GB memory, CUDA 10.0), exact dataset sizes and splits for all benchmarks, and precise baseline implementations including library versions, commit hashes, and any custom modifications. revision: yes

  2. Referee: [Library Design] While the use of dedicated CUDA kernels is highlighted as key to efficiency, the paper would benefit from including complexity analysis or pseudocode for the mini-batch handling routine to demonstrate how it achieves better performance than standard PyTorch operations for variable-sized graphs.

    Authors: We appreciate the suggestion. The revised manuscript now includes both a complexity analysis (O(N + E) for the collate routine versus O(B * max_size) for padded baselines) and pseudocode for the mini-batch collation procedure in Section 3.2, clarifying how sparse tensor construction and dynamic batching avoid unnecessary padding overhead. revision: yes

Circularity Check

0 steps flagged

No significant circularity detected

full rationale

The paper introduces PyTorch Geometric as a software library for graph deep learning, describing its data structures, CUDA kernels, mini-batch routines, and included methods from prior literature, then reports empirical throughput and accuracy benchmarks. No load-bearing mathematical derivations, fitted parameters renamed as predictions, or self-referential equations exist; performance claims rest on external comparative studies and open-source implementation rather than internal construction. Self-citations, if present, are not used to justify uniqueness theorems or ansatzes that reduce the central contribution to its own inputs.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

The work rests on the standard assumption that PyTorch provides reliable GPU tensor operations and that graph data can be represented as sparse adjacency matrices and feature tensors.

axioms (1)
  • standard math PyTorch supplies efficient sparse tensor operations on GPU
    The library is built directly on top of PyTorch and relies on its sparse tensor support for acceleration.

pith-pipeline@v0.9.0 · 5384 in / 1115 out tokens · 36183 ms · 2026-05-13T19:56:34.171049+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

What do these tags mean?
matches
The paper's claim is directly supported by a theorem in the formal canon.
supports
The theorem supports part of the paper's argument, but the paper may add assumptions or extra steps.
extends
The paper goes beyond the formal theorem; the theorem is a base layer rather than the whole result.
uses
The paper appears to rely on the theorem as machinery.
contradicts
The paper's claim conflicts with a theorem or certificate in the canon.
unclear
Pith found a possible connection, but the passage is too broad, indirect, or ambiguous to say the theorem truly supports the claim.

Forward citations

Cited by 26 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Learning Scenario Reduction for Two-Stage Robust Optimization with Discrete Uncertainty

    cs.AI 2026-05 conditional novelty 7.0

    NeurPRISE trains a GNN-Transformer via imitation learning to mimic a lookahead heuristic for scenario reduction in 2RO, delivering 7-200x speedups with competitive regret on three test problems and zero-shot generalization.

  2. Frequency-Space Mechanics: A Sequence and Coordinate-Free Representation for Protein Function Prediction

    q-bio.BM 2026-05 unverdicted novelty 7.0

    Vibrational mode graphs from molecular dynamics enable sequence-free protein function prediction via graph neural networks, with entrainment improving signals for collective dynamics.

  3. ATLAS: Efficient Out-of-Core Inference for Billion-Scale Graph Neural Networks

    cs.DC 2026-05 unverdicted novelty 7.0

    ATLAS achieves 12-30x faster out-of-core full-graph GNN inference on graphs up to 4B edges by switching to broadcast-based layer-wise execution with graph reordering, minimum-pending-message eviction, and GPU-accelera...

  4. PiGGO: Physics-Guided Learnable Graph Kalman Filters for Virtual Sensing of Nonlinear Dynamic Structures under Uncertainty

    cs.LG 2026-04 unverdicted novelty 7.0

    PiGGO integrates a learned graph neural ODE as the continuous-time dynamics model within an extended Kalman filter to enable online virtual sensing and uncertainty-aware state estimation for nonlinear dynamic systems ...

  5. HopRank: Self-Supervised LLM Preference-Tuning on Graphs for Few-Shot Node Classification

    cs.CL 2026-04 unverdicted novelty 7.0

    HopRank is a self-supervised LLM-tuning method that turns node classification into link prediction via hierarchical hop-based preference sampling, matching supervised GNN performance with zero labeled data on text-att...

  6. Hypergraph Neural Diffusion: A PDE-Inspired Framework for Hypergraph Message Passing

    cs.LG 2026-04 unverdicted novelty 7.0

    HND models hypergraph feature propagation as an anisotropic diffusion process governed by a continuous-time PDE, discretized into stable neural layers with energy dissipation and boundedness guarantees.

  7. High-dimensional inference for the $\gamma$-ray sky with differentiable programming

    astro-ph.HE 2026-04 unverdicted novelty 7.0

    A differentiable forward model and likelihood enable probabilistic inference over many spatial morphologies for the Galactic Center gamma-ray Excess using variational methods on GPUs.

  8. Towards Near-Real-Time Telemetry-Aware Routing with Neural Routing Algorithms

    cs.LG 2026-04 unverdicted novelty 7.0

    LOGGIA is a delay-aware graph neural routing algorithm using pre-training and RL that outperforms shortest-path and other neural methods in realistic network simulations.

  9. TOPOS: High-Fidelity and Efficient Industry-Grade 3D Head Generation

    cs.CV 2026-05 unverdicted novelty 6.0

    TOPOS creates high-fidelity 3D heads with fixed industry topology from single images via a specialized VAE with Perceiver Resampler and a rectified flow transformer.

  10. SAGE: A Self-Evolving Agentic Graph-Memory Engine for Structure-Aware Associative Memory

    cs.AI 2026-05 unverdicted novelty 6.0

    SAGE is a self-evolving agentic graph-memory engine that dynamically constructs and refines structured memory graphs via writer-reader feedback, yielding performance gains on multi-hop QA, open-domain retrieval, and l...

  11. Invariant-Based Diagnostics for Graph Benchmarks

    cs.LG 2026-05 unverdicted novelty 6.0

    Graph invariants serve as expressive, task-agnostic baselines that characterize structural heterogeneity and match trained models across 26 datasets, indicating that expressivity is not the primary driver of performance.

  12. Mochi: Aligning Pre-training and Inference for Efficient Graph Foundation Models via Meta-Learning

    cs.LG 2026-04 unverdicted novelty 6.0

    Mochi aligns pre-training with inference via meta-learning for efficient graph foundation models, matching or exceeding prior models on 25 datasets with 8-27x less training time.

  13. Robustness of Spatio-temporal Graph Neural Networks for Fault Location in Partially Observable Distribution Grids

    cs.LG 2026-04 unverdicted novelty 6.0

    Measured-only graph topologies enable STGNNs to achieve up to 11-point F1 gains and 6x faster training versus full-topology GNNs and RNN baselines for fault location in partially observable distribution grids.

  14. TACENR: Task-Agnostic Contrastive Explanations for Node Representations

    cs.LG 2026-04 unverdicted novelty 6.0

    TACENR introduces a contrastive-learning method that identifies the most influential attribute, proximity, and structural features in node representations in a task-agnostic manner.

  15. LogosKG: Hardware-Optimized Scalable and Interpretable Knowledge Graph Retrieval

    cs.CL 2026-04 unverdicted novelty 6.0

    LogosKG delivers a novel hardware-aligned system for efficient multi-hop retrieval on billion-edge knowledge graphs without sacrificing fidelity, demonstrated via biomedical KG-LLM applications.

  16. A Structure-Preserving Graph Neural Solver for Parametric Hyperbolic Conservation Laws

    physics.comp-ph 2026-04 unverdicted novelty 6.0

    A structure-preserving GNN solver for parametric hyperbolic conservation laws achieves superior long-horizon stability and orders-of-magnitude speedups over high-resolution simulations on supersonic flow benchmarks.

  17. PUFFIN: Protein Unit Discovery with Functional Supervision

    q-bio.BM 2026-04 unverdicted novelty 6.0

    PUFFIN discovers protein units by jointly learning structural partitioning of residue graphs and functional supervision via a graph neural network with structure-aware pooling.

  18. TOPCELL: Topology Optimization of Standard Cell via LLMs

    cs.LG 2026-04 unverdicted novelty 6.0

    TOPCELL reformulates standard cell topology optimization as an LLM generative task with GRPO fine-tuning, outperforming base models and matching exhaustive solvers with 85.91x speedup in 2nm/7nm industrial flows.

  19. Disorder-induced chirality in superconductor-ferromagnet heterostructures revealed by neutron scattering and multiscale modeling

    cond-mat.mtrl-sci 2026-04 unverdicted novelty 6.0

    Chemical disorder plus compositional gradients in FePd films produce finite Dzyaloshinskii-Moriya interactions that stabilize chiral magnetic modulations with mixed Bloch-Néel character.

  20. FlexMS is a flexible framework for benchmarking deep learning-based mass spectrum prediction tools in metabolomics

    cs.AI 2026-02 unverdicted novelty 6.0

    FlexMS is a new flexible benchmarking framework that lets researchers dynamically combine deep learning architectures and evaluate their mass spectrum prediction performance on public metabolomics datasets using multi...

  21. Astro Generative Network: A Variational Framework for Controlled Node Insertion in Incomplete Complex Networks

    cs.SI 2026-05 unverdicted novelty 5.0

    AGN is a variational framework for inserting plausible new nodes into incomplete networks by latent sampling and similarity attachment, shown on synthetic data to keep clustering and modularity changes modest compared...

  22. Compositional Quantum Heuristics for Max-Clique Detection

    quant-ph 2026-05 unverdicted novelty 5.0

    Compositional quantum circuits with symmetry-induced invariant losses produce trainable equivariant quantum GNNs that generalize on max-clique problems and improve hybrid recursive search accuracy and scalability.

  23. From Spherical to Gaussian: A Comparative Analysis of Point Cloud Cropping Strategies in Large-Scale 3D Environments

    cs.CV 2026-05 unverdicted novelty 5.0

    Gaussian and linear cropping strategies for large point clouds improve 3D neural network performance over spherical crops, especially in outdoor scenes, and achieve new state-of-the-art results.

  24. A Universal Space of Brain Dynamics for Unveiling Cognitive Transitions and Individual Differences

    q-bio.QM 2026-05 unverdicted novelty 5.0

    UBD creates a universal space for brain dynamics that predicts fMRI signals with Pearson's r greater than 0.9 across eight states and 963 subjects, revealing mechanisms of cognitive transitions and individual differences.

  25. Robustness of Spatio-temporal Graph Neural Networks for Fault Location in Partially Observable Distribution Grids

    cs.LG 2026-04 unverdicted novelty 5.0

    Measured-only STGNNs (RGATv2, RGSAGE) achieve up to 11 F1 points higher and 6x faster training than RNN baselines for fault location on the IEEE 123-bus feeder under partial observability.

  26. On Improving Graph Neural Networks for QSAR by Pre-training on Extended-Connectivity Fingerprints

    cs.LG 2026-05 unverdicted novelty 4.0

    Pre-training GNNs on ECFP prediction produces statistically significant QSAR gains on five of six Biogen benchmarks with OOD splits, but underperforms on heterogeneous datasets and complex endpoints like binding affinity.

Reference graph

Works this paper leans on

52 extracted references · 52 canonical work pages · cited by 25 Pith papers · 2 internal anchors

  1. [1]

    P. W. Battaglia, J. B. Hamrick, V. Bapst, A. Sanchez - Gonzalez, V. F. Zambaldi, M. Malinowski, A. Tacchetti, D. Raposo, A. Santoro, R. Faulkner, C . G \" u l c ehre, F. Song, A. J. Ballard, J. Gilmer, G. E. Dahl, A. Vaswani, K. Allen, C. Nash, V. Langston, C. Dyer, N. Heess, D. Wierstra, P. Kohli, M. Botvinick, O. Vinyals, Y. Li, and R. Pascanu. Relation...

  2. [2]

    F. M. Bianchi, D. Grattarola, L. Livi, and C. Alippi. Graph neural networks with convolutional ARMA filters. CoRR, abs/1901.01343, 2019

  3. [3]

    F. Bogo, J. Romero, M. Loper, and M. J. Black. FAUST : Dataset and evaluation for 3D mesh registration. In CVPR, 2014

  4. [4]

    Bojchevski and S

    A. Bojchevski and S. G \"u nnemann. Deep gaussian embedding of attributed graphs: Unsupervised inductive learning via ranking. In ICLR, 2018

  5. [5]

    Boschee, J

    E. Boschee, J. Lautenschlager, S. O'Brien, S. Shellman, J. Starz, and M. Ward. ICEWS coded event data. Harvard Dataverse, 2015

  6. [6]

    M. M. Bronstein, J. Bruna, Y. LeCun, A. Szlam, and P. Vandergheynst. Geometric deep learning: Going beyond euclidean data. In Signal Processing Magazine, 2017

  7. [7]

    Cai and Y

    C. Cai and Y. Wang. A simple yet effective baseline for non-attribute graph classification. CoRR, abs/1811.03508, 2018

  8. [8]

    Cangea, P

    C. Cangea, P. Veli c kovi \' c , N. Jovanovi \' c , T. N. Kipf, and P. Li \` o . Towards sparse hierarchical graph classifiers. In NeurIPS-W, 2018

  9. [9]

    A. X. Chang, T. Funkhouser, L. J. Guibas, P. Hanrahan, Q. Huang, Z. Li, S. Savarese, M. Savva, S. Song, H. Su, J. Xiao, L. Yi, and F. Yu. ShapeNet : An information-rich 3D model repository. CoRR, abs/1512.03012, 2015

  10. [10]

    Defferrard, X

    M. Defferrard, X. Bresson, and P. Vandergheynst. Convolutional neural networks on graphs with fast localized spectral filtering. In NIPS, 2016

  11. [11]

    T. Derr, Y. Ma, and J. Tang. Signed graph convolutional networks. In ICDM, 2018

  12. [12]

    I. S. Dhillon, Y. Guan, and B. Kulis. Weighted graph cuts without eigenvectors: A multilevel approach. In TPAMI, 2007

  13. [13]

    B. O. Fagginger Auer and R. H. Bisseling. A GPU algorithm for greedy graph matching. In Facing the Multicore - Challenge II - Aspects of New Paradigms and Technologies in Parallel Computing , 2011

  14. [14]

    M. Fey. Just jump: Dynamic neighborhood aggregation in graph neural networks. In ICLR-W, 2019

  15. [15]

    M. Fey, J. E. Lenssen, F. Weichert, and H. M \"u ller. SplineCNN : Fast geometric deep learning with continuous B -spline kernels. In CVPR, 2018

  16. [16]

    Gao and S

    H. Gao and S. Ji. Graph U - N et. https://openreview.net/forum?id=HJePRoAct7, 2018. Submitted to ICLR

  17. [17]

    Gilmer, S

    J. Gilmer, S. S. Schoenholz, P. F. Riley, O. Vinyals, and G. E. Dahl. Neural message passing for quantum chemistry. In ICML, 2017

  18. [18]

    Guerrero, Y

    P. Guerrero, Y. Kleiman, M. Ovsjanikov, and N. J. Mitra. PCPNet : Learning local shape properties from raw point clouds. Computer Graphics Forum, 37, 2018

  19. [19]

    W. L. Hamilton, R. Ying, and J. Leskovec. Inductive representation learning on large graphs. In NIPS, 2017

  20. [20]

    W. Jin, C. Zhang, P. Szekely, and X. Ren. Recurrent event network for reasoning over temporal knowledge graphs. In ICLR-W, 2019

  21. [21]

    Kersting, N

    K. Kersting, N. M. Kriege, C. Morris, P. Mutzel, and M. Neumann. Benchmark data sets for graph kernels. http://graphkernels.cs.tu-dortmund.de, 2016

  22. [22]

    T. N. Kipf and M. Welling. Variational graph auto-encoders. In NIPS-W, 2016

  23. [23]

    T. N. Kipf and M. Welling. Semi-supervised classification with graph convolutional networks. In ICLR, 2017

  24. [24]

    Klicpera, A

    J. Klicpera, A. Bojchevski, and S. G \"u nnemann. Predict then propagate: Graph neural networks meet personalized PageRank . In ICLR, 2019

  25. [25]

    Kumar, F

    S. Kumar, F. Spezzano, V. Subrahmanian, and C. Faloutsos. Edge weight prediction in weighted signed networks. In ICDM, 2016

  26. [26]

    Leetaru and P

    K. Leetaru and P. A. Schrodt. GDELT : Global data on events, location, and tone. ISA Annual Convention, 2013

  27. [27]

    Y. Li, D. Tarlow, M. Brockschmidt, and R. Zemel. Gated graph sequence neural networks. In ICLR, 2016

  28. [28]

    Y. Li, R. Bu, M. Sun, W. Wu, X. Di, and B. Chen. PointCNN : Convolution on X -transformed points. In NeurIPS, 2018

  29. [29]

    Montavon, M

    G. Montavon, M. Rupp, V. Gobre, A. Vazquez-Mayagoitia, K. Hansen, A. Tkatchenko, K. M \"u ller, and O. A. von Lilienfeld. Machine learning of molecular electronic properties in chemical compound space. New Journal of Physics, 2013

  30. [30]

    Monti, D

    F. Monti, D. Boscaini, J. Masci, E. Rodol \` a , J. Svoboda, and M. M. Bronstein. Geometric deep learning on graphs and manifolds using mixture model CNN s. In CVPR, 2017

  31. [31]

    Morris, M

    C. Morris, M. Ritzert, M. Fey, W. L. Hamilton, J. E. Lenssen, G. Rattan, and M. Grohe. W eisfeiler and L eman go neural: Higher-order graph neural networks. In AAAI, 2019

  32. [32]

    S. Pan, R. Hu, G. Long, J. Jiang, L. Yao, and C. Zhang. Adversarially regularized graph autoencoder for graph embedding. In IJCAI, 2018

  33. [33]

    Paszke, S

    A. Paszke, S. Gross, S. Chintala, G. Chanan, E. Yang, Z. DeVito, Z. Lin, A. Desmaison, L. Antiga, and A. Lerer. Automatic differentiation in P y T orch. In NIPS-W, 2017

  34. [34]

    C. R. Qi, L. Yi, H. Su, and L. J. Guibas. PointNet++ : Deep hierarchical feature learning on point sets in a metric space. In NIPS, 2017

  35. [35]

    Ramakrishnan, P

    R. Ramakrishnan, P. O. Dral, M. Rupp, and O. A. von Lilienfeld. Quantum chemistry structures and properties of 134 kilo molecules. Scientific Data, 2014

  36. [36]

    Ranjan, T

    A. Ranjan, T. Bolkart, S. Sanyal, and M. J. Black. Generating 3D faces using convolutional mesh autoencoders. In ECCV, 2018

  37. [37]

    M. S. Schlichtkrull, T. N. Kipf, P. Bloem, R. van den Berg , I. Titov, and M. Welling. Modeling relational data with graph convolutional networks. In ESWC, 2018

  38. [38]

    G. Sen, G. Namata, M. Bilgic, and L. Getoor. Collective classification in network data. AI Magazine, 29, 2008

  39. [39]

    Shchur, M

    O. Shchur, M. Mumme, A. Bojchevski, and S. G \"u nnemann. Pitfalls of graph neural network evaluation. In NeurIPS-W, 2018

  40. [40]

    Simonovsky and N

    M. Simonovsky and N. Komodakis. Dynamic edge-conditioned filters in convolutional neural networks on graphs. In CVPR, 2017

  41. [41]

    K. K. Thekumparampil, C. Wang, S. Oh, and L. Li. Attention-based graph neural network for semi-supervised learning. CoRR, abs/1803.03735, 2018

  42. [42]

    Veli c kovi \' c , G

    P. Veli c kovi \' c , G. Cucurull, A. Casanova, A. Romero, P. Li \` o , and Y. Bengio. Graph attention networks. In ICLR, 2018

  43. [43]

    Veli c kovi \' c , W

    P. Veli c kovi \' c , W. Fedus, W. L. Hamilton, P. Li \` o , Y. Bengio, and R. D. Hjeml. Deep graph infomax. In ICLR, 2019

  44. [44]

    Vinyals, S

    O. Vinyals, S. Bengio, and M. Kudlur. Order matters: Sequence to sequence for sets. In ICLR, 2016

  45. [45]

    M. Wang, L. Yu, A. Gan, D. Zheng, Y. Gai, Z. Ye, M. Li, J. Zhou, Q. Huang, J. Zhao, H. Lin, C. Ma, D. Deng, Q. Guo, H. Zhang, J. Li, A. J. Smola, and Z. Zhang. Deep graph library. http://dgl.ai, 2018 a

  46. [46]

    Y. Wang, Y. Sun, Z. Liu, S. E. Sarma, M. M. Bronstein, and J. M. Solomon. Dynamic graph CNN for learning on point clouds. CoRR, abs/1801.07829, 2018 b

  47. [47]

    F. Wu, T. Zhang, A. H. de Souza Jr., C. Fifty, T. Yu, and K. Q. Weinberger. Simplifying graph convolutional networks. CoRR, abs/1902.07153, 2019

  48. [48]

    Z. Wu, S. Song, A. Khosla, F. Yu, L. Zhang, X. Tang, and J. Xiao. 3D ShapeNets : A deep representation for volumetric shapes. In CVPR, 2015

  49. [49]

    K. Xu, C. Li, Y. Tian, T. Sonobe, K. Kawarabayashi, and S. Jegelka. Representation learning on graphs with jumping knowledge networks. In ICML, 2018

  50. [50]

    K. Xu, W. Hu, J. Leskovec, and S. Jegelka. How powerful are graph neural networks? In ICLR, 2019

  51. [51]

    R. Ying, J. You, C. Morris, X. Ren, W. Hamilton, and J. Leskovec. Hierarchical graph representation learning with differentiable pooling. In NeurIPS, 2018

  52. [52]

    Zhang, Z

    M. Zhang, Z. Cui, M. Neumann, and Y. Chen. An end-to-end deep learning architecture for graph classification. In AAAI, 2018