Recognition: unknown
TACENR: Task-Agnostic Contrastive Explanations for Node Representations
Pith reviewed 2026-05-10 03:22 UTC · model grok-4.3
The pith
A contrastive learning method identifies the attribute, proximity, and structural features that shape node representations in graphs, even without a downstream task.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
TACENR learns a similarity function in the representation space via contrastive learning to rank the features most responsible for a node's embedding; the method operates task-agnostically but extends to supervised use, and experiments confirm that proximity and structural features play a major role while the supervised version performs on par with task-specific baselines.
What carries the argument
The contrastive similarity function that scores how much each candidate feature contributes to similarity between a node and its positive or negative examples in the learned representation space.
If this is right
- Proximity and structural features contribute more to node representations than previously emphasized in explanation work.
- Task-agnostic explanations can be obtained without retraining or access to downstream labels.
- The same contrastive machinery yields explanations comparable to supervised methods when labels become available.
- Explanations can address the full structure of a representation rather than isolated dimensions.
Where Pith is reading between the lines
- If the identified features prove stable across random seeds or model architectures, they could serve as a diagnostic for whether a graph model has captured intended structural signals.
- The approach might extend to other embedding techniques such as knowledge-graph or point-cloud representations by swapping the contrastive objective.
- Identifying dominant proximity features could guide data collection priorities, such as which neighbor attributes to measure more accurately.
Load-bearing premise
The similarity function learned by contrastive training isolates the features that truly determine the node representation rather than merely correlating with it.
What would settle it
Perturb the top-ranked features identified by TACENR for a node and observe whether the representation vector changes substantially less than when perturbing lower-ranked features.
Figures
read the original abstract
Graph representation learning has achieved notable success in encoding graph-structured data into latent vector spaces, enabling a wide range of downstream tasks. However, these node representations remain opaque and difficult to interpret. Existing explainability methods primarily focus on supervised settings or on explaining individual representation dimensions, leaving a critical gap in explaining the overall structure of node representations. In this paper, we propose TACENR (Task-Agnostic Contrastive Explanations for Node Representations), a local explanation method that identifies not only attribute features but also proximity and structural ones that contribute the most in the representation space. TACENR builds on contrastive learning, through which we learn a similarity function in the representation space, revealing which are the features that play an important role in the representation of a node. While our focus is on task-agnostic explanations, TACENR can be applied to supervised scenarios as well. Experimental results demonstrate that proximity and structural features play a significant role in shaping node representations and that our supervised variant performs comparably to existing task-specific approaches in identifying the most impactful features.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The paper introduces TACENR, a local explanation method for node representations learned by graph neural networks. It employs contrastive learning to train a similarity function over representations, thereby identifying the most important attribute, proximity, and structural features that shape each node's embedding. The work emphasizes a task-agnostic setting while also presenting a supervised variant; the central empirical claims are that proximity and structural features play a significant role and that the supervised variant matches existing task-specific explainers in identifying impactful features.
Significance. A reliable task-agnostic explainer that surfaces proximity and structural contributions would address a genuine gap in GNN interpretability. If the contrastive similarity function can be shown to isolate mechanistically relevant features rather than mere correlations, the approach could be adopted for post-hoc analysis of pretrained embeddings across multiple downstream tasks.
major comments (2)
- [Abstract and §4] Abstract and §4 (Experiments): the claim that 'proximity and structural features play a significant role' and that the supervised variant 'performs comparably' is asserted without any reported datasets, metrics, baselines, or quantitative results. Because these statements constitute the primary empirical support for the method's utility, the absence of concrete evidence makes it impossible to evaluate whether the data actually substantiate the central claims.
- [§3] §3 (Method): the contrastive objective is presented as revealing 'which are the features that play an important role in the representation of a node.' This interpretation assumes that the learned similarity distinguishes causal drivers from statistical associations induced by message-passing biases. The paper does not provide a diagnostic (e.g., controlled synthetic graphs where ground-truth causal features are known) to test whether positive/negative pair construction isolates mechanistic contributions or merely co-varying ones.
minor comments (2)
- [§3] Notation for the similarity function and the feature attribution scores should be introduced once and used consistently; currently the abstract and method description employ slightly different phrasing for the same quantities.
- [§2] The manuscript would benefit from a short related-work subsection that explicitly contrasts TACENR with existing GNN explanation techniques that also incorporate structural or proximity information.
Simulated Author's Rebuttal
We thank the referee for their thoughtful comments on our work. We provide detailed responses to each major comment and indicate the revisions we will make to strengthen the manuscript.
read point-by-point responses
-
Referee: [Abstract and §4] Abstract and §4 (Experiments): the claim that 'proximity and structural features play a significant role' and that the supervised variant 'performs comparably' is asserted without any reported datasets, metrics, baselines, or quantitative results. Because these statements constitute the primary empirical support for the method's utility, the absence of concrete evidence makes it impossible to evaluate whether the data actually substantiate the central claims.
Authors: We acknowledge the referee's point that the abstract and §4 could benefit from more explicit reporting of the experimental details to substantiate the claims. We will revise the abstract to briefly describe the datasets, metrics, and key quantitative findings. In §4, we will add or clarify the presentation of results, including specific datasets used, the metrics for evaluating feature importance, the baselines compared, and the numerical results demonstrating the significant role of proximity and structural features as well as the performance of the supervised variant. revision: yes
-
Referee: [§3] §3 (Method): the contrastive objective is presented as revealing 'which are the features that play an important role in the representation of a node.' This interpretation assumes that the learned similarity distinguishes causal drivers from statistical associations induced by message-passing biases. The paper does not provide a diagnostic (e.g., controlled synthetic graphs where ground-truth causal features are known) to test whether positive/negative pair construction isolates mechanistic contributions or merely co-varying ones.
Authors: This is a valid concern regarding the interpretation of the contrastive explanations. Our method identifies features that contribute to the similarity in the learned representation space, which reflects the GNN's encoding process. To address the distinction between causal and correlational features, we will incorporate experiments on synthetic graphs with known ground-truth structural and attribute influences (e.g., graphs generated with specific motifs or feature dependencies). These will serve as a diagnostic to show that the positive and negative pair construction recovers the intended influential features rather than spurious correlations. revision: yes
Circularity Check
No significant circularity; method relies on standard contrastive learning without self-referential derivations
full rationale
The paper proposes TACENR using contrastive learning to learn a similarity function that identifies impactful features (attributes, proximity, structural) in node representations. No mathematical derivations, equations, or prediction steps are described that reduce by construction to fitted inputs or self-citations. Claims rest on experimental results comparing to baselines, with the core approach being a direct application of contrastive objectives rather than a tautological chain. The task-agnostic and supervised variants are presented as extensions of existing techniques without load-bearing self-references or uniqueness theorems imported from prior author work.
Axiom & Free-Parameter Ledger
Reference graph
Works this paper leans on
-
[1]
IEEE TKDE (2020)
Ahmed, N.K., Rossi, R.A., Lee, J.B., Willke, T.L., Zhou, R., Kong, X., Eldardiry, H.: Role-based graph embeddings. IEEE TKDE (2020)
2020
- [2]
-
[3]
arXiv preprint arXiv:1905.13686 (2019)
Baldassarre, F., Azizpour, H.: Explainability techniques for gcns. arXiv:1905.13686 (2019) TACENR: Task-Agnostic Contrastive Explanations for Node Representations 21
-
[4]
In: ECML-PKDD (2021)
Duval, A., Malliaros, F.D.: Graphsvx: Shapley value explanations for gnns. In: ECML-PKDD (2021)
2021
-
[5]
In: WWW (2019)
Fan, W., Ma, Y., Li, Q., He, Y., Zhao, E., Tang, J., Yin, D.: Graph neural networks for social recommendation. In: WWW (2019)
2019
-
[6]
Fast Graph Representation Learning with PyTorch Geometric
Fey, M., Lenssen, J.E.: Fast graph representation learning with pytorch geometric. arXiv:1903.02428 (2019)
work page internal anchor Pith review arXiv 1903
-
[7]
NeurIPS (2017)
Fout, A., Byrd, J., Shariat, B., Ben-Hur, A.: Protein interface prediction using graph convolutional networks. NeurIPS (2017)
2017
-
[8]
IEEE TKDE (2022)
Funke, T., Khosla, M., Rathee, M., Anand, A.: Zorro: Valid, sparse, and stable explanations in gnns. IEEE TKDE (2022)
2022
-
[9]
In: ACM DL (1998)
Giles, C.L., Bollacker, K.D., Lawrence, S.: Citeseer: An automatic citation indexing system. In: ACM DL (1998)
1998
-
[10]
Giorgi, F., Silvestri, F., Tolomei, G.: Combinex: Counterfactual explanations via feature and structural perturbations. arXiv:2502.10111 (2025)
-
[11]
In: KDD (2016)
Grover, A., Leskovec, J.: node2vec: Scalable feature learning for networks. In: KDD (2016)
2016
-
[12]
In: UAI (2011)
Gu, Q., Li, Z., Han, J.: Generalized fisher score for feature selection. In: UAI (2011)
2011
-
[13]
NeurIPS (2017)
Hamilton, W., Ying, Z., Leskovec, J.: Inductive representation learning on large graphs. NeurIPS (2017)
2017
-
[14]
In: KDD (2019)
Han, P., Yang, P., Zhao, P., Shang, S., Liu, Y., Zhou, J., Gao, X., Kalnis, P.: Gcn- mf: disease-gene association identification by graph convolutional networks and matrix factorization. In: KDD (2019)
2019
-
[15]
IEEE TKDE (2022)
Huang, Q., Yamada, M., Tian, Y., Singh, D., Chang, Y.: Graphlime: Local inter- pretable model explanations for gnns. IEEE TKDE (2022)
2022
-
[16]
In: ICLR (2024)
Kang, H., Han, G., Park, H.: Unr-explainer: Counterfactual explanations for un- supervised gnns. In: ICLR (2024)
2024
-
[17]
Semi-Supervised Classification with Graph Convolutional Networks
Kipf, T.N., Welling, M.: Semi-supervised classification with gcns. arXiv:1609.02907 (2016)
work page internal anchor Pith review arXiv 2016
-
[18]
NeurIPS (2020)
Luo, D., Cheng, W., Xu, D., Yu, W., Zong, B., Chen, H., Zhang, X.: Parameterized explainer for graph neural networks. NeurIPS (2020)
2020
-
[19]
McCallum,A.K.,Nigam,K.,Rennie,J.,Seymore,K.:Automatingtheconstruction of internet portals with machine learning. Inf. Retrieval (2000)
2000
-
[20]
Neural Comput
Nandan, M., Mitra, S., De, D.: Graphxai: Survey on explainable gnns. Neural Comput. Appl. (2025)
2025
-
[21]
What do gnns actually learn? towards understanding their representations, 2024
Nikolentzos, G., Chatzianastasis, M., Vazirgiannis, M.: What do gnns ac- tually learn? towards understanding their representations. arXiv preprint arXiv:2304.10851 (2023)
-
[22]
JMLR (2011)
Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., Blondel, M., Prettenhofer, P., Weiss, R., Dubourg, V., et al.: Scikit-learn: Machine learning in python. JMLR (2011)
2011
-
[23]
IEEE TKDE (2024)
Piaggesi, S., Khosla, M., Panisson, A., Anand, A.: Dine: Dimensional interpretabil- ity of node embeddings. IEEE TKDE (2024)
2024
-
[24]
In: CVPR (2019)
Pope, P.E., Kolouri, S., Rostami, M., Martin, C.E., Hoffmann, H.: Explainability methods for graph cnns. In: CVPR (2019)
2019
-
[25]
why should i trust you?
Ribeiro, M.T., Singh, S., Guestrin, C.: “why should i trust you?” explaining the predictions of any classifier. In: KDD (2016)
2016
-
[26]
Schlichtkrull, M.S., De Cao, N., Titov, I.: Interpreting graph neural networks for nlp with differentiable edge masking. arXiv:2010.00577 (2020)
-
[27]
IEEE TPAMI (2021) 22 V
Schnake, T., Eberle, O., Lederer, J., Nakajima, S., Schütt, K.T., Müller, K.R., Montavon, G.: Higher-order explanations of gnns via relevant walks. IEEE TPAMI (2021) 22 V. Papanikou and E. Pitoura
2021
-
[28]
Schwarzenberg, R., Hübner, M., Harbecke, D., Alt, C., Hennig, L.: Layerwise rele- vance visualization in convolutional text graph classifiers. arXiv:1909.10911 (2019)
-
[29]
Sen, P., Namata, G., Bilgic, M., Getoor, L., Galligher, B., Eliassi-Rad, T.: Collec- tive classification in network data. AI Mag. (2008)
2008
-
[30]
Shafi, Z., Chatterjee, A., Eliassi-Rad, T.: Generating human understandable ex- planations for node embeddings. arXiv:2406.07642 (2024)
-
[31]
Velickovic, P., Cucurull, G., Casanova, A., Romero, A., Lio, P., Bengio, Y.: Graph attention networks. arXiv:1710.10903 (2017)
work page internal anchor Pith review arXiv 2017
-
[32]
NeurIPS (2020)
Vu, M., Thai, M.T.: Pgm-explainer: Probabilistic graphical model explanations for gnns. NeurIPS (2020)
2020
-
[33]
NeurIPS (2022)
Xie, Y., Katariya, S., Tang, X., Huang, E., Rao, N., Subbian, K., Ji, S.: Task- agnostic graph explanations. NeurIPS (2022)
2022
-
[34]
Neural computation (2014)
Yamada, M., Jitkrittum, W., Sigal, L., Xing, E.P., Sugiyama, M.: High-dimensional feature selection by feature-wise kernelized lasso. Neural computation (2014)
2014
-
[35]
In: 2020 IEEE international conference on data mining (ICDM)
Yang, L., Gu, J., Wang, C., Cao, X., Zhai, L., Jin, D., Guo, Y.: Toward unsu- pervised graph neural network: Interactive clustering and embedding via optimal transport. In: 2020 IEEE international conference on data mining (ICDM). pp. 1358–1363. IEEE (2020)
2020
-
[36]
NeurIPS (2019)
Ying, Z., Bourgeois, D., You, J., Zitnik, M., Leskovec, J.: Gnnexplainer: Generating explanations for graph neural networks. NeurIPS (2019)
2019
-
[37]
IEEE TPAMI (2022)
Yuan, H., Yu, H., Gui, S., Ji, S.: Explainability in graph neural networks: A taxo- nomic survey. IEEE TPAMI (2022)
2022
-
[38]
In: AIES (2021)
Zhang, Y., Defazio, D., Ramesh, A.: Relex: A model-agnostic relational model explainer. In: AIES (2021)
2021
-
[39]
Bioinformatics (2017) APPENDIX This appendix provides additional details on our experimental configuration and further evaluations of the weighting-based similarity methods
Zitnik, M., Leskovec, J.: Predicting multicellular function through multi-layer tis- sue networks. Bioinformatics (2017) APPENDIX This appendix provides additional details on our experimental configuration and further evaluations of the weighting-based similarity methods. A Experimental Setup All experiments are conducted on a Linux machine with an NVIDIA...
2017
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.