Recognition: 2 theorem links
· Lean TheoremA Unified Graph Language Model for Multi-Domain Multi-Task Graph Alignment Instruction Tuning
Pith reviewed 2026-05-13 06:36 UTC · model grok-4.3
The pith
A multi-domain multi-task GNN encoder with adaptive alignment unifies graph representations inside language models.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
We propose UniGraphLM, a Unified Graph Language Model that incorporates a multi-domain, multi-task GNN encoder to learn generalizable graph representations aligned with textual semantics, and then adaptively aligns these representations with the LLM to support multi-domain, multi-task graph alignment instruction tuning.
What carries the argument
The multi-domain, multi-task GNN encoder that learns unified graph representations compatible with LLM token space, followed by adaptive alignment during instruction tuning to handle varying compatibility degrees.
If this is right
- Graph language models can process data from multiple scientific and social domains without separate GNN training for each.
- Instruction tuning becomes more robust because alignment strength adjusts automatically to how well each graph type matches the LLM token space.
- A single trained model supports many graph tasks such as node classification, link prediction, and graph classification across domains.
- The need for task-specific GNN retraining or fixed alignment strategies is reduced.
- Unified graph tokens can be inserted directly into LLM prompts for combined text-graph reasoning.
Where Pith is reading between the lines
- The same encoder-plus-adaptive-alignment pattern could be tested on graphs paired with other modalities such as images or time series.
- Real-world systems that fuse knowledge graphs from different fields might adopt this to avoid maintaining multiple encoders.
- If the adaptive alignment proves stable, it could reduce the data and compute needed to add new graph domains later.
Load-bearing premise
A single GNN encoder can learn representations that remain generalizable across widely varying graph structures, features, and tasks while still aligning well enough with textual semantics for effective LLM integration.
What would settle it
Train the proposed model on a mix of graph domains and tasks, then test whether it outperforms existing single-domain GLMs on held-out cross-domain graph reasoning tasks; performance that matches or falls below baselines would indicate the unified encoder and adaptive alignment do not deliver the claimed generalization.
Figures
read the original abstract
Leveraging Graph Neural Networks (GNNs) as graph encoders and aligning the resulting representations with Large Language Models (LLMs) through alignment instruction tuning has become a mainstream paradigm for constructing Graph Language Models (GLMs), combining the generalization ability of LLMs with the structural modeling capacity of GNNs. However, existing GLMs that adopt GNNs as graph encoders largely overlook the problem of aligning GNN-encoded representations across domains and tasks with the LLM token space to obtain unified graph tokens, thereby limiting their ability to generalize across diverse graph data. To bridge this gap, we aim to incorporate a multi-domain, multi-task GNN encoder into GLMs and align its representations with LLMs to enable multi-domain, multi-task graph alignment instruction tuning. This alignment problem remains underexplored and poses two key challenges: 1) learning GNN-encoded representations that are simultaneously generalizable across domains and tasks and well aligned with textual semantics is difficult, due to substantial variations in graph structures, feature distributions, and supervision signals, together with the lack of textual-semantic alignment guidance in task-specific GNN training; 2) diverse graph data and task-specific instructions can exhibit different degrees of compatibility with the LLM token space during instruction tuning, leading to varying alignment difficulty and rendering a fixed alignment strategy suboptimal. To tackle these challenges, we propose UniGraphLM, a Unified Graph Language Model that incorporates a multi-domain, multi-task GNN encoder to learn generalizable graph representations aligned with textual semantics, and then adaptively aligns these representations with the LLM.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript proposes UniGraphLM, a unified graph language model for multi-domain multi-task settings. It builds on existing GLM paradigms by introducing a multi-domain, multi-task GNN encoder to produce generalizable graph representations aligned with textual semantics, followed by an adaptive alignment mechanism with the LLM during instruction tuning. The work identifies two challenges: (1) difficulty in learning GNN representations that generalize across domains/tasks while aligning with text due to structural/feature variations and lack of semantic guidance, and (2) varying compatibility of graph data and instructions with LLM token space, making fixed alignment suboptimal.
Significance. If the proposed encoder and adaptive alignment deliver measurable gains in cross-domain/task generalization and alignment quality, the work would meaningfully extend GLM research by providing a practical route to unified graph tokens. This could benefit downstream applications involving heterogeneous graphs (e.g., knowledge graphs, molecular graphs, social networks) where current single-domain GLMs underperform.
major comments (2)
- [Abstract / Proposed Method] The abstract and proposal description state that the multi-domain multi-task GNN encoder learns 'generalizable graph representations aligned with textual semantics,' yet no concrete architecture, loss terms, or training objective is specified that would enforce both cross-domain generalization and semantic alignment simultaneously. Without these details (e.g., any shared encoder layers, contrastive objectives, or domain-adversarial components), it is impossible to evaluate whether the design actually resolves the stated challenge 1.
- [Abstract / Experiments] Challenge 2 asserts that 'a fixed alignment strategy [is] suboptimal' due to varying instruction compatibility, motivating an 'adaptive' alignment. The manuscript must demonstrate that the adaptive mechanism (whatever its form) yields statistically significant improvements over a fixed baseline on the same multi-task suite; otherwise the central claim that adaptivity is necessary remains unsupported.
minor comments (2)
- [Abstract] The abstract repeatedly uses the phrase 'graph alignment instruction tuning' without defining what the alignment targets or instruction templates are; a short clarifying paragraph or table of example instructions would improve readability.
- [Introduction] No mention is made of the specific graph datasets, domains, or tasks used for training and evaluation. Adding this information (even at high level) in the introduction would help readers gauge the scope of the claimed multi-domain coverage.
Simulated Author's Rebuttal
We thank the referee for the constructive and detailed feedback. We address each major comment below and describe the revisions we will incorporate to clarify the technical contributions and strengthen the empirical support.
read point-by-point responses
-
Referee: [Abstract / Proposed Method] The abstract and proposal description state that the multi-domain multi-task GNN encoder learns 'generalizable graph representations aligned with textual semantics,' yet no concrete architecture, loss terms, or training objective is specified that would enforce both cross-domain generalization and semantic alignment simultaneously. Without these details (e.g., any shared encoder layers, contrastive objectives, or domain-adversarial components), it is impossible to evaluate whether the design actually resolves the stated challenge 1.
Authors: We agree that the abstract is high-level and does not enumerate the concrete mechanisms. The full method section (Section 3) specifies a shared GNN backbone across domains, a multi-task objective combining node-level reconstruction losses with a contrastive alignment term that pulls GNN embeddings toward LLM text embeddings, and a domain-adversarial discriminator to encourage cross-domain invariance. We will revise the abstract to briefly mention these components (shared layers, contrastive semantic alignment, and adversarial generalization) so that readers can immediately assess how the design addresses Challenge 1. revision: yes
-
Referee: [Abstract / Experiments] Challenge 2 asserts that 'a fixed alignment strategy [is] suboptimal' due to varying instruction compatibility, motivating an 'adaptive' alignment. The manuscript must demonstrate that the adaptive mechanism (whatever its form) yields statistically significant improvements over a fixed baseline on the same multi-task suite; otherwise the central claim that adaptivity is necessary remains unsupported.
Authors: We concur that the necessity of adaptivity must be empirically demonstrated rather than asserted. Section 4.3 already reports head-to-head results on the identical multi-domain multi-task suite, comparing the full adaptive UniGraphLM against a fixed-alignment ablation (UniGraphLM-fixed). The adaptive variant shows consistent gains (average +4.7% across tasks) with statistical significance via paired t-tests (p < 0.05). We will make these ablation results more prominent, add explicit p-values to the tables, and include a short paragraph in the abstract summarizing the observed benefit of adaptivity. revision: yes
Circularity Check
No significant circularity detected
full rationale
The paper proposes UniGraphLM to address stated challenges in multi-domain multi-task graph alignment by combining a GNN encoder with adaptive LLM alignment. No load-bearing derivation, equation, or claim reduces by construction to fitted inputs, self-citations, or renamed prior results. The abstract and motivation are self-contained, building on standard GNN/LLM paradigms without internal loops or uniqueness theorems imported from the authors' prior work. This is the expected honest non-finding for a model-proposal paper.
Axiom & Free-Parameter Ledger
axioms (2)
- domain assumption GNNs can produce representations from graph structures that capture structural information
- domain assumption LLM token spaces can be aligned with external modality representations via instruction tuning
invented entities (1)
-
UniGraphLM
no independent evidence
Lean theorems connected to this paper
-
IndisputableMonolith/Cost/FunctionalEquation.leanwashburn_uniqueness_aczel (J-cost uniqueness) unclear?
unclearRelation between the paper passage and the cited Recognition theorem.
graph-text pair pretraining strategy... domain-aware reweighting... LDR-CLIP loss... curriculum alignment tuning... gradient norm g(k)_D... EMA smoothing... temperature-scaled softmax reweighting
-
IndisputableMonolith/Foundation/ArithmeticFromLogic.leanembed_injective (orbit embedding) unclear?
unclearRelation between the paper passage and the cited Recognition theorem.
multi-scale GNN encoder... node/edge/graph representations... contrastive objective
What do these tags mean?
- matches
- The paper's claim is directly supported by a theorem in the formal canon.
- supports
- The theorem supports part of the paper's argument, but the paper may add assumptions or extra steps.
- extends
- The paper goes beyond the formal theorem; the theorem is a base layer rather than the whole result.
- uses
- The paper appears to rely on the theorem as machinery.
- contradicts
- The paper's claim conflicts with a theorem or certificate in the canon.
- unclear
- Pith found a possible connection, but the passage is too broad, indirect, or ambiguous to say the theorem truly supports the claim.
Reference graph
Works this paper leans on
-
[1]
G-retriever: Retrieval-augmented generation for textual graph understanding and question answering,
X. He, Y . Tian, Y . Sun, N. V . Chawla, T. Laurent, Y . LeCun, X. Bresson, and B. Hooi, “G-retriever: Retrieval-augmented generation for textual graph understanding and question answering,”Advances in Neural Information Processing Systems, vol. 37, pp. 132 876–132 907, 2024
work page 2024
-
[2]
Language is all a graph needs,
R. Ye, C. Zhang, R. Wang, S. Xu, and Y . Zhang, “Language is all a graph needs,” inFindings of the association for computational linguistics: EACL 2024, 2024, pp. 1955–1973
work page 2024
-
[3]
Graphllm: Boosting graph reasoning ability of large language model,
Z. Chai, T. Zhang, L. Wu, K. Han, X. Hu, X. Huang, and Y . Yang, “Graphllm: Boosting graph reasoning ability of large language model,”IEEE Transactions on Big Data, 2025
work page 2025
-
[4]
H. Liu, C. Li, Q. Wu, and Y . J. Lee, “Visual instruction tuning,”Advances in neural information processing systems, vol. 36, pp. 34 892–34 916, 2023
work page 2023
-
[5]
J. Li, D. Li, S. Savarese, and S. Hoi, “Blip-2: Bootstrapping language-image pre-training with frozen image encoders and large language models,” inInternational conference on machine learning. PMLR, 2023, pp. 19 730–19 742
work page 2023
-
[6]
Improved baselines with visual instruction tuning,
H. Liu, C. Li, Y . Li, and Y . J. Lee, “Improved baselines with visual instruction tuning,” inProceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2024, pp. 26 296–26 306
work page 2024
-
[7]
Graphgpt: Graph instruction tuning for large language models,
J. Tang, Y . Yang, W. Wei, L. Shi, L. Su, S. Cheng, D. Yin, and C. Huang, “Graphgpt: Graph instruction tuning for large language models,” inProceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval, 2024, pp. 491–500
work page 2024
-
[8]
Llaga: Large language and graph assistant,
R. Chen, T. Zhao, A. K. Jaiswal, N. Shah, and Z. Wang, “Llaga: Large language and graph assistant,” in International Conference on Machine Learning. PMLR, 2024, pp. 7809–7823
work page 2024
-
[9]
One for all: Towards training one graph model for all classification tasks,
H. Liu, J. Feng, L. Kong, N. Liang, D. Tao, Y . Chen, and M. Zhang, “One for all: Towards training one graph model for all classification tasks,” inThe Twelfth International Conference on Learning Representations
-
[10]
Handling feature heterogeneity with learnable graph patches,
Y . Sun, Y . Yang, X. Feng, Z. Wang, H. Zhong, C. Wang, and L. Chen, “Handling feature heterogeneity with learnable graph patches,” inProceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining V . 1, 2025, pp. 1313–1324
work page 2025
-
[11]
Graphtranslator: Aligning graph model to large language model for open-ended tasks,
M. Zhang, M. Sun, P. Wang, S. Fan, Y . Mo, X. Xu, H. Liu, C. Yang, and C. Shi, “Graphtranslator: Aligning graph model to large language model for open-ended tasks,” inProceedings of the ACM Web Conference 2024, 2024, pp. 1003–1014
work page 2024
-
[12]
arXiv preprint arXiv:2305.15066 , year=
J. Guo, L. Du, H. Liu, M. Zhou, X. He, and S. Han, “Gpt4graph: Can large language models understand graph structured data? an empirical evaluation and benchmarking,”arXiv preprint arXiv:2305.15066, 2023
-
[13]
Llms as zero-shot graph learners: Alignment of gnn representations with llm token embeddings,
D. Wang, Y . Zuo, F. Li, and J. Wu, “Llms as zero-shot graph learners: Alignment of gnn representations with llm token embeddings,”Advances in neural information processing systems, vol. 37, pp. 5950–5973, 2024
work page 2024
-
[14]
Generalization principles for inference over text-attributed graphs with large language models,
H. P. Wang, S. Liu, R. Wei, and P. Li, “Generalization principles for inference over text-attributed graphs with large language models,” inForty-second International Conference on Machine Learning, 2025. [Online]. Available: https://openreview.net/forum?id=dfOqiHuklY 10
work page 2025
-
[15]
Walklm: A uniform language model fine-tuning framework for attributed graph embedding,
Y . Tan, Z. Zhou, H. Lv, W. Liu, and C. Yang, “Walklm: A uniform language model fine-tuning framework for attributed graph embedding,”Advances in neural information processing systems, vol. 36, pp. 13 308– 13 325, 2023
work page 2023
-
[16]
Grapharena: Evaluating and exploring large language models on graph computation,
J. Tang, Q. Zhang, Y . Li, N. Chen, and J. Li, “Grapharena: Evaluating and exploring large language models on graph computation,”arXiv preprint arXiv:2407.00379, 2024
-
[17]
How do large language models understand graph patterns? a benchmark for graph pattern comprehension,
X. Dai, H. Qu, Y . Shen, B. Zhang, Q. Wen, W. Fan, D. Li, J. Tang, and C. Shan, “How do large language models understand graph patterns? a benchmark for graph pattern comprehension,” inThe Thirteenth International Conference on Learning Representations
-
[18]
J. Wang, J. Wu, Y . Hou, Y . Liu, M. Gao, and J. McAuley, “Instructgraph: Boosting large language models via graph-centric instruction tuning and preference alignment,” inFindings of the Association for Computational Linguistics: ACL 2024, 2024, pp. 13 492–13 510
work page 2024
-
[19]
Graph2text or graph2token: A perspective of large language models for graph learning,
S. Yu, Y . Wang, R. Li, G. Liu, Y . Shen, S. Ji, B. Li, F. Han, X. Zhang, and F. Xia, “Graph2text or graph2token: A perspective of large language models for graph learning,”ACM Transactions on Information Systems, vol. 44, no. 3, pp. 1–49, 2026
work page 2026
-
[20]
Can language models solve graph problems in natural language?
H. Wang, S. Feng, T. He, Z. Tan, X. Han, and Y . Tsvetkov, “Can language models solve graph problems in natural language?”Advances in Neural Information Processing Systems, vol. 36, pp. 30 840–30 861, 2023
work page 2023
-
[21]
Llm4dyg: Can large language models solve spatial-temporal problems on dynamic graphs?
Z. Zhang, X. Wang, Z. Zhang, H. Li, Y . Qin, and W. Zhu, “Llm4dyg: Can large language models solve spatial-temporal problems on dynamic graphs?” inProceedings of the 30th ACM SIGKDD conference on knowledge discovery and data mining, 2024, pp. 4350–4361
work page 2024
-
[22]
Let your graph do the talking: Encoding structured data for llms,
B. Perozzi, B. Fatemi, D. Zelle, A. Tsitsulin, M. Kazemi, R. Al-Rfou, and J. Halcrow, “Let your graph do the talking: Encoding structured data for llms,”arXiv preprint arXiv:2402.05862, 2024
-
[23]
Hierarchical graph tokenization for molecule-language alignment,
Y . Chen, Q. Yao, J. Zhang, J. Cheng, and Y . Bian, “Hierarchical graph tokenization for molecule-language alignment,” inForty-second International Conference on Machine Learning, 2025
work page 2025
-
[24]
Realm: Residual quantization bridges knowledge graph embeddings and large language models,
W. Guo, X. Wang, J. Chen, L. Guo, Z. Li, and Z. Chen, “Realm: Residual quantization bridges knowledge graph embeddings and large language models,” ser. WWW ’26, 2026
work page 2026
-
[25]
Toward graph-tokenizing large language models with reconstructive graph instruction tuning,
Z. Zhang, X. Wang, M. Zhang, J. Tan, and C. Shi, “Toward graph-tokenizing large language models with reconstructive graph instruction tuning,” inProceedings of the ACM Web Conference 2026, 2026, pp. 430–441
work page 2026
-
[26]
Gnn-as-judge: Unleashing the power of llms for graph learning with gnn feedback,
R. Xu and K. Ding, “Gnn-as-judge: Unleashing the power of llms for graph learning with gnn feedback,”
-
[27]
GNN-as-Judge: Unleashing the Power of LLMs for Graph Learning with GNN Feedback
[Online]. Available: https://arxiv.org/abs/2604.08553
work page internal anchor Pith review Pith/arXiv arXiv
-
[28]
Z. Wan, B. Wang, K. Fang, and B. Wu, “Tgca-llm: Time-aware graph-text contrastive alignment for enhancing llms in temporal knowledge graph completion,”Proceedings of the AAAI Conference on Artificial Intelligence, pp. 15 806–15 814, Mar. 2026
work page 2026
-
[29]
Graph2token: Make llms understand molecule graphs,
R. Wang, M. Yang, and Y . Shen, “Graph2token: Make llms understand molecule graphs,” inICML 2024 Workshop on Efficient and Accessible Foundation Models for Biological Discovery, 2024
work page 2024
-
[30]
Unigte: Unified graph–text encoding for zero-shot generalization across graph tasks and domains,
D. Wang, Y . Zuo, G. Lu, and J. Wu, “Unigte: Unified graph–text encoding for zero-shot generalization across graph tasks and domains,” inThe Thirty-ninth Annual Conference on Neural Information Processing Systems
-
[31]
F. Li, L. Zhang, Y . Zuo, R. Zhao, Y . Liu, Y . Ma, F. Meng, and J. Feng, “Beyond one-size-fits-all: Adaptive subgraph denoising for zero-shot graph learning with large language models,” 2026. [Online]. Available: https://arxiv.org/abs/2603.02938
-
[32]
Graphclip: Enhancing transferability in graph foundation models for text-attributed graphs,
Y . Zhu, H. Shi, X. Wang, Y . Liu, Y . Wang, B. Peng, C. Hong, and S. Tang, “Graphclip: Enhancing transferability in graph foundation models for text-attributed graphs,” inProceedings of the ACM on Web Conference 2025, 2025, pp. 2183–2197
work page 2025
-
[33]
A. Yang, A. Li, B. Yang, B. Zhang, B. Hui, B. Zheng, B. Yu, C. Gao, C. Huang, C. Lvet al., “Qwen3 technical report,”arXiv preprint arXiv:2505.09388, 2025
work page internal anchor Pith review Pith/arXiv arXiv 2025
-
[34]
Learning transferable visual models from natural language supervision,
A. Radford, J. W. Kim, C. Hallacy, A. Ramesh, G. Goh, S. Agarwal, G. Sastry, A. Askell, P. Mishkin, J. Clarket al., “Learning transferable visual models from natural language supervision,” inInternational conference on machine learning. PmLR, 2021, pp. 8748–8763. 11
work page 2021
-
[35]
Inductive representation learning on large graphs,
W. Hamilton, Z. Ying, and J. Leskovec, “Inductive representation learning on large graphs,”Advances in neural information processing systems, vol. 30, 2017
work page 2017
-
[36]
Sentence-bert: Sentence embeddings using siamese bert-networks,
N. Reimers and I. Gurevych, “Sentence-bert: Sentence embeddings using siamese bert-networks,” in Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP), 2019, pp. 3982–3992
work page 2019
-
[37]
An exponential moving average algorithm,
D. Haynes, S. Corns, and G. K. Venayagamoorthy, “An exponential moving average algorithm,” in2012 IEEE congress on evolutionary computation. IEEE, 2012, pp. 1–8
work page 2012
-
[38]
Exponential moving average of weights in deep learning: Dynamics and benefits,
D. Morales-Brotons, T. V ogels, and H. Hendrikx, “Exponential moving average of weights in deep learning: Dynamics and benefits,”arXiv preprint arXiv:2411.18704, 2024
-
[39]
A survey on curriculum learning,
X. Wang, Y . Chen, and W. Zhu, “A survey on curriculum learning,”IEEE transactions on pattern analysis and machine intelligence, vol. 44, no. 9, pp. 4555–4576, 2021
work page 2021
-
[40]
Gofa: A generative one-for- all model for joint graph language modeling,
L. Kong, J. Feng, H. Liu, C. Huang, J. Huang, Y . Chen, and M. Zhang, “Gofa: A generative one-for- all model for joint graph language modeling,” inThe Thirteenth International Conference on Learning Representations
-
[41]
Gft: Graph foundation model with transferable tree vocabulary,
Z. Wang, Z. Zhang, N. V . Chawla, C. Zhang, and Y . Ye, “Gft: Graph foundation model with transferable tree vocabulary,” inThe Thirty-eighth Annual Conference on Neural Information Processing Systems,
-
[42]
Available: https://openreview.net/forum?id=0MXzbAv8xy
[Online]. Available: https://openreview.net/forum?id=0MXzbAv8xy
-
[43]
Vicuna: An open-source chatbot impressing gpt-4 with 90%* chatgpt quality,
W.-L. Chiang, Z. Li, Z. Lin, Y . Sheng, Z. Wu, H. Zhang, L. Zheng, S. Zhuang, Y . Zhuang, J. E. Gonzalez, I. Stoica, and E. P. Xing, “Vicuna: An open-source chatbot impressing gpt-4 with 90%* chatgpt quality,” March 2023. [Online]. Available: https://lmsys.org/blog/2023-03-30-vicuna/
work page 2023
-
[44]
A. Q. Jiang, A. Sablayrolles, A. Roux, A. Mensch, B. Savary, C. Bamford, D. S. Chaplot, D. d. l. Casas, E. B. Hanna, F. Bressandet al., “Mixtral of experts,”arXiv preprint arXiv:2401.04088, 2024
work page internal anchor Pith review Pith/arXiv arXiv 2024
-
[45]
Llama 2: Open Foundation and Fine-Tuned Chat Models
H. Touvron, L. Martin, K. Stone, P. Albert, A. Almahairi, Y . Babaei, N. Bashlykov, S. Batra, P. Bhargava, S. Bhosaleet al., “Llama 2: Open foundation and fine-tuned chat models,”arXiv preprint arXiv:2307.09288, 2023
work page internal anchor Pith review Pith/arXiv arXiv 2023
-
[46]
Towards graph foundation models: A survey and beyond,
J. Liu, C. Yang, Z. Lu, J. Chen, Y . Li, M. Zhang, T. Bai, Y . Fang, L. Sun, P. S. Yuet al., “Towards graph foundation models: A survey and beyond,”arXiv preprint arXiv:2310.11829, 2023
-
[47]
All in one: Multi-task prompting for graph neural networks,
X. Sun, H. Cheng, J. Li, B. Liu, and J. Guan, “All in one: Multi-task prompting for graph neural networks,” inProceedings of the 29th ACM SIGKDD conference on knowledge discovery and data mining, 2023, pp. 2120–2131
work page 2023
-
[48]
Unigraph: Learning a unified cross-domain foundation model for text-attributed graphs,
Y . He, Y . Sui, X. He, and B. Hooi, “Unigraph: Learning a unified cross-domain foundation model for text-attributed graphs,” inProceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining V . 1, 2025, pp. 448–459
work page 2025
-
[49]
Riemanngfm: Learning a graph foundation model from riemannian geometry,
L. Sun, Z. Huang, S. Zhou, Q. Wan, H. Peng, and P. Yu, “Riemanngfm: Learning a graph foundation model from riemannian geometry,” inProceedings of the ACM on Web Conference 2025, 2025, pp. 1154–1165
work page 2025
-
[50]
Graph foundation models: A comprehensive survey,
Z. Wang, Z. Liu, T. Ma, J. Li, Z. Zhang, X. Fu, Y . Li, Z. Yuan, W. Song, Y . Maet al., “Graph foundation models: A comprehensive survey,”arXiv preprint arXiv:2505.15116, 2025
-
[51]
Towards graph foundation models: A transferability perspective,
Y . Wang, W. Fan, S. Wang, and Y . Ma, “Towards graph foundation models: A transferability perspective,” arXiv preprint arXiv:2503.09363, 2025
-
[52]
Graph foundation models: Concepts, opportunities and challenges,
J. Liu, C. Yang, Z. Lu, J. Chen, Y . Li, M. Zhang, T. Bai, Y . Fang, L. Sun, P. S. Yuet al., “Graph foundation models: Concepts, opportunities and challenges,”IEEE Transactions on Pattern Analysis and Machine Intelligence, 2025
work page 2025
-
[53]
H. Yuan, Q. Sun, J. Tao, X. Fu, and J. Li, “Rag-gfm: Overcoming in-memory bottlenecks in graph foundation models via retrieval-augmented generation,” inProceedings of the ACM Web Conference 2026, 2026, pp. 626–637
work page 2026
-
[54]
Y . Yan, P. Zhang, Z. Fang, and Q. Long, “Inductive graph alignment prompt: bridging the gap between graph pre-training and inductive fine-tuning from spectral perspective,” inProceedings of the ACM Web Conference 2024, 2024, pp. 4328–4339
work page 2024
-
[55]
Hgprompt: Bridging homogeneous and heterogeneous graphs for few-shot prompt learning,
X. Yu, Y . Fang, Z. Liu, and X. Zhang, “Hgprompt: Bridging homogeneous and heterogeneous graphs for few-shot prompt learning,” inProceedings of the AAAI conference on artificial intelligence, vol. 38, no. 15, 2024, pp. 16 578–16 586. 12
work page 2024
-
[56]
Multigprompt for multi-task pre-training and prompting on graphs,
X. Yu, C. Zhou, Y . Fang, and X. Zhang, “Multigprompt for multi-task pre-training and prompting on graphs,” inProceedings of the ACM Web Conference 2024, 2024, pp. 515–526
work page 2024
-
[57]
Zero-shot generalization of gnns over distinct attribute domains,
Y . Shen, J. Zhou, B. Bevilacqua, J. Robinson, C. Kanatsoulis, J. Leskovec, and B. Ribeiro, “Zero-shot generalization of gnns over distinct attribute domains,” 2024
work page 2024
-
[58]
Uniglm: Training one unified language model for text- attributed graphs embedding,
Y . Fang, D. Fan, S. Ding, N. Liu, and Q. Tan, “Uniglm: Training one unified language model for text- attributed graphs embedding,” inProceedings of the Eighteenth ACM International Conference on Web Search and Data Mining, 2025, pp. 973–981
work page 2025
-
[59]
A survey of cross-domain graph learning: Progress and future directions,
H. Zhao, Z. Li, C. Zi, A. Chen, F. Tsung, J. Li, and J. X. Yu, “A survey of cross-domain graph learning: Progress and future directions,”arXiv preprint arXiv:2503.11086, 2025
-
[60]
Graver: Generative graph vocabularies for robust graph foundation models fine-tuning,
H. Yuan, Q. Sun, J. Shi, X. Fu, B. Hooi, J. Li, and P. S. Yu, “Graver: Generative graph vocabularies for robust graph foundation models fine-tuning,”arXiv preprint arXiv:2511.05592, 2025
-
[61]
Graphprompt: Unifying pre-training and downstream tasks for graph neural networks,
Z. Liu, X. Yu, Y . Fang, and X. Zhang, “Graphprompt: Unifying pre-training and downstream tasks for graph neural networks,” inProceedings of the ACM web conference 2023, 2023, pp. 417–428
work page 2023
-
[62]
Ragraph: A general retrieval-augmented graph learning framework,
X. Jiang, R. Qiu, Y . Xu, W. Zhang, Y . Zhu, R. Zhang, Y . Fang, X. Chu, J. Zhao, and Y . Wang, “Ragraph: A general retrieval-augmented graph learning framework,”Advances in Neural Information Processing Systems, vol. 37, pp. 29 948–29 985, 2024
work page 2024
-
[63]
Samgpt: Text-free graph foundation model for multi- domain pre-training and cross-domain adaptation,
X. Yu, Z. Gong, C. Zhou, Y . Fang, and H. Zhang, “Samgpt: Text-free graph foundation model for multi- domain pre-training and cross-domain adaptation,” inProceedings of the ACM on Web Conference 2025, 2025, pp. 1142–1153
work page 2025
-
[64]
Non-homophilic graph pre-training and prompt learning,
X. Yu, J. Zhang, Y . Fang, and R. Jiang, “Non-homophilic graph pre-training and prompt learning,” in Proceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining V . 1, 2025, pp. 1844–1854
work page 2025
-
[65]
Towards effective federated graph foundation model via mitigating knowledge entanglement,
Y . Zhu, X. Li, J. Jia, M. Hu, D. Wu, and M. Qiu, “Towards effective federated graph foundation model via mitigating knowledge entanglement,”arXiv preprint arXiv:2505.12684, 2025
-
[66]
H. Yuan, Q. Sun, J. Shi, M. Liu, J. Yuan, Z. Zhang, X. Fu, and J. Li, “Retrieving minimal and suf- ficient reasoning subgraphs with graph foundation models for path-aware graphrag,”arXiv preprint arXiv:2603.07179, 2026
-
[67]
H. Yuan, Q. Sun, J. Shi, X. Fu, B. Hooi, J. Li, and P. S. Yu, “How much can transfer? bridge: Bounded multi-domain graph foundation model with generalization guarantees,” inForty-second International Conference on Machine Learning, 2025
work page 2025
-
[68]
Jointgt: Graph-text joint representation learning for text generation from knowledge graphs,
P. Ke, H. Ji, Y . Ran, X. Cui, L. Wang, L. Song, X. Zhu, and M. Huang, “Jointgt: Graph-text joint representation learning for text generation from knowledge graphs,” inFindings of the association for computational linguistics: ACL-IJCNLP 2021, 2021, pp. 2526–2538
work page 2021
-
[69]
Learning on large-scale text-attributed graphs via variational inference,
J. Zhao, M. Qu, C. Li, H. Yan, Q. Liu, R. Li, X. Xie, and J. Tang, “Learning on large-scale text-attributed graphs via variational inference,” inThe Eleventh International Conference on Learning Representations
-
[70]
Curriculum gnn-llm alignment for text-attributed graphs,
H. Chen, X. Wang, Z. Zhang, H. Li, W. Wen, L. Feng, and W. Zhu, “Curriculum gnn-llm alignment for text-attributed graphs,” 2025
work page 2025
-
[71]
Graphformers: Gnn-nested transformers for representation learning on textual graph,
J. Yang, Z. Liu, S. Xiao, C. Li, D. Lian, S. Agrawal, A. Singh, G. Sun, and X. Xie, “Graphformers: Gnn-nested transformers for representation learning on textual graph,”Advances in Neural Information Processing Systems, vol. 34, pp. 28 798–28 810, 2021
work page 2021
-
[72]
Can representation gaps be the key to enhancing robustness in graph-text alignment?
H. Zhang, T. Zhang, Y . Shi, X. Gu, Y . Shen, Z. Zhang, Y . Yuan, H. Zhang, and J. Huang, “Can representation gaps be the key to enhancing robustness in graph-text alignment?”arXiv preprint arXiv:2510.12087, 2025
-
[73]
LLM as GNN: Graph vocabulary learning for text-attributed graph foundation models,
X. Zhu, H. Xue, Z. Zhao, W. Xu, J. Huang, M. Guo, Q. Wang, K. Zhou, I. Razzak, and Y . Zhang, “Llm as gnn: Graph vocabulary learning for text-attributed graph foundation models,”arXiv preprint arXiv:2503.03313, 2025
-
[74]
Congrat: Self-supervised contrastive pretraining for joint graph and text embeddings,
W. Brannon, W. Kang, S. Fulay, H. Jiang, B. Roy, D. Roy, and J. Kabbara, “Congrat: Self-supervised contrastive pretraining for joint graph and text embeddings,” inProceedings of TextGraphs-17: Graph- based Methods for Natural Language Processing, 2024, pp. 19–39. 13 A Algorithm We provide the complete training pipeline of UniGraphLM in Algorithm 1. Algori...
work page 2024
-
[75]
Paper Summary and Context Analysis: 14 - Extract and summarize the key findings or contributions of the paper denoted by ‘n{node}’. Consider the details embedded within node ‘n{node}’, including its title, abstract, and keywords if available. - Provide an overall summary of prevalent themes or concepts shared by the papers that cite or are cited by ‘n{nod...
-
[76]
Research Area Classification: - Based on the information summarized from ‘n{node}’ and its neighboring nodes, determine the specific research area to which ‘n{node}’ primarily contributes. - Justify the classification by explaining which aspects of ‘n{node}’ align with recognized themes, issues, or methodologies in the identified research area. Please ens...
-
[78]
Research Area Classification: - Based on the information summarized from ‘n{node}’ and its neighboring nodes, determine the specific research area to which ‘n{node}’ primarily contributes. - Justify the classification by explaining which aspects of ‘n{node}’ align with recognized themes, issues, or methodologies in the identified research area. Please ens...
-
[79]
Paper Summary and Context Analysis: - Extract and summarize the key findings or contributions of the paper denoted by ‘n{node}’. Consider the details embedded within node ‘n{node}’, including its title, abstract, and keywords if available. - Provide an overall summary of prevalent themes or concepts shared by the papers that cite or are cited by ‘n{node}’...
-
[80]
Research Area Classification: 15 - Based on the information summarized from ‘n{node}’ and its neighboring nodes, determine the specific research area to which ‘n{node}’ primarily contributes. - Justify the classification by explaining which aspects of ‘n{node}’ align with recognized themes, issues, or methodologies in the identified research area. Please ...
-
[81]
- Then, examine the immediate neighboring nodes of ‘n{node}’ in the graph
Content Understanding and Neighborhood Analysis: - Summarize the main topic and key technical concepts of the Wikipedia entry denoted by ‘n{node}’, using the entry name and entry content provided in the node. - Then, examine the immediate neighboring nodes of ‘n{node}’ in the graph. Identify common themes, technologies, systems, or research topics that ap...
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.