{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,21]],"date-time":"2026-01-21T17:59:31Z","timestamp":1769018371187,"version":"3.49.0"},"publisher-location":"New York, NY, USA","reference-count":142,"publisher":"ACM","funder":[{"DOI":"10.13039\/https:\/\/doi.org\/10.13039\/100000001","name":"NSF (National Science Foundation)","doi-asserted-by":"publisher","award":["IIS-2321504, IIS-2334193, IIS-2340346, IIS-2217239, CNS-2426514, and CMMI-2146076"],"award-info":[{"award-number":["IIS-2321504, IIS-2334193, IIS-2340346, IIS-2217239, CNS-2426514, and CMMI-2146076"]}],"id":[{"id":"10.13039\/https:\/\/doi.org\/10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,8,3]]},"DOI":"10.1145\/3711896.3736568","type":"proceedings-article","created":{"date-parts":[[2025,8,3]],"date-time":"2025-08-03T20:52:41Z","timestamp":1754254361000},"page":"6184-6194","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["Graph Foundation Models: Challenges, Methods, and Open Questions"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-7670-6777","authenticated-orcid":false,"given":"Zehong","family":"Wang","sequence":"first","affiliation":[{"name":"University of Notre Dame, Notre Dame, IN, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8349-7926","authenticated-orcid":false,"given":"Chuxu","family":"Zhang","sequence":"additional","affiliation":[{"name":"University of Connecticut, Storrs, CT, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1878-817X","authenticated-orcid":false,"given":"Jundong","family":"Li","sequence":"additional","affiliation":[{"name":"University of Virginia, Charlottesville, VA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3932-5956","authenticated-orcid":false,"given":"Nitesh","family":"Chawla","sequence":"additional","affiliation":[{"name":"University of Notre Dame, Notre Dame, IN, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-6002-125X","authenticated-orcid":false,"given":"Yanfang","family":"Ye","sequence":"additional","affiliation":[{"name":"University of Notre Dame, Notre Dame, IN, USA"}]}],"member":"320","published-online":{"date-parts":[[2025,8,3]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al.","author":"Achiam Josh","year":"2023","unstructured":"Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, Florencia Leoni Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al. 2023. Gpt-4 technical report. arXiv (2023)."},{"key":"e_1_3_2_1_2_1","unstructured":"Uri Alon and Eran Yahav. 2021. On the Bottleneck of Graph Neural Networks and its Practical Implications. In ICLR."},{"key":"e_1_3_2_1_3_1","volume-title":"Sequential modeling enables scalable learning for large vision models. arXiv","author":"Bai Yutong","year":"2023","unstructured":"Yutong Bai, Xinyang Geng, Karttikeya Mangalam, Amir Bar, Alan Yuille, Trevor Darrell, Jitendra Malik, and Alexei A Efros. 2023. Sequential modeling enables scalable learning for large vision models. arXiv (2023)."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"crossref","unstructured":"Yutong Bai Xinyang Geng Karttikeya Mangalam Amir Bar Alan L Yuille Trevor Darrell Jitendra Malik and Alexei A Efros. 2024. Sequential modeling enables scalable learning for large vision models. In CVPR.","DOI":"10.1109\/CVPR52733.2024.02157"},{"key":"e_1_3_2_1_5_1","unstructured":"Ilyes Batatia Philipp Benner Yuan Chiang Alin M Elena D\u00e1vid P Kov\u00e1cs Janosh Riebesell Xavier R Advincula Mark Asta Matthew Avaylon William J Baldwin et al. 2023. A foundation model for atomistic materials chemistry. arXiv (2023)."},{"key":"e_1_3_2_1_6_1","volume-title":"Zhiyi Li, Gabriela Moisescu-Pareja, Oleksandr Dymov, Samuel Maddrell-Mander, Callum McLean, Frederik Wenkel, Luis M\u00fcller, et al.","author":"Beaini Dominique","year":"2023","unstructured":"Dominique Beaini, Shenyang Huang, Joao Alex Cunha, Zhiyi Li, Gabriela Moisescu-Pareja, Oleksandr Dymov, Samuel Maddrell-Mander, Callum McLean, Frederik Wenkel, Luis M\u00fcller, et al. 2023. Towards foundational models for molecular learning on large-scale multi-task datasets. arXiv (2023)."},{"key":"e_1_3_2_1_7_1","volume-title":"Position: Graph Learning Will Lose Relevance Due To Poor Benchmarks. arXiv","author":"Bechler-Speicher Maya","year":"2025","unstructured":"Maya Bechler-Speicher, Ben Finkelshtein, Fabrizio Frasca, Luis M\u00fcller, Jan T\u00f6nshoff, Antoine Siraudin, Viktor Zaverkin, Michael M Bronstein, Mathias Niepert, Bryan Perozzi, et al. 2025. Position: Graph Learning Will Lose Relevance Due To Poor Benchmarks. arXiv (2025)."},{"key":"e_1_3_2_1_8_1","unstructured":"Beatrice Bevilacqua Joshua Robinson Jure Leskovec and Bruno Ribeiro. 2025. Holographic Node Representations: Pre-training Task-Agnostic Node Embeddings. In ICLR."},{"key":"e_1_3_2_1_9_1","volume-title":"A comparative study of training algorithms for supervised machine learning. IJSCE","author":"Bhavsar Hetal","year":"2012","unstructured":"Hetal Bhavsar and Amit Ganatra. 2012. A comparative study of training algorithms for supervised machine learning. IJSCE (2012)."},{"key":"e_1_3_2_1_10_1","unstructured":"Rishi Bommasani Drew A Hudson Ehsan Adeli Russ Altman Simran Arora Sydney von Arx Michael S Bernstein Jeannette Bohg Antoine Bosselut Emma Brunskill et al. 2021. On the opportunities and risks of foundation models. arXiv (2021)."},{"key":"e_1_3_2_1_11_1","volume-title":"Graph theory","author":"Bondy John Adrian","unstructured":"John Adrian Bondy and Uppaluri Siva Ramachandra Murty. 2008. Graph theory. Springer Publishing Company, Incorporated."},{"key":"e_1_3_2_1_12_1","unstructured":"Tom Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared D Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell et al. 2020. Language models are few-shot learners. In NeurIPS."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/3447786.3456233"},{"key":"e_1_3_2_1_14_1","volume-title":"Graphwiz: An instruction-following language model for graph computational problems. In KDD.","author":"Chen Nuo","year":"2024","unstructured":"Nuo Chen, Yuhan Li, Jianheng Tang, and Jia Li. 2024a. Graphwiz: An instruction-following language model for graph computational problems. In KDD."},{"key":"e_1_3_2_1_15_1","volume-title":"Neil Shah, and Zhangyang Wang.","author":"Chen Runjin","year":"2024","unstructured":"Runjin Chen, Tong Zhao, AJAY KUMAR JAISWAL, Neil Shah, and Zhangyang Wang. 2024c. LLaGA: Large Language and Graph Assistant. In ICML."},{"key":"e_1_3_2_1_16_1","unstructured":"Zhikai Chen Haitao Mao Jingzhe Liu Yu Song Bingheng Li Wei Jin Bahare Fatemi Anton Tsitsulin Bryan Perozzi Hui Liu et al. 2024b. Text-space Graph Foundation Models: Comprehensive Benchmarks and New Insights. arXiv (2024)."},{"key":"e_1_3_2_1_17_1","volume-title":"Boosting Graph Foundation Model from Structural Perspective. arXiv","author":"Cheng Yao","year":"2024","unstructured":"Yao Cheng, Yige Zhao, Jianxiang Yu, and Xiang Li. 2024. Boosting Graph Foundation Model from Structural Perspective. arXiv (2024)."},{"key":"e_1_3_2_1_18_1","volume-title":"A Prompt-Based Knowledge Graph Foundation Model for Universal In-Context Reasoning. NeurIPS","author":"Cui Yuanning","year":"2025","unstructured":"Yuanning Cui, Zequn Sun, and Wei Hu. 2025. A Prompt-Based Knowledge Graph Foundation Model for Universal In-Context Reasoning. NeurIPS (2025)."},{"key":"e_1_3_2_1_19_1","volume-title":"Alice Wang, Enrico Palumbo, and Mounia Lalmas.","author":"Damianou Andreas","year":"2024","unstructured":"Andreas Damianou, Francesco Fabbri, Paul Gigioli, Marco De Nadai, Alice Wang, Enrico Palumbo, and Mounia Lalmas. 2024. Towards graph foundation models for personalization. In WWW."},{"key":"e_1_3_2_1_20_1","volume-title":"Its All Graph To Me: Single-Model Graph Representation Learning on Multiple Domains. In NeurIPS 2023 Workshop.","author":"Davies Alex","year":"2023","unstructured":"Alex Davies, Riku Green, Nirav Ajmeri, and Telmo Silva Filho. 2023. Its All Graph To Me: Single-Model Graph Representation Learning on Multiple Domains. In NeurIPS 2023 Workshop."},{"key":"e_1_3_2_1_21_1","volume-title":"Feature engineering for machine learning and data analytics","author":"Dong Guozhu","unstructured":"Guozhu Dong and Huan Liu. 2018. Feature engineering for machine learning and data analytics. CRC press."},{"key":"e_1_3_2_1_22_1","unstructured":"Alexey Dosovitskiy Lucas Beyer Alexander Kolesnikov Dirk Weissenborn Xiaohua Zhai Thomas Unterthiner Mostafa Dehghani Matthias Minderer Georg Heigold Sylvain Gelly Jakob Uszkoreit and Neil Houlsby. 2021. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In ICLR."},{"key":"e_1_3_2_1_23_1","volume-title":"An appraisal of some shortest-path algorithms. Operations research","author":"Dreyfus Stuart E","year":"1969","unstructured":"Stuart E Dreyfus. 1969. An appraisal of some shortest-path algorithms. Operations research (1969)."},{"key":"e_1_3_2_1_24_1","volume-title":"Anh Tuan Luu, and Dominique Beaini","author":"Dwivedi Vijay Prakash","year":"2022","unstructured":"Vijay Prakash Dwivedi, Ladislav Ramp\u00e1\u0161ek, Mikhail Galkin, Ali Parviz, Guy Wolf, Anh Tuan Luu, and Dominique Beaini. 2022. Long Range Graph Benchmark. In NeurIPS."},{"key":"e_1_3_2_1_25_1","unstructured":"Wenqi Fan Shijie Wang Jiani Huang Zhikai Chen Yu Song Wenzhuo Tang Haitao Mao Hui Liu Xiaorui Liu Dawei Yin et al. 2024. Graph machine learning in the era of large language models (llms). arXiv (2024)."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"crossref","unstructured":"Yi Fang Dongzhe Fan Sirui Ding Ninghao Liu and Qiaoyu Tan. 2024. UniGLM: Training One Unified Language Model for Text-Attributed Graphs. In WSDM.","DOI":"10.1145\/3701551.3703586"},{"key":"e_1_3_2_1_27_1","volume-title":"Rishabh Ranjan, Joshua Robinson, Rex Ying, Jiaxuan You, and Jure Leskovec.","author":"Fey Matthias","year":"2024","unstructured":"Matthias Fey, Weihua Hu, Kexin Huang, Jan Eric Lenssen, Rishabh Ranjan, Joshua Robinson, Rex Ying, Jiaxuan You, and Jure Leskovec. 2024. Position: Relational deep learning-graph representation learning on relational databases. In ICML."},{"key":"e_1_3_2_1_28_1","unstructured":"Mikhail Galkin Xinyu Yuan Hesham Mostafa Jian Tang and Zhaocheng Zhu. 2024a. Towards foundation models for knowledge graph reasoning. In ICLR."},{"key":"e_1_3_2_1_29_1","volume-title":"A Foundation Model for Zero-shot Logical Query Reasoning. arXiv","author":"Galkin Mikhail","year":"2024","unstructured":"Mikhail Galkin, Jincheng Zhou, Bruno Ribeiro, Jian Tang, and Zhaocheng Zhu. 2024b. A Foundation Model for Zero-shot Logical Query Reasoning. arXiv (2024)."},{"key":"e_1_3_2_1_30_1","volume-title":"A Graph is Worth K Words: Euclideanizing Graph using Pure Transformer. arXiv","author":"Gao Zhangyang","year":"2024","unstructured":"Zhangyang Gao, Daize Dong, Cheng Tan, Jun Xia, Bozhen Hu, and Stan Z Li. 2024. A Graph is Worth K Words: Euclideanizing Graph using Pure Transformer. arXiv (2024)."},{"key":"e_1_3_2_1_31_1","unstructured":"Justin Gilmer Samuel S Schoenholz Patrick F Riley Oriol Vinyals and George E Dahl. 2017. Neural message passing for quantum chemistry. In ICML."},{"key":"e_1_3_2_1_32_1","volume-title":"Pseudocode-Injection Magic: Enabling LLMs to Tackle Graph Computational Tasks. arXiv","author":"Gong Chang","year":"2025","unstructured":"Chang Gong, Wanrui Bian, Zhijie Zhang, and Weiguo Zheng. 2025. Pseudocode-Injection Magic: Enabling LLMs to Tackle Graph Computational Tasks. arXiv (2025)."},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"crossref","unstructured":"Aditya Grover and Jure Leskovec. 2016. node2vec: Scalable feature learning for networks. In KDD.","DOI":"10.1145\/2939672.2939754"},{"key":"e_1_3_2_1_34_1","unstructured":"Will Hamilton Zhitao Ying and Jure Leskovec. 2017. Inductive representation learning on large graphs. In NeurIPS."},{"key":"e_1_3_2_1_35_1","unstructured":"Xueting Han Zhenhuan Huang Bang An and Jing Bai. 2021. Adaptive transfer learning on graph neural networks. In KDD."},{"key":"e_1_3_2_1_36_1","unstructured":"Kaiming He Xinlei Chen Saining Xie Yanghao Li Piotr Doll\u00e1r and Ross Girshick. 2022. Masked autoencoders are scalable vision learners. In CVPR."},{"key":"e_1_3_2_1_37_1","volume-title":"UniGraph: Learning a Cross-Domain Graph Foundation Model From Natural Language. arXiv","author":"He Yufei","year":"2024","unstructured":"Yufei He and Bryan Hooi. 2024. UniGraph: Learning a Cross-Domain Graph Foundation Model From Natural Language. arXiv (2024)."},{"key":"e_1_3_2_1_38_1","volume-title":"Generalizing Graph Transformers Across Diverse Graphs and Tasks via Pre-Training on Industrial-Scale Data. arXiv","author":"He Yufei","year":"2024","unstructured":"Yufei He, Zhenyu Hou, Yukuo Cen, Feng He, Xu Cheng, and Bryan Hooi. 2024. Generalizing Graph Transformers Across Diverse Graphs and Tasks via Pre-Training on Industrial-Scale Data. arXiv (2024)."},{"key":"e_1_3_2_1_39_1","unstructured":"Yufei He Yuan Sui Xiaoxin He Yue Liu Yifei Sun and Bryan Hooi. 2025. UniGraph2: Learning a Unified Embedding Space to Bind Multimodal Graphs. In WWW."},{"key":"e_1_3_2_1_40_1","volume-title":"Graphalign: Pretraining one graph neural network on multiple graphs via feature alignment. arXiv","author":"Hou Zhenyu","year":"2024","unstructured":"Zhenyu Hou, Haozhan Li, Yukuo Cen, Jie Tang, and Yuxiao Dong. 2024. Graphalign: Pretraining one graph neural network on multiple graphs via feature alignment. arXiv (2024)."},{"key":"e_1_3_2_1_41_1","volume-title":"Beyond text: A deep dive into large language models' ability on understanding graph data. arXiv","author":"Hu Yuntong","year":"2023","unstructured":"Yuntong Hu, Zheng Zhang, and Liang Zhao. 2023. Beyond text: A deep dive into large language models' ability on understanding graph data. arXiv (2023)."},{"key":"e_1_3_2_1_42_1","unstructured":"Zhengyu Hu Yichuan Li Zhengyu Chen Jingang Wang Han Liu Kyumin Lee and Kaize Ding. 2024. Let`s Ask GNN: Empowering Large Language Model for Graph In-Context Learning. In EMNLP."},{"key":"e_1_3_2_1_43_1","volume-title":"Prodigy: Enabling in-context learning over graphs. In NeurIPS.","author":"Huang Qian","year":"2023","unstructured":"Qian Huang, Hongyu Ren, Peng Chen, Gregor Kr\u017emanc, Daniel Zeng, Percy S Liang, and Jure Leskovec. 2023. Prodigy: Enabling in-context learning over graphs. In NeurIPS."},{"key":"e_1_3_2_1_44_1","volume-title":"How Expressive are Knowledge Graph Foundation Models? arXiv","author":"Huang Xingyue","year":"2025","unstructured":"Xingyue Huang, Pablo Barcel\u00f3, Michael M Bronstein,Ismail. Ilkan Ceylan, Mikhail Galkin, Juan L Reutter, and Miguel Romero Orth. 2025. How Expressive are Knowledge Graph Foundation Models? arXiv (2025)."},{"key":"e_1_3_2_1_45_1","volume-title":"Matko Bo\u0161njak, Alex Vitvitskyi, Yulia Rubanova, et al.","author":"Ibarz Borja","year":"2022","unstructured":"Borja Ibarz, Vitaly Kurin, George Papamakarios, Kyriacos Nikiforou, Mehdi Bennani, R\u00f3bert Csord\u00e1s, Andrew Joseph Dudzik, Matko Bo\u0161njak, Alex Vitvitskyi, Yulia Rubanova, et al. 2022. A generalist neural algorithmic learner. In LoG."},{"key":"e_1_3_2_1_46_1","volume-title":"NT-LLM: A Novel Node Tokenizer for Integrating Graph Structure into Large Language Models. arXiv","author":"Ji Yanbiao","year":"2024","unstructured":"Yanbiao Ji, Chang Liu, Xin Chen, Yue Ding, Dan Luo, Mei Li, Wenqing Lin, and Hongtao Lu. 2024. NT-LLM: A Novel Node Tokenizer for Integrating Graph Structure into Large Language Models. arXiv (2024)."},{"key":"e_1_3_2_1_47_1","volume":"201","author":"Jia Ruoxi","unstructured":"Ruoxi Jia, David Dao, Boxin Wang, Frances Ann Hubis, Nick Hynes, Nezihe Merve G\u00fcrel, Bo Li, Ce Zhang, Dawn Song, and Costas J Spanos. 2019. Towards efficient data valuation based on the shapley value. In AISTATS.","journal-title":"Costas J Spanos."},{"key":"e_1_3_2_1_48_1","unstructured":"Xinke Jiang Rihong Qiu Yongxin Xu Wentao Zhang Yichen Zhu Ruizhe Zhang Yuchen Fang Xu Chu Junfeng Zhao and Yasha Wang. 2024. RAGraph: A General Retrieval-Augmented Graph Learning Framework. In NeurIPS."},{"key":"e_1_3_2_1_49_1","volume-title":"Large language models on graphs: A comprehensive survey. TKDE","author":"Jin Bowen","year":"2024","unstructured":"Bowen Jin, Gang Liu, Chi Han, Meng Jiang, Heng Ji, and Jiawei Han. 2024. Large language models on graphs: A comprehensive survey. TKDE (2024)."},{"key":"e_1_3_2_1_50_1","unstructured":"Yongcheng Jing Chongbin Yuan Li Ju Yiding Yang Xinchao Wang and Dacheng Tao. 2023. Deep graph reprogramming. In CVPR."},{"key":"e_1_3_2_1_51_1","unstructured":"Mingxuan Ju Tong Zhao Qianlong Wen Wenhao Yu Neil Shah Yanfang Ye and Chuxu Zhang. 2023. Multi-task self-supervised graph neural networks enable stronger task generalization. In ICLR."},{"key":"e_1_3_2_1_52_1","volume-title":"Kipf and Max Welling","author":"Thomas","year":"2017","unstructured":"Thomas N. Kipf and Max Welling. 2017. Semi-Supervised Classification with Graph Convolutional Networks. In ICLR."},{"key":"e_1_3_2_1_53_1","volume-title":"ICML 2024 Workshop.","author":"Klaser Kerstin","year":"2024","unstructured":"Kerstin Klaser, Blazej Banaszewski, Samuel Maddrell-Mander, Callum McLean, Luis M\u00fcller, Ali Parviz, Shenyang Huang, and Andrew W Fitzgibbon. 2024. Minimol: A parameter-efficient foundation model for molecular learning. In ICML 2024 Workshop."},{"key":"e_1_3_2_1_54_1","volume-title":"GOFA: A Generative One-For-All Model for Joint Graph Language Modeling. In ICLR.","author":"Kong Lecheng","year":"2025","unstructured":"Lecheng Kong, Jiarui Feng, Hao Liu, Chengsong Huang, Jiaxin Huang, Yixin Chen, and Muhan Zhang. 2025. GOFA: A Generative One-For-All Model for Joint Graph Language Modeling. In ICLR."},{"key":"e_1_3_2_1_55_1","unstructured":"Guohao Li Matthias Muller Ali Thabet and Bernard Ghanem. 2019. DeepGCNs: Can GCNs go as deep as CNNs?. In ICCV."},{"key":"e_1_3_2_1_56_1","volume-title":"Are Large Language Models In-Context Graph Learners? arXiv","author":"Li Jintang","year":"2025","unstructured":"Jintang Li, Ruofan Wu, Yuchang Zhu, Huizhe Zhang, Liang Chen, and Zibin Zheng. 2025. Are Large Language Models In-Context Graph Learners? arXiv (2025)."},{"key":"e_1_3_2_1_57_1","volume-title":"A survey of graph meets large language model: Progress and future directions. arXiv","author":"Li Yuhan","year":"2023","unstructured":"Yuhan Li, Zhixun Li, Peisong Wang, Jia Li, Xiangguo Sun, Hong Cheng, and Jeffrey Xu Yu. 2023. A survey of graph meets large language model: Progress and future directions. arXiv (2023)."},{"key":"e_1_3_2_1_58_1","volume-title":"Jeffrey Xu Yu, and Jia Li","author":"Li Yuhan","year":"2024","unstructured":"Yuhan Li, Peisong Wang, Zhixun Li, Jeffrey Xu Yu, and Jia Li. 2024. ZeroG: Investigating Cross-dataset Zero-shot Transferability in Graphs. In KDD."},{"key":"e_1_3_2_1_59_1","volume-title":"LangGFM: A Large Language Model Alone Can be a Powerful Graph Foundation Model. arXiv","author":"Lin Tianqianjin","year":"2024","unstructured":"Tianqianjin Lin, Pengwei Yan, Kaisong Song, Zhuoren Jiang, Yangyang Kang, Jun Lin, Weikang Yuan, Junjie Cao, Changlong Sun, and Xiaozhong Liu. 2024b. LangGFM: A Large Language Model Alone Can be a Powerful Graph Foundation Model. arXiv (2024)."},{"key":"e_1_3_2_1_60_1","unstructured":"Zhenghao Lin Zhibin Gou Yeyun Gong Xiao Liu Yelong Shen Ruochen Xu Chen Lin Yujiu Yang Jian Jiao Nan Duan et al. 2024a. Rho-1: Not all tokens are what you need. arXiv (2024)."},{"key":"e_1_3_2_1_61_1","unstructured":"Hao Liu Jiarui Feng Lecheng Kong Ningyue Liang Dacheng Tao Yixin Chen and Muhan Zhang. 2024b. One for all: Towards training one graph model for all classification tasks. In ICLR."},{"key":"e_1_3_2_1_62_1","volume-title":"2024 d. One Model for One Graph: A New Perspective for Pretraining with Cross-domain Graphs. arXiv","author":"Liu Jingzhe","year":"2024","unstructured":"Jingzhe Liu, Haitao Mao, Zhikai Chen, Wenqi Fan, Mingxuan Ju, Tong Zhao, Neil Shah, and Jiliang Tang. 2024 d. One Model for One Graph: A New Perspective for Pretraining with Cross-domain Graphs. arXiv (2024)."},{"key":"e_1_3_2_1_63_1","unstructured":"Jiawei Liu Cheng Yang Zhiyuan Lu Junze Chen Yibo Li Mengmei Zhang Ting Bai Yuan Fang Lichao Sun Philip S Yu et al. 2023b. Towards graph foundation models: A survey and beyond. arXiv (2023)."},{"key":"e_1_3_2_1_64_1","volume-title":"2024 e. Git-mol: A multi-modal large language model for molecular science with graph, image, and text. Computers in biology and medicine","author":"Liu Pengfei","year":"2024","unstructured":"Pengfei Liu, Yiming Ren, Jun Tao, and Zhixiang Ren. 2024 e. Git-mol: A multi-modal large language model for molecular science with graph, image, and text. Computers in biology and medicine (2024)."},{"key":"e_1_3_2_1_65_1","volume-title":"Moleculargpt: Open large language model (llm) for few-shot molecular property prediction. arXiv","author":"Liu Yuyan","year":"2024","unstructured":"Yuyan Liu, Sirui Ding, Sheng Zhou, Wenqi Fan, and Qiaoyu Tan. 2024a. Moleculargpt: Open large language model (llm) for few-shot molecular property prediction. arXiv (2024)."},{"key":"e_1_3_2_1_66_1","volume-title":"Graph self-supervised learning: A survey. TKDE","author":"Liu Yixin","year":"2022","unstructured":"Yixin Liu, Ming Jin, Shirui Pan, Chuan Zhou, Yu Zheng, Feng Xia, and S Yu Philip. 2022. Graph self-supervised learning: A survey. TKDE (2022)."},{"key":"e_1_3_2_1_67_1","unstructured":"Yixin Liu Shiyuan Li Yu Zheng Qingfeng Chen Chengqi Zhang and Shirui Pan. 2025. Arc: a generalist graph anomaly detector with in-context learning. In NeurIPS."},{"key":"e_1_3_2_1_68_1","unstructured":"Zheyuan Liu Xiaoxin He Yijun Tian and Nitesh V Chawla. 2024c. Can we soft prompt LLMs for graph learning tasks?. In WWW."},{"key":"e_1_3_2_1_69_1","unstructured":"Zhiyuan Liu Sihang Li Yanchen Luo Hao Fei Yixin Cao Kenji Kawaguchi Xiang Wang and Tat-Seng Chua. 2023a. MolCA: Molecular Graph-Language Modeling with Cross-Modal Projector and Uni-Modal Adapter. In EMNLP."},{"key":"e_1_3_2_1_70_1","volume-title":"Graphprompt: Unifying pre-training and downstream tasks for graph neural networks. In WWW.","author":"Liu Zemin","year":"2023","unstructured":"Zemin Liu, Xingtong Yu, Yuan Fang, and Xinming Zhang. 2023c. Graphprompt: Unifying pre-training and downstream tasks for graph neural networks. In WWW."},{"key":"e_1_3_2_1_71_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.findings-acl.658"},{"key":"e_1_3_2_1_72_1","volume-title":"Enhance Graph Alignment for Large Language Models. arXiv","author":"Luo Haitong","year":"2024","unstructured":"Haitong Luo, Xuying Meng, Suhang Wang, Tianxiang Zhao, Fali Wang, Hanyun Cao, and Yujun Zhang. 2024a. Enhance Graph Alignment for Large Language Models. arXiv (2024)."},{"key":"e_1_3_2_1_73_1","volume-title":"Graphinstruct: Empowering large language models with graph understanding and reasoning capability. arXiv","author":"Luo Zihan","year":"2024","unstructured":"Zihan Luo, Xiran Song, Hong Huang, Jianxun Lian, Chenhao Zhang, Jinqi Jiang, and Xing Xie. 2024b. Graphinstruct: Empowering large language models with graph understanding and reasoning capability. arXiv (2024)."},{"key":"e_1_3_2_1_74_1","doi-asserted-by":"crossref","unstructured":"Batta Mahesh et al. 2020. Machine learning algorithms-a review. IJSR (2020).","DOI":"10.21275\/ART20203995"},{"key":"e_1_3_2_1_75_1","unstructured":"Haitao Mao Zhikai Chen Wenzhuo Tang Jianan Zhao Yao Ma Tong Zhao Neil Shah Michael Galkin and Jiliang Tang. 2024. Graph Foundation Models Are Already Here. In ICML."},{"key":"e_1_3_2_1_76_1","volume-title":"Ebrahim Songhori, Shen Wang, Young-Joon Lee, Eric Johnson, Omkar Pathak, Azade Nova, et al.","author":"Mirhoseini Azalia","year":"2021","unstructured":"Azalia Mirhoseini, Anna Goldie, Mustafa Yazgan, Joe Wenjie Jiang, Ebrahim Songhori, Shen Wang, Young-Joon Lee, Eric Johnson, Omkar Pathak, Azade Nova, et al. 2021. A graph placement methodology for fast chip design. Nature (2021)."},{"key":"e_1_3_2_1_77_1","volume-title":"Ron Levie, Derek Lim, Michael M. Bronstein, Martin Grohe, and Stefanie Jegelka.","author":"Morris Christopher","year":"2024","unstructured":"Christopher Morris, Fabrizio Frasca, Nadav Dym, Haggai Maron, Ismail Ilkan Ceylan, Ron Levie, Derek Lim, Michael M. Bronstein, Martin Grohe, and Stefanie Jegelka. 2024. Position: Future Directions in the Theory of Graph Machine Learning. In ICML."},{"key":"e_1_3_2_1_78_1","volume-title":"Zero-shot Generalist Graph Anomaly Detection with Unified Neighborhood Prompts. arXiv","author":"Niu Chaoxi","year":"2024","unstructured":"Chaoxi Niu, Hezhe Qiao, Changlu Chen, Ling Chen, and Guansong Pang. 2024. Zero-shot Generalist Graph Anomaly Detection with Unified Neighborhood Prompts. arXiv (2024)."},{"key":"e_1_3_2_1_79_1","doi-asserted-by":"publisher","DOI":"10.1145\/2623330.2623732"},{"key":"e_1_3_2_1_80_1","volume-title":"AnomalyGFM: Graph Foundation Model for Zero\/Few-shot Anomaly Detection. arXiv","author":"Qiao Hezhe","year":"2025","unstructured":"Hezhe Qiao, Chaoxi Niu, Ling Chen, and Guansong Pang. 2025. AnomalyGFM: Graph Foundation Model for Zero\/Few-shot Anomaly Detection. arXiv (2025)."},{"key":"e_1_3_2_1_81_1","doi-asserted-by":"publisher","DOI":"10.1145\/3394486.3403168"},{"key":"e_1_3_2_1_82_1","doi-asserted-by":"crossref","unstructured":"Xubin Ren Jiabin Tang Dawei Yin Nitesh Chawla and Chao Huang. 2024a. A survey of large language models for graphs. In KDD.","DOI":"10.1145\/3637528.3671460"},{"key":"e_1_3_2_1_83_1","doi-asserted-by":"crossref","unstructured":"Xubin Ren Wei Wei Lianghao Xia Lixin Su Suqi Cheng Junfeng Wang Dawei Yin and Chao Huang. 2024b. Representation learning with large language models for recommendation. In WWW.","DOI":"10.1145\/3589334.3645458"},{"key":"e_1_3_2_1_84_1","unstructured":"Yu Rong Yatao Bian Tingyang Xu Weiyang Xie Ying Wei Wenbing Huang and Junzhou Huang. 2020. Self-supervised graph transformer on large-scale molecular data. In NeurIPS."},{"key":"e_1_3_2_1_85_1","volume-title":"Understanding transformer reasoning capabilities via graph algorithms. NeurIPS","author":"Sanford Clayton","year":"2025","unstructured":"Clayton Sanford, Bahare Fatemi, Ethan Hall, Anton Tsitsulin, Mehran Kazemi, Jonathan Halcrow, Bryan Perozzi, and Vahab Mirrokni. 2025. Understanding transformer reasoning capabilities via graph algorithms. NeurIPS (2025)."},{"key":"e_1_3_2_1_86_1","volume-title":"From molecules to materials: Pre-training large generalizable models for atomic property prediction. arXiv","author":"Shoghi Nima","year":"2023","unstructured":"Nima Shoghi, Adeesh Kolluru, John R Kitchin, Zachary W Ulissi, C Lawrence Zitnick, and Brandon M Wood. 2023. From molecules to materials: Pre-training large generalizable models for atomic property prediction. arXiv (2023)."},{"key":"e_1_3_2_1_87_1","volume-title":"A molecular multimodal foundation model associating molecule graphs with natural language. arXiv","author":"Su Bing","year":"2022","unstructured":"Bing Su, Dazhao Du, Zhao Yang, Yujie Zhou, Jiangmeng Li, Anyi Rao, Hao Sun, Zhiwu Lu, and Ji-Rong Wen. 2022. A molecular multimodal foundation model associating molecule graphs with natural language. arXiv (2022)."},{"key":"e_1_3_2_1_88_1","doi-asserted-by":"crossref","unstructured":"Li Sun Zhenhao Huang Suyang Zhou Qiqi Wan Hao Peng and Philip Yu. 2025 a. RiemannGFM: Learning a Graph Foundation Model from Riemannian Geometry. In WWW.","DOI":"10.1145\/3696410.3714952"},{"key":"e_1_3_2_1_89_1","volume-title":"Gppt: Graph pre-training and prompt tuning to generalize graph neural networks. In KDD.","author":"Sun Mingchen","year":"2022","unstructured":"Mingchen Sun, Kaixiong Zhou, Xin He, Ying Wang, and Xin Wang. 2022. Gppt: Graph pre-training and prompt tuning to generalize graph neural networks. In KDD."},{"key":"e_1_3_2_1_90_1","unstructured":"Xiangguo Sun Hong Cheng Jia Li Bo Liu and Jihong Guan. 2023a. All in one: Multi-task prompting for graph neural networks. In KDD."},{"key":"e_1_3_2_1_91_1","volume-title":"Graph prompt learning: A comprehensive survey and beyond. arXiv","author":"Sun Xiangguo","year":"2023","unstructured":"Xiangguo Sun, Jiawen Zhang, Xixi Wu, Hong Cheng, Yun Xiong, and Jia Li. 2023b. Graph prompt learning: A comprehensive survey and beyond. arXiv (2023)."},{"key":"e_1_3_2_1_92_1","volume-title":"2025 b. GraphICL: Unlocking Graph Learning Potential in LLMs through Structured Prompt Design. arXiv","author":"Sun Yuanfu","year":"2025","unstructured":"Yuanfu Sun, Zhengnan Ma, Yi Fang, Jing Ma, and Qiaoyu Tan. 2025 b. GraphICL: Unlocking Graph Learning Potential in LLMs through Structured Prompt Design. arXiv (2025)."},{"key":"e_1_3_2_1_93_1","doi-asserted-by":"crossref","unstructured":"Yifei Sun Yang Yang Xiao Feng Zijun Wang Haoyang Zhong Chunping Wang and Lei Chen. 2025 c. Handling Feature Heterogeneity with Learnable Graph Patches. In KDD.","DOI":"10.1145\/3690624.3709242"},{"key":"e_1_3_2_1_94_1","doi-asserted-by":"publisher","DOI":"10.1145\/2736277.2741093"},{"key":"e_1_3_2_1_95_1","volume-title":"Cross-domain graph data scaling: A showcase with diffusion models. arXiv","author":"Tang Wenzhuo","year":"2024","unstructured":"Wenzhuo Tang, Haitao Mao, Danial Dervovic, Ivan Brugere, Saumitra Mishra, Yuying Xie, and Jiliang Tang. 2024. Cross-domain graph data scaling: A showcase with diffusion models. arXiv (2024)."},{"key":"e_1_3_2_1_96_1","unstructured":"Crawl4ai team. 2024. Crawl4ai. crawl4ai.com."},{"key":"e_1_3_2_1_97_1","unstructured":"Petar Veli\u010dkovi\u0107 Guillem Cucurull Arantxa Casanova Adriana Romero Pietro Li\u00f2 and Yoshua Bengio. 2018. Graph Attention Networks. In ICLR."},{"key":"e_1_3_2_1_98_1","volume-title":"Digress: Discrete denoising diffusion for graph generation. arXiv","author":"Vignac Clement","year":"2022","unstructured":"Clement Vignac, Igor Krawczuk, Antoine Siraudin, Bohan Wang, Volkan Cevher, and Pascal Frossard. 2022. Digress: Discrete denoising diffusion for graph generation. arXiv (2022)."},{"key":"e_1_3_2_1_99_1","volume-title":"Graph kernels. JMLR","author":"Vishwanathan S Vichy N","year":"2010","unstructured":"S Vichy N Vishwanathan, Nicol N Schraudolph, Risi Kondor, and Karsten M Borgwardt. 2010. Graph kernels. JMLR (2010)."},{"key":"e_1_3_2_1_100_1","volume-title":"A tutorial on spectral clustering. Statistics and computing","author":"Luxburg Ulrike Von","year":"2007","unstructured":"Ulrike Von Luxburg. 2007. A tutorial on spectral clustering. Statistics and computing (2007)."},{"key":"e_1_3_2_1_101_1","doi-asserted-by":"publisher","DOI":"10.1109\/CogMI52975.2021.00026"},{"key":"e_1_3_2_1_102_1","volume-title":"LLMs as Zero-shot Graph Learners: Alignment of GNN Representations with LLM Token Embeddings. NeurIPS","author":"Wang Duo","year":"2024","unstructured":"Duo Wang, Yuan Zuo, Fengzhi Li, and Junjie Wu. 2024c. LLMs as Zero-shot Graph Learners: Alignment of GNN Representations with LLM Token Embeddings. NeurIPS (2024)."},{"key":"e_1_3_2_1_103_1","volume-title":"Instructgraph: Boosting large language models via graph-centric instruction tuning and preference alignment. arXiv","author":"Wang Jianing","year":"2024","unstructured":"Jianing Wang, Junda Wu, Yupeng Hou, Yao Liu, Ming Gao, and Julian McAuley. 2024a. Instructgraph: Boosting large language models via graph-centric instruction tuning and preference alignment. arXiv (2024)."},{"key":"e_1_3_2_1_104_1","volume-title":"Towards Graph Foundation Models: The Perspective of Zero-shot Reasoning on Knowledge Graphs. arXiv","author":"Wang Kai","year":"2024","unstructured":"Kai Wang and Siqiang Luo. 2024. Towards Graph Foundation Models: The Perspective of Zero-shot Reasoning on Knowledge Graphs. arXiv (2024)."},{"key":"e_1_3_2_1_105_1","volume-title":"Graph agent: Explicit reasoning agent for graphs. arXiv","author":"Wang Qinyong","year":"2023","unstructured":"Qinyong Wang, Zhenxiang Gao, and Rong Xu. 2023. Graph agent: Explicit reasoning agent for graphs. arXiv (2023)."},{"key":"e_1_3_2_1_106_1","doi-asserted-by":"crossref","unstructured":"Zehong Wang Sidney Liu Zheyuan Zhang Tianyi Ma Chuxu Zhang and Yanfang Ye. 2025 b. Can LLMs Convert Graphs to Text-Attributed Graphs?. In NAACL.","DOI":"10.18653\/v1\/2025.naacl-long.65"},{"key":"e_1_3_2_1_107_1","volume-title":"2025 a","author":"Wang Zehong","year":"2025","unstructured":"Zehong Wang, Zheyuan Liu, Tianyi Ma, Jiazheng Li, Zheyuan Zhang, Xingbo Fu, Yiyang Li, Zhengqing Yuan, Wei Song, Yijun Ma, Qingkai Zeng, Xiusi Chen, Jianan Zhao, Jundong Li, Meng Jiang, Pietro Lio, Nitesh Chawla, Chuxu Zhang, and Yanfang Ye. 2025 a. Graph Foundation Models: A Comprehensive Survey. arXiv preprint arXiv:2505.15116 (2025)."},{"key":"e_1_3_2_1_108_1","volume-title":"GFT: Graph Foundation Model with Transferable Tree Vocabulary. In The Thirty-eighth Annual Conference on Neural Information Processing Systems.","author":"Wang Zehong","year":"2024","unstructured":"Zehong Wang, Zheyuan Zhang, Nitesh V Chawla, Chuxu Zhang, and Yanfang Ye. 2024b. GFT: Graph Foundation Model with Transferable Tree Vocabulary. In The Thirty-eighth Annual Conference on Neural Information Processing Systems."},{"key":"e_1_3_2_1_109_1","unstructured":"Zehong Wang Zheyuan Zhang Tianyi Ma Nitesh V Chawla Chuxu Zhang and Yanfang Ye. 2025 c. Neural Graph Pattern Machine. In ICML."},{"key":"e_1_3_2_1_110_1","volume-title":"2025 d","author":"Wang Zehong","unstructured":"Zehong Wang, Zheyuan Zhang, Tianyi Ma, Nitesh V Chawla, Chuxu Zhang, and Yanfang Ye. 2025 d. Towards Graph Foundation Models: Learning Generalities Across Graphs via Task-trees. In ICML."},{"key":"e_1_3_2_1_111_1","volume-title":"Llmrec: Large language models with graph augmentation for recommendation. In WSDM.","author":"Wei Wei","year":"2024","unstructured":"Wei Wei, Xubin Ren, Jiabin Tang, Qinyong Wang, Lixin Su, Suqi Cheng, Junfeng Wang, Dawei Yin, and Chao Huang. 2024. Llmrec: Large language models with graph augmentation for recommendation. In WSDM."},{"key":"e_1_3_2_1_112_1","volume-title":"Towards a general gnn framework for combinatorial optimization. arXiv","author":"Wenkel Frederik","year":"2024","unstructured":"Frederik Wenkel, Semih Cant\u00fcrk, Michael Perlmutter, and Guy Wolf. 2024. Towards a general gnn framework for combinatorial optimization. arXiv (2024)."},{"key":"e_1_3_2_1_113_1","volume-title":"The shapley value. Handbook of game theory with economic applications","author":"Winter Eyal","year":"2002","unstructured":"Eyal Winter. 2002. The shapley value. Handbook of game theory with economic applications (2002)."},{"key":"e_1_3_2_1_114_1","volume-title":"Graph Foundation Models for Recommendation: A Comprehensive Survey. arXiv","author":"Wu Bin","year":"2025","unstructured":"Bin Wu, Yihang Wang, Yuanhao Zeng, Jiawei Liu, Jiashu Zhao, Cheng Yang, Yawen Li, Long Xia, Dawei Yin, and Chuan Shi. 2025. Graph Foundation Models for Recommendation: A Comprehensive Survey. arXiv (2025)."},{"key":"e_1_3_2_1_115_1","volume-title":"A comprehensive survey on graph neural networks. TNNLS","author":"Wu Zonghan","year":"2020","unstructured":"Zonghan Wu, Shirui Pan, Fengwen Chen, Guodong Long, Chengqi Zhang, and S Yu Philip. 2020. A comprehensive survey on graph neural networks. TNNLS (2020)."},{"key":"e_1_3_2_1_116_1","volume-title":"Mole-BERT: Rethinking Pre-training Graph Neural Networks for Molecules. In The Eleventh International Conference on Learning Representations.","author":"Xia Jun","year":"2023","unstructured":"Jun Xia, Chengshuai Zhao, Bozhen Hu, Zhangyang Gao, Cheng Tan, Yue Liu, Siyuan Li, and Stan Z Li. 2023. Mole-BERT: Rethinking Pre-training Graph Neural Networks for Molecules. In The Eleventh International Conference on Learning Representations."},{"key":"e_1_3_2_1_117_1","volume-title":"AnyGraph: Graph Foundation Model in the Wild. arXiv","author":"Xia Lianghao","year":"2024","unstructured":"Lianghao Xia and Chao Huang. 2024. AnyGraph: Graph Foundation Model in the Wild. arXiv (2024)."},{"key":"e_1_3_2_1_118_1","volume-title":"Opengraph: Towards open graph foundation models. In EMNLP.","author":"Xia Lianghao","year":"2024","unstructured":"Lianghao Xia, Ben Kao, and Chao Huang. 2024. Opengraph: Towards open graph foundation models. In EMNLP."},{"key":"e_1_3_2_1_119_1","volume-title":"Graph Linearization Methods for Reasoning on Graphs with Large Language Models. arXiv","author":"Xypolopoulos Christos","year":"2024","unstructured":"Christos Xypolopoulos, Guokan Shang, Xiao Fei, Giannis Nikolentzos, Hadi Abdine, Iakovos Evdaimon, Michail Chatzianastasis, Giorgos Stamou, and Michalis Vazirgiannis. 2024. Graph Linearization Methods for Reasoning on Graphs with Large Language Models. arXiv (2024)."},{"key":"e_1_3_2_1_120_1","unstructured":"Yuchen Yan Peiyan Zhang Zheng Fang and Qingqing Long. 2024. Inductive Graph Alignment Prompt: Bridging the Gap between Graph Pre-training and Inductive Fine-tuning From Spectral Perspective. In WWW."},{"key":"e_1_3_2_1_121_1","volume-title":"GraphAgent: Agentic Graph Language Assistant. arXiv","author":"Yang Yuhao","year":"2024","unstructured":"Yuhao Yang, Jiabin Tang, Lianghao Xia, Xingchen Zou, Yuxuan Liang, and Chao Huang. 2024. GraphAgent: Agentic Graph Language Assistant. arXiv (2024)."},{"key":"e_1_3_2_1_122_1","unstructured":"Ruosong Ye Caiqi Zhang Runhui Wang Shuyuan Xu and Yongfeng Zhang. 2024. Language is all a graph needs. In EACL."},{"key":"e_1_3_2_1_123_1","unstructured":"Lijun Yu Jos\u00e9 Lezama Nitesh B Gundavarapu Luca Versari Kihyuk Sohn David Minnen Yong Cheng Agrim Gupta Xiuye Gu Alexander G Hauptmann et al. 2024b. Language Model Beats Diffusion-Tokenizer is Key to Visual Generation. In ICLR."},{"key":"e_1_3_2_1_124_1","volume-title":"Hgprompt: Bridging homogeneous and heterogeneous graphs for few-shot prompt learning. In AAAI.","author":"Yu Xingtong","year":"2024","unstructured":"Xingtong Yu, Yuan Fang, Zemin Liu, and Xinming Zhang. 2024a. Hgprompt: Bridging homogeneous and heterogeneous graphs for few-shot prompt learning. In AAAI."},{"key":"e_1_3_2_1_125_1","volume-title":"2025 a","author":"Yu Xingtong","unstructured":"Xingtong Yu, Zechuan Gong, Chang Zhou, Yuan Fang, and Hui Zhang. 2025 a. SAMGPT: Text-free Graph Foundation Model for Multi-domain Pre-training and Cross-domain Adaptation. In WWW."},{"key":"e_1_3_2_1_126_1","unstructured":"Xinmiao Yu Meng Qu Xiaocheng Feng and Bing Qin. 2024c. GraphAgent: Exploiting Large Language Models for Interpretable Learning on Text-attributed Graphs. https:\/\/openreview.net\/forum?id=L3jATpVEGv"},{"key":"e_1_3_2_1_127_1","unstructured":"Xingtong Yu Jie Zhang Yuan Fang and Renhe Jiang. 2025 b. Non-homophilic graph pre-training and prompt learning. In KDD."},{"key":"e_1_3_2_1_128_1","unstructured":"Xingtong Yu Chang Zhou Yuan Fang and Xinming Zhang. 2024 d. MultiGPrompt for Multi-Task Pre-Training and Prompting on Graphs. In WWW."},{"key":"e_1_3_2_1_129_1","volume-title":"Florence: A new foundation model for computer vision. arXiv","author":"Yuan Lu","year":"2021","unstructured":"Lu Yuan, Dongdong Chen, Yi-Ling Chen, Noel Codella, Xiyang Dai, Jianfeng Gao, Houdong Hu, Xuedong Huang, Boxin Li, Chunyuan Li, et al. 2021. Florence: A new foundation model for computer vision. arXiv (2021)."},{"key":"e_1_3_2_1_130_1","volume-title":"Graph-toolformer: To empower llms with graph reasoning ability via prompt augmented by chatgpt. arXiv","author":"Zhang Jiawei","year":"2023","unstructured":"Jiawei Zhang. 2023. Graph-toolformer: To empower llms with graph reasoning ability via prompt augmented by chatgpt. arXiv (2023)."},{"key":"e_1_3_2_1_131_1","volume-title":"Unimot: Unified molecule-text language model with discrete token representation. arXiv","author":"Zhang Juzheng","year":"2024","unstructured":"Juzheng Zhang, Yatao Bian, Yongqiang Chen, and Quanming Yao. 2024a. Unimot: Unified molecule-text language model with discrete token representation. arXiv (2024)."},{"key":"e_1_3_2_1_132_1","volume-title":"Graphtranslator: Aligning graph model to large language model for open-ended tasks. In WWW.","author":"Zhang Mengmei","year":"2024","unstructured":"Mengmei Zhang, Mingwei Sun, Peng Wang, Shen Fan, Yanhu Mo, Xiaoxiao Xu, Hong Liu, Cheng Yang, and Chuan Shi. 2024c. Graphtranslator: Aligning graph model to large language model for open-ended tasks. In WWW."},{"key":"e_1_3_2_1_133_1","volume-title":"Gcoder: Improving large language model for generalized graph problem solving. arXiv","author":"Zhang Qifan","year":"2024","unstructured":"Qifan Zhang, Xiaobin Hong, Jianheng Tang, Nuo Chen, Yuhan Li, Wenzhong Li, Jing Tang, and Jia Li. 2024b. Gcoder: Improving large language model for generalized graph problem solving. arXiv (2024)."},{"key":"e_1_3_2_1_134_1","volume-title":"Meta-transformer: A unified framework for multimodal learning. arXiv","author":"Zhang Yiyuan","year":"2023","unstructured":"Yiyuan Zhang, Kaixiong Gong, Kaipeng Zhang, Hongsheng Li, Yu Qiao, Wanli Ouyang, and Xiangyu Yue. 2023. Meta-transformer: A unified framework for multimodal learning. arXiv (2023)."},{"key":"e_1_3_2_1_135_1","doi-asserted-by":"crossref","unstructured":"Zheyuan Zhang Zehong Wang Shifu Hou Evan Hall Landon Bachman Jasmine White Vincent Galassi Nitesh V Chawla Chuxu Zhang and Yanfang Ye. 2024 d. Diet-ODIN: A Novel Framework for Opioid Misuse Detection with Interpretable Dietary Patterns. In KDD.","DOI":"10.1145\/3637528.3671587"},{"key":"e_1_3_2_1_136_1","volume-title":"Varun Sameer Taneja","author":"Zhang Zheyuan","unstructured":"Zheyuan Zhang, Zehong Wang, Tianyi Ma, Varun Sameer Taneja, Sofia Nelson, Nhi Ha Lan Le, Keerthiram Murugesan, Mingxuan Ju, Nitesh V Chawla, Chuxu Zhang, et al. 2024 e. MOPI-HFRS: A Multi-objective Personalized Health-aware Food Recommendation System with LLM-enhanced Interpretation. arXiv (2024)."},{"key":"e_1_3_2_1_137_1","doi-asserted-by":"crossref","unstructured":"Haihong Zhao Aochuan Chen Xiangguo Sun Hong Cheng and Jia Li. 2024a. All in One and One for All: A Simple yet Effective Method towards Cross-domain Graph Pretraining. In KDD.","DOI":"10.1145\/3637528.3671913"},{"key":"e_1_3_2_1_138_1","volume-title":"A Survey on Self-Supervised Graph Foundation Models: Knowledge-Based Perspective. arXiv","author":"Zhao Ziwen","year":"2024","unstructured":"Ziwen Zhao, Yixin Su, Yuhua Li, Yixiong Zou, Ruixuan Li, and Rui Zhang. 2024b. A Survey on Self-Supervised Graph Foundation Models: Knowledge-Based Perspective. arXiv (2024)."},{"key":"e_1_3_2_1_139_1","doi-asserted-by":"crossref","unstructured":"Ce Zhou Qian Li Chen Li Jun Yu Yixin Liu Guangjing Wang Kai Zhang Cheng Ji Qiben Yan Lifang He et al. 2023. A comprehensive survey on pretrained foundation models: A history from bert to chatgpt. arXiv (2023).","DOI":"10.1007\/s13042-024-02443-6"},{"key":"e_1_3_2_1_140_1","volume-title":"Investigating Instruction Tuning Large Language Models on Graphs. arXiv","author":"Zhu Kerui","year":"2024","unstructured":"Kerui Zhu, Bo-Wei Huang, Bowen Jin, Yizhu Jiao, Ming Zhong, Kevin Chang, Shou-De Lin, and Jiawei Han. 2024a. Investigating Instruction Tuning Large Language Models on Graphs. arXiv (2024)."},{"key":"e_1_3_2_1_141_1","unstructured":"Qi Zhu Carl Yang Yidan Xu Haonan Wang Chao Zhang and Jiawei Han. 2021. Transfer learning of graph neural networks with ego-graph information maximization. In NeurIPS."},{"key":"e_1_3_2_1_142_1","volume-title":"Graphclip: Enhancing transferability in graph foundation models for text-attributed graphs. arXiv","author":"Zhu Yun","year":"2024","unstructured":"Yun Zhu, Haizhou Shi, Xiaotang Wang, Yongchao Liu, Yaoke Wang, Boci Peng, Chuntao Hong, and Siliang Tang. 2024b. Graphclip: Enhancing transferability in graph foundation models for text-attributed graphs. arXiv (2024)."}],"event":{"name":"KDD '25: The 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining","location":"Toronto ON Canada","acronym":"KDD '25","sponsor":["SIGMOD ACM Special Interest Group on Management of Data","SIGKDD ACM Special Interest Group on Knowledge Discovery in Data"]},"container-title":["Proceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining V.2"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3711896.3736568","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,16]],"date-time":"2025-08-16T14:31:01Z","timestamp":1755354661000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3711896.3736568"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,3]]},"references-count":142,"alternative-id":["10.1145\/3711896.3736568","10.1145\/3711896"],"URL":"https:\/\/doi.org\/10.1145\/3711896.3736568","relation":{},"subject":[],"published":{"date-parts":[[2025,8,3]]},"assertion":[{"value":"2025-08-03","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}