{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T05:50:58Z","timestamp":1776059458067,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":109,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,8,24]],"date-time":"2024-08-24T00:00:00Z","timestamp":1724457600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"NSFC","award":["62206067"],"award-info":[{"award-number":["62206067"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,8,25]]},"DOI":"10.1145\/3637528.3671456","type":"proceedings-article","created":{"date-parts":[[2024,8,25]],"date-time":"2024-08-25T04:54:55Z","timestamp":1724561695000},"page":"6545-6554","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":11,"title":["Graph Intelligence with Large Language Models and Prompt Learning"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-6362-4385","authenticated-orcid":false,"given":"Jia","family":"Li","sequence":"first","affiliation":[{"name":"The Hong Kong University of Science and Technology (Guangzhou), Guangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2224-4634","authenticated-orcid":false,"given":"Xiangguo","family":"Sun","sequence":"additional","affiliation":[{"name":"The Chinese University of Hong Kong, HongKong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1324-5819","authenticated-orcid":false,"given":"Yuhan","family":"Li","sequence":"additional","affiliation":[{"name":"The Hong Kong University of Science and Technology (Guangzhou), Guangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6750-9002","authenticated-orcid":false,"given":"Zhixun","family":"Li","sequence":"additional","affiliation":[{"name":"The Chinese University of Hong Kong, HongKong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4673-2587","authenticated-orcid":false,"given":"Hong","family":"Cheng","sequence":"additional","affiliation":[{"name":"The Chinese University of Hong Kong, HongKong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9738-827X","authenticated-orcid":false,"given":"Jeffrey Xu","family":"Yu","sequence":"additional","affiliation":[{"name":"The Chinese University of Hong Kong, HongKong, China"}]}],"member":"320","published-online":{"date-parts":[[2024,8,24]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"crossref","unstructured":"Iz Beltagy Kyle Lo et al. 2019. SciBERT: A pretrained language model for scientific text. arXiv preprint arXiv:1903.10676 (2019).","DOI":"10.18653\/v1\/D19-1371"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"crossref","unstructured":"William Brannon Suyash Fulay et al. 2023. ConGraT: Self-Supervised Contrastive Pretraining for Joint Graph and Text Embeddings. arXiv preprint arXiv:2305.14321 (2023).","DOI":"10.18653\/v1\/2024.textgraphs-1.2"},{"key":"e_1_3_2_1_3_1","first-page":"1877","article-title":"Language models are few-shot learners","volume":"33","author":"Brown Tom","year":"2020","unstructured":"Tom Brown, Benjamin Mann, et al. 2020. Language models are few-shot learners. NeurIPS, Vol. 33 (2020), 1877--1901.","journal-title":"NeurIPS"},{"key":"e_1_3_2_1_4_1","unstructured":"He Cao Zijing Liu et al. 2023. InstructMol: Multi-Modal Integration for Building a Versatile and Reliable Molecular Assistant in Drug Discovery. arXiv preprint arXiv:2311.16208 (2023)."},{"key":"e_1_3_2_1_5_1","unstructured":"Ziwei Chai Tianjie Zhang et al. 2023. GraphLLM: Boosting Graph Reasoning Ability of Large Language Model. arXiv preprint arXiv:2310.05845 (2023)."},{"key":"e_1_3_2_1_6_1","unstructured":"Shantanu Chandra Pushkar Mishra et al. 2020. Graph-based modeling of online communities for fake news detection. arXiv preprint arXiv:2008.06274 (2020)."},{"key":"e_1_3_2_1_7_1","volume-title":"ULTRA-DP: Unifying Graph Pre-training with Multi-task Graph Dual Prompt. arXiv preprint","author":"Chen Mouxiang","year":"2023","unstructured":"Mouxiang Chen, Zemin Liu, Chenghao Liu, Jundong Li, Qiheng Mao, and Jianling Sun. 2023. ULTRA-DP: Unifying Graph Pre-training with Multi-task Graph Dual Prompt. arXiv preprint (2023)."},{"key":"e_1_3_2_1_8_1","volume-title":"GraphWiz: An Instruction-Following Language Model for Graph Problems. arXiv preprint arXiv:2402.16029","author":"Chen Nuo","year":"2024","unstructured":"Nuo Chen, Yuhan Li, Jianheng Tang, and Jia Li. 2024. GraphWiz: An Instruction-Following Language Model for Graph Problems. arXiv preprint arXiv:2402.16029 (2024)."},{"key":"e_1_3_2_1_9_1","unstructured":"Zhikai Chen Haitao Mao et al. 2023. Exploring the potential of large language models (llms) in learning on graphs. arXiv preprint arXiv:2307.03393 (2023)."},{"key":"e_1_3_2_1_10_1","unstructured":"Zhikai Chen Haitao Mao et al. 2024. Label-free Node Classification on Graphs with Large Language Models (LLMS). In ICLR."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"crossref","unstructured":"Jiashun Cheng Man Li et al. 2023. Wiener Graph Deconvolutional Network Improves Graph Self-Supervised Learning. In AAAI. 7131--7139.","DOI":"10.1609\/aaai.v37i6.25870"},{"key":"e_1_3_2_1_12_1","unstructured":"Eli Chien Wei-Cheng Chang et al. 2021. Node feature extraction by self-supervised multi-scale neighborhood prediction. arXiv preprint arXiv:2111.00064 (2021)."},{"key":"e_1_3_2_1_13_1","volume-title":"Palm: Scaling language modeling with pathways. arXiv preprint arXiv:2204.02311","author":"Chowdhery Aakanksha","year":"2022","unstructured":"Aakanksha Chowdhery, Sharan Narang, et al. 2022. Palm: Scaling language modeling with pathways. arXiv preprint arXiv:2204.02311 (2022)."},{"key":"e_1_3_2_1_14_1","unstructured":"Hyung Won Chung Le Hou et al. 2022. Scaling instruction-finetuned language models. arXiv preprint arXiv:2210.11416 (2022)."},{"key":"e_1_3_2_1_15_1","unstructured":"Debarati Das Ishaan Gupta et al. 2023. Which Modality should I use--Text Motif or Image?: Understanding Graphs with Large Language Models. arXiv preprint arXiv:2311.09862 (2023)."},{"key":"e_1_3_2_1_16_1","volume-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv preprint","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv preprint (2019)."},{"key":"e_1_3_2_1_17_1","unstructured":"Qingxiu Dong Lei Li et al. 2022. A survey for in-context learning. arXiv preprint arXiv:2301.00234 (2022)."},{"key":"e_1_3_2_1_18_1","volume-title":"Simteg: A frustratingly simple approach improves textual graph learning. arXiv preprint arXiv:2308.02565","author":"Duan Keyu","year":"2023","unstructured":"Keyu Duan, Qian Liu, et al. 2023. Simteg: A frustratingly simple approach improves textual graph learning. arXiv preprint arXiv:2308.02565 (2023)."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"crossref","unstructured":"Carl Edwards ChengXiang Zhai et al. 2021. Text2mol: Cross-modal molecule retrieval with natural language queries. In EMNLP. 595--607.","DOI":"10.18653\/v1\/2021.emnlp-main.47"},{"key":"e_1_3_2_1_20_1","volume-title":"Prompt tuning for graph neural networks. arXiv preprint","author":"Fang Taoran","year":"2022","unstructured":"Taoran Fang, Yunchao Zhang, Yang Yang, and Chunping Wang. 2022. Prompt tuning for graph neural networks. arXiv preprint (2022)."},{"key":"e_1_3_2_1_21_1","unstructured":"Taoran Fang Yunchao Zhang Yang Yang Chunping Wang and Lei Chen. 2023. Universal Prompt Tuning for Graph Neural Networks. In NeurIPS."},{"key":"e_1_3_2_1_22_1","volume-title":"Graph: Encoding Graphs for Large Language Models. arXiv preprint arXiv:2310.04560","author":"Fatemi Bahare","year":"2023","unstructured":"Bahare Fatemi, Jonathan Halcrow, et al. 2023. Talk like a Graph: Encoding Graphs for Large Language Models. arXiv preprint arXiv:2310.04560 (2023)."},{"key":"e_1_3_2_1_23_1","unstructured":"Chelsea Finn Pieter Abbeel and Sergey Levine. 2017. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks. In ICML."},{"key":"e_1_3_2_1_24_1","unstructured":"Tianyu Gao Adam Fisch and Danqi Chen. 2021. Making Pre-Trained Language Models Better Few-Shot Learners. In ACL. 3816--3830."},{"key":"e_1_3_2_1_25_1","volume-title":"Enhancing Graph Neural Networks with Structure-Based Prompt. arXiv preprint","author":"Ge Qingqing","year":"2023","unstructured":"Qingqing Ge, Zeyuan Zhao, Yiding Liu, Anfeng Cheng, Xiang Li, Shuaiqiang Wang, and Dawei Yin. 2023. Enhancing Graph Neural Networks with Structure-Based Prompt. arXiv preprint (2023)."},{"key":"e_1_3_2_1_26_1","volume-title":"Prompt Tuning for Multi-View Graph Contrastive Learning. arXiv preprint","author":"Gong Chenghua","year":"2023","unstructured":"Chenghua Gong, Xiang Li, Jianxiang Yu, Cheng Yao, Jiaqi Tan, Chengcheng Yu, and Dawei Yin. 2023. Prompt Tuning for Multi-View Graph Contrastive Learning. arXiv preprint (2023)."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"crossref","unstructured":"Aditya Grover and Jure Leskovec. 2016. node2vec: Scalable feature learning for networks. In KDD. 855--864.","DOI":"10.1145\/2939672.2939754"},{"key":"e_1_3_2_1_28_1","unstructured":"Jiayan Guo Lun Du et al. 2023. GPT4Graph: Can Large Language Models Understand Graph Structured Data? An Empirical Evaluation and Benchmarking. arXiv preprint arXiv:2305.15066 (2023)."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"crossref","unstructured":"Yuxin Guo Cheng Yang Yuluo Chen Jixi Liu Chuan Shi and Junping Du. 2023. A Data-Centric Framework to Endow Graph Neural Networks with Out-of-Distribution Detection Ability. In KDD. 638--648.","DOI":"10.1145\/3580305.3599244"},{"key":"e_1_3_2_1_30_1","volume-title":"NeurIPS","volume":"30","author":"Hamilton Will","year":"2017","unstructured":"Will Hamilton, Zhitao Ying, et al. 2017. Inductive representation learning on large graphs. NeurIPS, Vol. 30 (2017)."},{"key":"e_1_3_2_1_31_1","volume-title":"Features: LLM-Based Features for Text-Attributed Graphs. In ICLR.","author":"He Xiaoxin","year":"2024","unstructured":"Xiaoxin He, Xavier Bresson, et al. 2024. Explanations as Features: LLM-Based Features for Text-Attributed Graphs. In ICLR."},{"key":"e_1_3_2_1_32_1","volume-title":"Graphmae: Self-supervised masked graph autoencoders. In SIGKDD. 594--604.","author":"Hou Zhenyu","year":"2022","unstructured":"Zhenyu Hou, Xiao Liu, et al. 2022. Graphmae: Self-supervised masked graph autoencoders. In SIGKDD. 594--604."},{"key":"e_1_3_2_1_33_1","volume-title":"Beyond Text: A Deep Dive into Large Language Models' Ability on Understanding Graph Data. arXiv preprint arXiv:2310.04944","author":"Hu Yuntong","year":"2023","unstructured":"Yuntong Hu, Zheng Zhang, et al. 2023. Beyond Text: A Deep Dive into Large Language Models' Ability on Understanding Graph Data. arXiv preprint arXiv:2310.04944 (2023)."},{"key":"e_1_3_2_1_34_1","unstructured":"Jin Huang Xingjian Zhang et al. 2023. Can LLMs effectively leverage graph structural information: when and why. arXiv preprint arXiv:2309.16595 (2023)."},{"key":"e_1_3_2_1_35_1","volume-title":"PRODIGY: Enabling In-context Learning Over Graphs. In NeurIPS.","author":"Huang Qian","year":"2023","unstructured":"Qian Huang, Hongyu Ren, Peng Chen, Gregor Kr?manc, Daniel Zeng, Percy Liang, and Jure Leskovec. 2023. PRODIGY: Enabling In-context Learning Over Graphs. In NeurIPS."},{"key":"e_1_3_2_1_36_1","unstructured":"Xuanwen Huang Kaiqiao Han et al. 2023. Prompt-based Node Feature Extractor for Few-shot Learning on Text-Attributed Graphs. arXiv preprint arXiv:2309.02848 (2023)."},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2021.3070843"},{"key":"e_1_3_2_1_38_1","unstructured":"Bowen Jin Wentao Zhang et al. 2023. Learning Multiplex Embeddings on Text-rich Networks with One Text Encoder. arXiv preprint arXiv:2310.06684 (2023)."},{"key":"e_1_3_2_1_39_1","volume-title":"Patton: Language Model Pretraining on Text-Rich Networks. arXiv preprint arXiv:2305.12268","author":"Jin Bowen","year":"2023","unstructured":"Bowen Jin, Wentao Zhang, et al. 2023. Patton: Language Model Pretraining on Text-Rich Networks. arXiv preprint arXiv:2305.12268 (2023)."},{"key":"e_1_3_2_1_40_1","volume-title":"Proceedings of NAACL-HLT. 4171--4186","author":"Ming-Wei Chang Jacob Devlin","unstructured":"Jacob Devlin Ming-Wei Chang Kenton et al. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of NAACL-HLT. 4171--4186."},{"key":"e_1_3_2_1_41_1","unstructured":"Thomas N Kipf and Max Welling. 2016. Semi-Supervised Classification with Graph Convolutional Networks. In ICLR."},{"key":"e_1_3_2_1_42_1","unstructured":"Junnan Li Dongxu Li et al. 2023. Blip-2: Bootstrapping language-image pre-training with frozen image encoders and large language models. arXiv preprint arXiv:2301.12597 (2023)."},{"key":"e_1_3_2_1_43_1","volume-title":"GRENADE: Graph-Centric Language Model for Self-Supervised Representation Learning on Text-Attributed Graphs. arXiv preprint arXiv:2310.15109","author":"Li Yichuan","year":"2023","unstructured":"Yichuan Li, Kaize Ding, et al. 2023. GRENADE: Graph-Centric Language Model for Self-Supervised Representation Learning on Text-Attributed Graphs. arXiv preprint arXiv:2310.15109 (2023)."},{"key":"e_1_3_2_1_44_1","volume-title":"A survey of graph meets large language model: Progress and future directions. arXiv preprint","author":"Li Yuhan","year":"2023","unstructured":"Yuhan Li, Zhixun Li, Peisong Wang, Jia Li, Xiangguo Sun, Hong Cheng, and Jeffrey Xu Yu. 2023. A survey of graph meets large language model: Progress and future directions. arXiv preprint (2023)."},{"key":"e_1_3_2_1_45_1","volume-title":"Community question answering entity linking via leveraging auxiliary data. arXiv preprint arXiv:2205.11917","author":"Li Yuhan","year":"2022","unstructured":"Yuhan Li, Wei Shen, Jianbo Gao, and Yadong Wang. 2022. Community question answering entity linking via leveraging auxiliary data. arXiv preprint arXiv:2205.11917 (2022)."},{"key":"e_1_3_2_1_46_1","volume-title":"Jeffrey Xu Yu, and Jia Li","author":"Li Yuhan","year":"2024","unstructured":"Yuhan Li, Peisong Wang, Zhixun Li, Jeffrey Xu Yu, and Jia Li. 2024. ZeroG: Investigating Cross-dataset Zero-shot Transferability in Graphs. arXiv preprint arXiv:2402.11235 (2024)."},{"key":"e_1_3_2_1_47_1","volume-title":"Evaluating large language models on graphs: Performance insights and comparative analysis. arXiv preprint arXiv:2308.11224","author":"Liu Chang","year":"2023","unstructured":"Chang Liu and Bo Wu. 2023. Evaluating large language models on graphs: Performance insights and comparative analysis. arXiv preprint arXiv:2308.11224 (2023)."},{"key":"e_1_3_2_1_48_1","unstructured":"Hao Liu Jiarui Feng et al. 2024. One for All: Towards Training One Graph Model for All Classification Tasks. In ICLR."},{"key":"e_1_3_2_1_49_1","unstructured":"Jiawei Liu Cheng Yang et al. 2023. Towards Graph Foundation Models: A Survey and Beyond. arXiv preprint arXiv:2310.11829 (2023)."},{"key":"e_1_3_2_1_50_1","unstructured":"Pengfei Liu Yiming Ren et al. 2023. GIT-Mol: A Multi-modal Large Language Model for Molecular Science with Graph Image and Text. arXiv preprint arXiv:2308.06911 (2023)."},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1145\/3560815"},{"key":"e_1_3_2_1_52_1","unstructured":"Shengchao Liu Weili Nie et al. 2022. Multi-modal molecule structure-text model for text-based retrieval and editing. arXiv preprint arXiv:2212.10789 (2022)."},{"key":"e_1_3_2_1_53_1","volume-title":"Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu, Myle Ott, et al. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019)."},{"key":"e_1_3_2_1_54_1","unstructured":"Zhiyuan Liu Sihang Li et al. 2023. MolCA: Molecular Graph-Language Modeling with Cross-Modal Projector and Uni-Modal Adapter. arXiv preprint arXiv:2310.12798 (2023)."},{"key":"e_1_3_2_1_55_1","volume-title":"Graphprompt: Unifying Pre-Training and Downstream Tasks for Graph Neural Networks. In The Web Conference. 417--428","author":"Liu Zemin","year":"2023","unstructured":"Zemin Liu, Xingtong Yu, Yuan Fang, and Xinming Zhang. 2023 d. Graphprompt: Unifying Pre-Training and Downstream Tasks for Graph Neural Networks. In The Web Conference. 417--428."},{"key":"e_1_3_2_1_56_1","volume-title":"Chawla","author":"Ma Yihong","year":"2023","unstructured":"Yihong Ma, Ning Yan, Jiayu Li, Masood Mortazavi, and Nitesh V. Chawla. 2023. HetGPT: Harnessing the Power of Prompt Tuning in Pre-Trained Heterogeneous Graph Neural Networks. arXiv preprint (2023)."},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"crossref","unstructured":"Costas Mavromatis Vassilis N Ioannidis et al. 2023. Train Your Own GNN Teacher: Graph-Aware Distillation on Textual Graphs. arXiv preprint arXiv:2304.10668 (2023).","DOI":"10.1007\/978-3-031-43418-1_10"},{"key":"e_1_3_2_1_58_1","volume-title":"NeurIPS","volume":"26","author":"Mikolov Tomas","year":"2013","unstructured":"Tomas Mikolov, Ilya Sutskever, et al. 2013. Distributed representations of words and phrases and their compositionality. NeurIPS, Vol. 26 (2013)."},{"key":"e_1_3_2_1_59_1","unstructured":"Aaron van den Oord Yazhe Li et al. 2018. Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748 (2018)."},{"key":"e_1_3_2_1_61_1","unstructured":"Chen Qian Huayi Tang et al. 2023. Can large language models empower molecular property prediction? arXiv preprint arXiv:2307.07443 (2023)."},{"key":"e_1_3_2_1_62_1","unstructured":"Yijian Qin Xin Wang et al. 2023. Disentangled Representation Learning with Large Language Models for Text-Attributed Graphs. arXiv preprint arXiv:2310.18152 (2023)."},{"key":"e_1_3_2_1_63_1","volume-title":"Jong Wook Kim, et al","author":"Radford Alec","year":"2021","unstructured":"Alec Radford, Jong Wook Kim, et al. 2021. Learning transferable visual models from natural language supervision. In ICLR. 8748--8763."},{"key":"e_1_3_2_1_64_1","unstructured":"Xubin Ren Wei Wei et al. 2023. Representation Learning with Large Language Models for Recommendation. arXiv preprint arXiv:2310.15950 (2023)."},{"key":"e_1_3_2_1_65_1","first-page":"513","article-title":"Term-weighting approaches in automatic text retrieval","volume":"24","author":"Salton Gerard","year":"1988","unstructured":"Gerard Salton and Christopher Buckley. 1988. Term-weighting approaches in automatic text retrieval. IPM, Vol. 24, 5 (1988), 513--523.","journal-title":"IPM"},{"key":"e_1_3_2_1_66_1","doi-asserted-by":"crossref","unstructured":"Prithviraj Sen Galileo Namata et al. 2008. Collective classification in network data. AI magazine Vol. 29 3 (2008) 93--93.","DOI":"10.1609\/aimag.v29i3.2157"},{"key":"e_1_3_2_1_67_1","volume-title":"Md Abdullah Al Mamun, et al","author":"Shayegani Erfan","year":"2023","unstructured":"Erfan Shayegani, Md Abdullah Al Mamun, et al. 2023. Survey of Vulnerabilities in Large Language Models Revealed by Adversarial Attacks. arXiv preprint arXiv:2310.10844 (2023)."},{"key":"e_1_3_2_1_68_1","first-page":"2556","article-title":"Entity linking meets deep learning: Techniques and solutions","volume":"35","author":"Shen Wei","year":"2021","unstructured":"Wei Shen, Yuhan Li, Yinan Liu, Jiawei Han, Jianyong Wang, and Xiaojie Yuan. 2021. Entity linking meets deep learning: Techniques and solutions. IEEE Transactions on Knowledge and Data Engineering, Vol. 35, 3 (2021), 2556--2578.","journal-title":"IEEE Transactions on Knowledge and Data Engineering"},{"key":"e_1_3_2_1_69_1","unstructured":"Yaorui Shi An Zhang et al. 2023. ReLM: Leveraging Language Models for Enhanced Chemical Reaction Prediction. arXiv preprint arXiv:2310.13590 (2023)."},{"key":"e_1_3_2_1_70_1","volume-title":"Deep Prompt Tuning for Graph Transformers. arXiv preprint","author":"Shirkavand Reza","year":"2023","unstructured":"Reza Shirkavand and Heng Huang. 2023. Deep Prompt Tuning for Graph Transformers. arXiv preprint (2023)."},{"key":"e_1_3_2_1_71_1","unstructured":"Bing Su Dazhao Du et al. 2022. A molecular multimodal foundation model associating molecule graphs with natural language. arXiv preprint arXiv:2209.05481 (2022)."},{"key":"e_1_3_2_1_72_1","unstructured":"Fan-Yun Sun Jordan Hoffmann Vikas Verma and Jian Tang. 2020. InfoGraph: Unsupervised and Semi-supervised Graph-Level Representation Learning via Mutual Information Maximization. In ICLR."},{"key":"e_1_3_2_1_73_1","volume-title":"Gppt: Graph pre-training and prompt tuning to generalize graph neural networks. In SIGKDD. 1717--1727.","author":"Sun Mingchen","year":"2022","unstructured":"Mingchen Sun, Kaixiong Zhou, et al. 2022. Gppt: Graph pre-training and prompt tuning to generalize graph neural networks. In SIGKDD. 1717--1727."},{"key":"e_1_3_2_1_74_1","volume-title":"One: Multi-Task Prompting for Graph Neural Networks. In SIGKDD. 2120--2131.","author":"Sun Xiangguo","year":"2023","unstructured":"Xiangguo Sun, Hong Cheng, et al. 2023. All in One: Multi-Task Prompting for Graph Neural Networks. In SIGKDD. 2120--2131."},{"key":"e_1_3_2_1_75_1","unstructured":"Xiangguo Sun Hong Cheng et al. 2023. Self-supervised hypergraph representation learning for sociological analysis. TKDE (2023)."},{"key":"e_1_3_2_1_76_1","doi-asserted-by":"crossref","unstructured":"Xiangguo Sun Hongzhi Yin Bo Liu Hongxu Chen Jiuxin Cao Yingxia Shao and Nguyen Quoc Viet Hung. 2021. Heterogeneous Hypergraph Embedding for Graph Classification. In WSDM. 725--733.","DOI":"10.1145\/3437963.3441835"},{"key":"e_1_3_2_1_77_1","volume-title":"Graph prompt learning: A comprehensive survey and beyond. arXiv preprint","author":"Sun Xiangguo","year":"2023","unstructured":"Xiangguo Sun, Jiawen Zhang, Xixi Wu, Hong Cheng, Yun Xiong, and Jia Li. 2023. Graph prompt learning: A comprehensive survey and beyond. arXiv preprint (2023)."},{"key":"e_1_3_2_1_78_1","doi-asserted-by":"crossref","unstructured":"Qiaoyu Tan Ninghao Liu et al. 2023. S2GAE: Self-Supervised Graph Autoencoders are Generalizable Learners with Graph Masking. In WSDM. 787--795.","DOI":"10.1145\/3539597.3570404"},{"key":"e_1_3_2_1_79_1","unstructured":"Yanchao Tan Zihao Zhou et al. 2023. WalkLM: A Uniform Language Model Fine-tuning Framework for Attributed Graph Embedding. In NeurIPS."},{"key":"e_1_3_2_1_80_1","unstructured":"Zhen Tan Ruocheng Guo Kaize Ding and Huan Liu. 2023. Virtual Node Tuning for Few-shot Node Classification. In KDD. 2177--2188."},{"key":"e_1_3_2_1_81_1","unstructured":"Jiabin Tang Yuhao Yang et al. 2023. GraphGPT: Graph Instruction Tuning for Large Language Models. arXiv preprint arXiv:2310.13023 (2023)."},{"key":"e_1_3_2_1_82_1","volume-title":"Galactica: A large language model for science. arXiv preprint arXiv:2211.09085","author":"Taylor Ross","year":"2022","unstructured":"Ross Taylor, Marcin Kardas, et al. 2022. Galactica: A large language model for science. arXiv preprint arXiv:2211.09085 (2022)."},{"key":"e_1_3_2_1_83_1","volume-title":"Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971","author":"Touvron Hugo","year":"2023","unstructured":"Hugo Touvron, Thibaut Lavril, et al. 2023. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023)."},{"key":"e_1_3_2_1_84_1","volume-title":"NeurIPS","volume":"30","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, et al. 2017. Attention is all you need. NeurIPS, Vol. 30 (2017)."},{"key":"e_1_3_2_1_85_1","unstructured":"Petar Velickovic Guillem Cucurull et al. 2018. GRAPH ATTENTION NETWORKS. stat Vol. 1050 (2018) 4."},{"key":"e_1_3_2_1_86_1","unstructured":"Heng Wang Shangbin Feng et al. 2023. Can Language Models Solve Graph Problems in Natural Language? arXiv preprint arXiv:2305.10037 (2023)."},{"key":"e_1_3_2_1_87_1","unstructured":"Haishuai Wang Yang Gao et al. 2023. Graph Neural Architecture Search with GPT-4. arXiv preprint arXiv:2310.01436 (2023)."},{"key":"e_1_3_2_1_88_1","first-page":"24824","article-title":"Chain-of-thought prompting elicits reasoning in large language models","volume":"35","author":"Wei Jason","year":"2022","unstructured":"Jason Wei, Xuezhi Wang, et al. 2022. Chain-of-thought prompting elicits reasoning in large language models. NeurIPS, Vol. 35 (2022), 24824--24837.","journal-title":"NeurIPS"},{"key":"e_1_3_2_1_89_1","unstructured":"Wei Wei Xubin Ren et al. 2023. LLMRec: Large Language Models with Graph Augmentation for Recommendation. arXiv preprint arXiv:2311.00423 (2023)."},{"key":"e_1_3_2_1_90_1","volume-title":"Prompt Tuning on Graph-augmented Low-resource Text Classification. arXiv preprint arXiv:2307.10230","author":"Wen Zhihao","year":"2023","unstructured":"Zhihao Wen and Yuan Fang. 2023. Prompt Tuning on Graph-augmented Low-resource Text Classification. arXiv preprint arXiv:2307.10230 (2023)."},{"key":"e_1_3_2_1_91_1","doi-asserted-by":"crossref","unstructured":"Zhenqin Wu Bharath Ramsundar et al. 2018. MoleculeNet: a benchmark for molecular machine learning. Chemical science Vol. 9 2 (2018) 513--530.","DOI":"10.1039\/C7SC02664A"},{"key":"e_1_3_2_1_92_1","doi-asserted-by":"crossref","unstructured":"Han Xie Da Zheng et al. 2023. Graph-Aware Language Model Pre-Training on a Large Graph Corpus Can Help Multiple Graph Applications. arXiv preprint arXiv:2306.02592 (2023).","DOI":"10.1145\/3580305.3599833"},{"key":"e_1_3_2_1_93_1","unstructured":"Rui Xue Xipeng Shen et al. 2023. Efficient Large Language Models Fine-Tuning On Graphs. arXiv preprint arXiv:2312.04737 (2023)."},{"key":"e_1_3_2_1_94_1","unstructured":"Jingfeng Yang Hongye Jin et al. 2023. Harnessing the power of llms in practice: A survey on chatgpt and beyond. arXiv preprint arXiv:2304.13712 (2023)."},{"key":"e_1_3_2_1_95_1","first-page":"28798","article-title":"GraphFormers: GNN-nested transformers for representation learning on textual graph","volume":"34","author":"Yang Junhan","year":"2021","unstructured":"Junhan Yang, Zheng Liu, et al. 2021. GraphFormers: GNN-nested transformers for representation learning on textual graph. NeurIPS, Vol. 34 (2021), 28798--28810.","journal-title":"NeurIPS"},{"key":"e_1_3_2_1_96_1","volume-title":"Natural language is all a graph needs. arXiv preprint arXiv:2308.07134","author":"Ye Ruosong","year":"2023","unstructured":"Ruosong Ye, Caiqi Zhang, Runhui Wang, Shuyuan Xu, and Yongfeng Zhang. 2023. Natural language is all a graph needs. arXiv preprint arXiv:2308.07134 (2023)."},{"key":"e_1_3_2_1_97_1","unstructured":"Yuning You Tianlong Chen et al. 2020. Graph contrastive learning with augmentations. NeurIPS (2020) 5812--5823."},{"key":"e_1_3_2_1_98_1","unstructured":"Jianxiang Yu Yuxiang Ren et al. 2023. Empower Text-Attributed Graphs Learning with Large Language Models (LLMs). arXiv preprint arXiv:2310.09872 (2023)."},{"key":"e_1_3_2_1_99_1","unstructured":"Aohan Zeng Xiao Liu et al. 2022. Glm-130b: An open bilingual pre-trained model. arXiv preprint arXiv:2210.02414 (2022)."},{"key":"e_1_3_2_1_100_1","unstructured":"Muhan Zhang and Yixin Chen. 2018. Link Prediction Based on Graph Neural Networks. In NeurIPS."},{"key":"e_1_3_2_1_101_1","doi-asserted-by":"crossref","unstructured":"Zeyang Zhang Xin Wang et al. 2023. LLM4DyG: Can Large Language Models Solve Problems on Dynamic Graphs? arXiv preprint arXiv:2310.17110 (2023).","DOI":"10.1145\/3637528.3671709"},{"key":"e_1_3_2_1_102_1","volume-title":"GIMLET: A Unified Graph-Text Model for Instruction-Based Molecule Zero-Shot Learning. arXiv preprint arXiv:2306.13089","author":"Zhao Haiteng","year":"2023","unstructured":"Haiteng Zhao, Shengchao Liu, et al. 2023. GIMLET: A Unified Graph-Text Model for Instruction-Based Molecule Zero-Shot Learning. arXiv preprint arXiv:2306.13089 (2023)."},{"key":"e_1_3_2_1_103_1","unstructured":"Jianan Zhao Meng Qu et al. 2022. Learning on large-scale text-attributed graphs via variational inference. arXiv preprint arXiv:2210.14709 (2022)."},{"key":"e_1_3_2_1_104_1","volume-title":"Graphtext: Graph reasoning in text space. arXiv preprint arXiv:2310.01089","author":"Zhao Jianan","year":"2023","unstructured":"Jianan Zhao, Le Zhuo, et al. 2023. Graphtext: Graph reasoning in text space. arXiv preprint arXiv:2310.01089 (2023)."},{"key":"e_1_3_2_1_105_1","unstructured":"Wayne Xin Zhao Kun Zhou et al. 2023. A survey of large language models. arXiv preprint arXiv:2303.18223 (2023)."},{"key":"e_1_3_2_1_106_1","doi-asserted-by":"crossref","unstructured":"Jing Zhu Xiang Song et al. 2023. TouchUp-G: Improving Feature Representation through Graph-Centric Finetuning. arXiv preprint arXiv:2309.13885 (2023).","DOI":"10.1145\/3626772.3657978"},{"key":"e_1_3_2_1_107_1","volume-title":"SGL-PT: A Strong Graph Learner with Graph Prompt Tuning. arXiv preprint","author":"Zhu Yun","year":"2023","unstructured":"Yun Zhu, Jianhao Guo, and Siliang Tang. 2023. SGL-PT: A Strong Graph Learner with Graph Prompt Tuning. arXiv preprint (2023)."},{"key":"e_1_3_2_1_108_1","volume-title":"GraphControl: Adding Conditional Control to Universal Graph Pre-trained Models for Graph Domain Transfer Learning. arXiv preprint","author":"Zhu Yun","year":"2023","unstructured":"Yun Zhu, Yaoke Wang, Haizhou Shi, Zhenshuo Zhang, and Siliang Tang. 2023. GraphControl: Adding Conditional Control to Universal Graph Pre-trained Models for Graph Domain Transfer Learning. arXiv preprint (2023)."},{"key":"e_1_3_2_1_109_1","doi-asserted-by":"crossref","unstructured":"Yanqiao Zhu Yichen Xu et al. 2021. Graph contrastive learning with adaptive augmentation. In WWW. 2069--2080.","DOI":"10.1145\/3442381.3449802"},{"key":"e_1_3_2_1_110_1","doi-asserted-by":"crossref","unstructured":"Tao Zou Le Yu et al. 2023. Pretraining Language Models with Text-Attributed Heterogeneous Graphs. arXiv preprint arXiv:2310.12580 (2023).","DOI":"10.18653\/v1\/2023.findings-emnlp.692"}],"event":{"name":"KDD '24: The 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining","location":"Barcelona Spain","acronym":"KDD '24","sponsor":["SIGMOD ACM Special Interest Group on Management of Data","SIGKDD ACM Special Interest Group on Knowledge Discovery in Data"]},"container-title":["Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3637528.3671456","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3637528.3671456","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T00:03:26Z","timestamp":1750291406000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3637528.3671456"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,8,24]]},"references-count":109,"alternative-id":["10.1145\/3637528.3671456","10.1145\/3637528"],"URL":"https:\/\/doi.org\/10.1145\/3637528.3671456","relation":{},"subject":[],"published":{"date-parts":[[2024,8,24]]},"assertion":[{"value":"2024-08-24","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}