{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,10]],"date-time":"2026-04-10T10:04:31Z","timestamp":1775815471826,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":96,"publisher":"ACM","license":[{"start":{"date-parts":[[2025,4,22]],"date-time":"2025-04-22T00:00:00Z","timestamp":1745280000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"National Key Research and Devel- opment Plan of China","award":["2023YFB4502305"],"award-info":[{"award-number":["2023YFB4502305"]}]},{"name":"Ant Group through Ant Research Intern Program"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,4,22]]},"DOI":"10.1145\/3696410.3714801","type":"proceedings-article","created":{"date-parts":[[2025,5,5]],"date-time":"2025-05-05T16:42:02Z","timestamp":1746463322000},"page":"2183-2197","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":9,"title":["GraphCLIP: Enhancing Transferability in Graph Foundation Models for Text-Attributed Graphs"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8950-383X","authenticated-orcid":false,"given":"Yun","family":"Zhu","sequence":"first","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8431-3703","authenticated-orcid":false,"given":"Haizhou","family":"Shi","sequence":"additional","affiliation":[{"name":"Rutgers University, New Brunswick, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0001-6804-542X","authenticated-orcid":false,"given":"Xiaotang","family":"Wang","sequence":"additional","affiliation":[{"name":"Huazhong University of Science and Technology, Wuhan, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3440-9675","authenticated-orcid":false,"given":"Yongchao","family":"Liu","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-0957-5007","authenticated-orcid":false,"given":"Yaoke","family":"Wang","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0984-8740","authenticated-orcid":false,"given":"Boci","family":"Peng","sequence":"additional","affiliation":[{"name":"Peking University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-3472-6102","authenticated-orcid":false,"given":"Chuntao","family":"Hong","sequence":"additional","affiliation":[{"name":"Ant Group, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7356-9711","authenticated-orcid":false,"given":"Siliang","family":"Tang","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]}],"member":"320","published-online":{"date-parts":[[2025,4,22]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Invariant risk minimization. arXiv preprint arXiv:1907.02893","author":"Arjovsky Martin","year":"2019","unstructured":"Martin Arjovsky, L\u00e9on Bottou, Ishaan Gulrajani, and David Lopez-Paz. 2019. Invariant risk minimization. arXiv preprint arXiv:1907.02893 (2019)."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1007\/3-540-45848-4_59"},{"key":"e_1_3_2_1_3_1","volume-title":"Congrat: Self-supervised contrastive pretraining for joint graph and text embeddings. arXiv preprint arXiv:2305.14321","author":"Brannon William","year":"2023","unstructured":"William Brannon, Suyash Fulay, Hang Jiang, Wonjune Kang, Brandon Roy, Jad Kabbara, and Deb Roy. 2023. Congrat: Self-supervised contrastive pretraining for joint graph and text embeddings. arXiv preprint arXiv:2305.14321 (2023)."},{"key":"e_1_3_2_1_4_1","volume-title":"LLaGA: Large Language and Graph Assistant. In Forty-first International Conference on Machine Learning. https:\/\/openreview.net\/forum?id=B48Pzc4oKi","author":"Chen Runjin","year":"2024","unstructured":"Runjin Chen, Tong Zhao, AJAY KUMAR JAISWAL, Neil Shah, and Zhangyang Wang. 2024. LLaGA: Large Language and Graph Assistant. In Forty-first International Conference on Machine Learning. https:\/\/openreview.net\/forum?id=B48Pzc4oKi"},{"key":"e_1_3_2_1_5_1","unstructured":"Zhikai Chen Haitao Mao Jingzhe Liu Yu Song Bingheng Li Wei Jin Bahare Fatemi Anton Tsitsulin Bryan Perozzi Hui Liu et al. 2024. Text-space Graph Foundation Models: Comprehensive Benchmarks and New Insights. arXiv preprint arXiv:2406.10727 (2024)."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/3292500.3330925"},{"key":"e_1_3_2_1_7_1","volume-title":"Graph Neural Networks with Learnable Structural and Positional Representations. In International Conference on Learning Representations.","author":"Dwivedi Vijay Prakash","unstructured":"Vijay Prakash Dwivedi, Anh Tuan Luu, Thomas Laurent, Yoshua Bengio, and Xavier Bresson. [n. d.]. Graph Neural Networks with Learnable Structural and Positional Representations. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_8_1","volume-title":"Prompt tuning for graph neural networks. arXiv preprint arXiv:2209.15240","author":"Fang Taoran","year":"2022","unstructured":"Taoran Fang, Yunchao Zhang, Yang Yang, and Chunping Wang. 2022. Prompt tuning for graph neural networks. arXiv preprint arXiv:2209.15240 (2022)."},{"key":"e_1_3_2_1_9_1","volume-title":"TAGLAS: An atlas of text-attributed graph datasets in the era of large graph and language models. arXiv preprint arXiv:2406.14683","author":"Feng Jiarui","year":"2024","unstructured":"Jiarui Feng, Hao Liu, Lecheng Kong, Yixin Chen, and Muhan Zhang. 2024. TAGLAS: An atlas of text-attributed graph datasets in the era of large graph and language models. arXiv preprint arXiv:2406.14683 (2024)."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1145\/276675.276685"},{"key":"e_1_3_2_1_11_1","volume-title":"Proc. of NeurIPS","author":"Grill Jean-Bastien","year":"2020","unstructured":"Jean-Bastien Grill, Florian Strub, Florent Altch\u00e9, Corentin Tallec, Pierre Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Guo, Mohammad Gheshlaghi Azar, et al. 2020. Bootstrap your own latent-a new approach to self-supervised learning. Proc. of NeurIPS (2020)."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"e_1_3_2_1_13_1","volume-title":"Proc. of ICLR.","author":"He Pengcheng","year":"2020","unstructured":"Pengcheng He, Xiaodong Liu, Jianfeng Gao, and Weizhu Chen. 2020. DEBERTA: DECODING-ENHANCED BERT WITH DISENTANGLED ATTENTION. In Proc. of ICLR."},{"key":"e_1_3_2_1_14_1","volume-title":"Explanations as Features: LLM-Based Features for Text-Attributed Graphs. arXiv preprint arXiv:2305.19523","author":"He Xiaoxin","year":"2023","unstructured":"Xiaoxin He, Xavier Bresson, Thomas Laurent, and Bryan Hooi. 2023. Explanations as Features: LLM-Based Features for Text-Attributed Graphs. arXiv preprint arXiv:2305.19523 (2023)."},{"key":"e_1_3_2_1_15_1","volume-title":"Scaling laws for transfer. arXiv preprint arXiv:2102.01293","author":"Hernandez Danny","year":"2021","unstructured":"Danny Hernandez, Jared Kaplan, Tom Henighan, and Sam McCandlish. 2021. Scaling laws for transfer. arXiv preprint arXiv:2102.01293 (2021)."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3534678.3539321"},{"key":"e_1_3_2_1_17_1","volume-title":"Proc. of ICLR.","author":"Hu Edward J","year":"2021","unstructured":"Edward J Hu, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, Weizhu Chen, et al. 2021. LoRA: Low-Rank Adaptation of Large Language Models. In Proc. of ICLR."},{"key":"e_1_3_2_1_18_1","volume-title":"Proc. of NeurIPS","author":"Hu Weihua","year":"2020","unstructured":"Weihua Hu, Matthias Fey, Marinka Zitnik, Yuxiao Dong, Hongyu Ren, Bowen Liu, Michele Catasta, and Jure Leskovec. 2020. Open graph benchmark: Datasets for machine learning on graphs. Proc. of NeurIPS (2020)."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3589334.3645627"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1186\/s13321-020-00479-8"},{"key":"e_1_3_2_1_21_1","volume-title":"Large Language Models on Graphs: A Comprehensive Survey. arXiv preprint arXiv:2312.02783","author":"Jin Bowen","year":"2023","unstructured":"Bowen Jin, Gang Liu, Chi Han, Meng Jiang, Heng Ji, and Jiawei Han. 2023. Large Language Models on Graphs: A Comprehensive Survey. arXiv preprint arXiv:2312.02783 (2023)."},{"key":"e_1_3_2_1_22_1","volume-title":"Scaling laws for neural language models. arXiv preprint arXiv:2001.08361","author":"Kaplan Jared","year":"2020","unstructured":"Jared Kaplan, Sam McCandlish, Tom Henighan, Tom B Brown, Benjamin Chess, Rewon Child, Scott Gray, Alec Radford, Jeffrey Wu, and Dario Amodei. 2020. Scaling laws for neural language models. arXiv preprint arXiv:2001.08361 (2020)."},{"key":"e_1_3_2_1_23_1","volume-title":"Proc. of AACL.","author":"Ming-Wei Chang Jacob Devlin","year":"2019","unstructured":"Jacob Devlin Ming-Wei Chang Kenton and Lee Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proc. of AACL."},{"key":"e_1_3_2_1_24_1","volume-title":"Supervised contrastive learning. Advances in neural information processing systems 33","author":"Khosla Prannay","year":"2020","unstructured":"Prannay Khosla, Piotr Teterwak, ChenWang, Aaron Sarna, Yonglong Tian, Phillip Isola, Aaron Maschinot, Ce Liu, and Dilip Krishnan. 2020. Supervised contrastive learning. Advances in neural information processing systems 33 (2020), 18661--18673."},{"key":"e_1_3_2_1_25_1","volume-title":"Proc. of NeurIPS","author":"Kim Minseon","year":"2020","unstructured":"Minseon Kim, Jihoon Tack, and Sung Ju Hwang. 2020. Adversarial self-supervised contrastive learning. Proc. of NeurIPS (2020)."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1145\/3366423.3380052"},{"key":"e_1_3_2_1_27_1","volume-title":"Kipf and Max Welling","author":"Thomas","year":"2017","unstructured":"Thomas N. Kipf and Max Welling. 2017. Semi-Supervised Classification with Graph Convolutional Networks. In Proc. of ICLR."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00016"},{"key":"e_1_3_2_1_29_1","volume-title":"International conference on machine learning. PMLR, 12888--12900","author":"Li Junnan","year":"2022","unstructured":"Junnan Li, Dongxu Li, Caiming Xiong, and Steven Hoi. 2022. Blip: Bootstrapping language-image pre-training for unified vision-language understanding and generation. In International conference on machine learning. PMLR, 12888--12900."},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2024\/898"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/3637528.3671982"},{"key":"e_1_3_2_1_32_1","volume-title":"Victor Wai Kin Chan, and Jia Li","author":"Li Yuhan","year":"2024","unstructured":"Yuhan Li, Peisong Wang, Xiao Zhu, Aochuan Chen, Haiyun Jiang, Deng Cai, Victor Wai Kin Chan, and Jia Li. 2024. GLBench: A Comprehensive Benchmark for Graph with Large Language Models. arXiv preprint arXiv:2407.07457 (2024)."},{"key":"e_1_3_2_1_33_1","volume-title":"One For All: Towards Training One Graph Model For All Classification Tasks. In The Twelfth International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=4IT2pgc9v6","author":"Liu Hao","year":"2024","unstructured":"Hao Liu, Jiarui Feng, Lecheng Kong, Ningyue Liang, Dacheng Tao, Yixin Chen, and Muhan Zhang. 2024. One For All: Towards Training One Graph Model For All Classification Tasks. In The Twelfth International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=4IT2pgc9v6"},{"key":"e_1_3_2_1_34_1","volume-title":"Visual instruction tuning. Advances in neural information processing systems 36","author":"Liu Haotian","year":"2024","unstructured":"Haotian Liu, Chunyuan Li, Qingyang Wu, and Yong Jae Lee. 2024. Visual instruction tuning. Advances in neural information processing systems 36 (2024)."},{"key":"e_1_3_2_1_35_1","unstructured":"Jiawei Liu Cheng Yang Zhiyuan Lu Junze Chen Yibo Li Mengmei Zhang Ting Bai Yuan Fang Lichao Sun Philip S Yu et al. 2023. Towards graph foundation models: A survey and beyond. arXiv preprint arXiv:2310.11829 (2023)."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1145\/3560815"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3543507.3583386"},{"key":"e_1_3_2_1_38_1","volume-title":"Decoupled Weight Decay Regularization. In International Conference on Learning Representations.","author":"Loshchilov Ilya","unstructured":"Ilya Loshchilov and Frank Hutter. [n. d.]. Decoupled Weight Decay Regularization. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_39_1","volume-title":"Parameters Efficient Fine-Tuning for Long-Tailed Sequential Recommendation. In CAAI International Conference on Artificial Intelligence. Springer, 442--459","author":"Lv Zheqi","year":"2023","unstructured":"Zheqi Lv, Feng Wang, Shengyu Zhang, Wenqiao Zhang, Kun Kuang, and Fei Wu. 2023. Parameters Efficient Fine-Tuning for Long-Tailed Sequential Recommendation. In CAAI International Conference on Artificial Intelligence. Springer, 442--459."},{"key":"e_1_3_2_1_40_1","volume-title":"Optimize Incompatible Parameters through Compatibility-aware Knowledge Integration. arXiv preprint arXiv:2501.07596","author":"Lv Zheqi","year":"2025","unstructured":"Zheqi Lv, Keming Ye, ZishuWei, Qi Tian, Shengyu Zhang,Wenqiao Zhang,Wenjie Wang, Kun Kuang, Tat-Seng Chua, and Fei Wu. 2025. Optimize Incompatible Parameters through Compatibility-aware Knowledge Integration. arXiv preprint arXiv:2501.07596 (2025)."},{"key":"e_1_3_2_1_41_1","volume-title":"Collaboration of Large Language Models and Small Recommendation Models for Device-Cloud Recommendation. arXiv preprint arXiv:2501.05647","author":"Lv Zheqi","year":"2025","unstructured":"Zheqi Lv, Tianyu Zhan, Wenjie Wang, Xinyu Lin, Shengyu Zhang, Wenqiao Zhang, Jiwei Li, Kun Kuang, and Fei Wu. 2025. Collaboration of Large Language Models and Small Recommendation Models for Device-Cloud Recommendation. arXiv preprint arXiv:2501.05647 (2025)."},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/3543507.3583451"},{"key":"e_1_3_2_1_43_1","volume-title":"Wiki-cs: A wikipedia-based benchmark for graph neural networks. arXiv preprint arXiv:2007.02901","author":"Mernyei P\u00e9ter","year":"2020","unstructured":"P\u00e9ter Mernyei and Catalina Cangea. 2020. Wiki-cs: A wikipedia-based benchmark for graph neural networks. arXiv preprint arXiv:2007.02901 (2020)."},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1018"},{"key":"e_1_3_2_1_45_1","volume-title":"Revisiting catastrophic forgetting in class incremental learning. arXiv preprint arXiv:2107.12308","author":"Ni Zixuan","year":"2021","unstructured":"Zixuan Ni, Haizhou Shi, Siliang Tang, LonghuiWei, Qi Tian, and Yueting Zhuang. 2021. Revisiting catastrophic forgetting in class incremental learning. arXiv preprint arXiv:2107.12308 (2021)."},{"key":"e_1_3_2_1_46_1","volume-title":"Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748","author":"van den Oord Aaron","year":"2018","unstructured":"Aaron van den Oord, Yazhe Li, and Oriol Vinyals. 2018. Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748 (2018)."},{"key":"e_1_3_2_1_48_1","volume-title":"Towards Unified Multimodal Editing with Enhanced Knowledge Collaboration. arXiv preprint arXiv:2409.19872","author":"Pan Kaihang","year":"2024","unstructured":"Kaihang Pan, Zhaoyu Fan, Juncheng Li, Qifan Yu, Hao Fei, Siliang Tang, Richang Hong, Hanwang Zhang, and Qianru Sun. 2024. Towards Unified Multimodal Editing with Enhanced Knowledge Collaboration. arXiv preprint arXiv:2409.19872 (2024)."},{"key":"e_1_3_2_1_49_1","volume-title":"Self-supervised Meta-Prompt Learning with Meta-Gradient Regularization for Few-shot Generalization. arXiv preprint arXiv:2303.12314","author":"Pan Kaihang","year":"2023","unstructured":"Kaihang Pan, Juncheng Li, Hongye Song, Jun Lin, Xiaozhong Liu, and Siliang Tang. 2023. Self-supervised Meta-Prompt Learning with Meta-Gradient Regularization for Few-shot Generalization. arXiv preprint arXiv:2303.12314 (2023)."},{"key":"e_1_3_2_1_50_1","volume-title":"Auto-Encoding Morph-Tokens for Multimodal LLM. arXiv preprint arXiv:2405.01926","author":"Pan Kaihang","year":"2024","unstructured":"Kaihang Pan, Siliang Tang, Juncheng Li, Zhaoyu Fan, Wei Chow, Shuicheng Yan, Tat-Seng Chua, Yueting Zhuang, and Hanwang Zhang. 2024. Auto-Encoding Morph-Tokens for Multimodal LLM. arXiv preprint arXiv:2405.01926 (2024)."},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-70365-2_3"},{"key":"e_1_3_2_1_52_1","unstructured":"Boci Peng Yun Zhu Yongchao Liu Xiaohe Bo Haizhou Shi Chuntao Hong Yan Zhang and Siliang Tang. 2024. Graph Retrieval-Augmented Generation: A Survey. arXiv:2408.08921 [cs.AI] https:\/\/arxiv.org\/abs\/2408.08921"},{"key":"e_1_3_2_1_53_1","volume-title":"International conference on machine learning. PMLR, 8748--8763","author":"Radford Alec","year":"2021","unstructured":"Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. 2021. Learning transferable visual models from natural language supervision. In International conference on machine learning. PMLR, 8748--8763."},{"key":"e_1_3_2_1_54_1","volume-title":"Anh Tuan Luu, Guy Wolf, and Dominique Beaini.","author":"Ramp\u00e1\u0161ek Ladislav","year":"2022","unstructured":"Ladislav Ramp\u00e1\u0161ek, Mikhail Galkin, Vijay Prakash Dwivedi, Anh Tuan Luu, Guy Wolf, and Dominique Beaini. 2022. Recipe for a General, Powerful, Scalable Graph Transformer. Advances in Neural Information Processing Systems 35 (2022)."},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1410"},{"key":"e_1_3_2_1_56_1","volume-title":"Collective classification in network data. AI magazine","author":"Sen Prithviraj","year":"2008","unstructured":"Prithviraj Sen, Galileo Namata, Mustafa Bilgic, Lise Getoor, Brian Galligher, and Tina Eliassi-Rad. 2008. Collective classification in network data. AI magazine (2008)."},{"key":"e_1_3_2_1_57_1","volume-title":"Zheng Xu, John Dickerson, Christoph Studer, Larry S Davis, Gavin Taylor, and Tom Goldstein.","author":"Shafahi Ali","year":"2019","unstructured":"Ali Shafahi, Mahyar Najibi, Mohammad Amin Ghiasi, Zheng Xu, John Dickerson, Christoph Studer, Larry S Davis, Gavin Taylor, and Tom Goldstein. 2019. Adversarial training for free! Proc. of NeurIPS (2019)."},{"key":"e_1_3_2_1_58_1","volume-title":"Continual Learning of Large Language Models: A Comprehensive Survey. arXiv preprint arXiv:2404.16789","author":"Shi Haizhou","year":"2024","unstructured":"Haizhou Shi, Zihao Xu, Hengyi Wang, Weiyi Qin, Wenyuan Wang, Yibin Wang, ZifengWang, Sayna Ebrahimi, and HaoWang. 2024. Continual Learning of Large Language Models: A Comprehensive Survey. arXiv preprint arXiv:2404.16789 (2024)."},{"key":"e_1_3_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1145\/3534678.3539249"},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599256"},{"key":"e_1_3_2_1_61_1","volume-title":"Proc. of NeurIPS","author":"Suresh Susheel","year":"2021","unstructured":"Susheel Suresh, Pan Li, Cong Hao, and Jennifer Neville. 2021. Adversarial graph augmentation to improve graph contrastive learning. Proc. of NeurIPS (2021)."},{"key":"e_1_3_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1145\/3626772.3657775"},{"key":"e_1_3_2_1_63_1","volume-title":"ICLR 2021 Workshop on Geometrical and Topological Representation Learning.","author":"Thakoor Shantanu","year":"2021","unstructured":"Shantanu Thakoor, Corentin Tallec, Mohammad Gheshlaghi Azar, R\u00e9mi Munos, Petar Velickovic, and Michal Valko. 2021. Bootstrapped representation learning on graphs. In ICLR 2021 Workshop on Geometrical and Topological Representation Learning."},{"key":"e_1_3_2_1_64_1","unstructured":"Hugo Touvron Louis Martin Kevin Stone Peter Albert Amjad Almahairi Yasmine Babaei Nikolay Bashlykov Soumya Batra Prajjwal Bhargava Shruti Bhosale et al. 2023. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288 (2023)."},{"key":"e_1_3_2_1_65_1","unstructured":"Raja Vavekanand and Kira Sam. 2024. Llama 3.1: An In-Depth Analysis of the Next-Generation Large Language Model."},{"key":"e_1_3_2_1_66_1","volume-title":"Proc. of ICLR.","author":"Velickovic Petar","year":"2018","unstructured":"Petar Velickovic, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Li\u00f2, and Yoshua Bengio. 2018. Graph Attention Networks. In Proc. of ICLR."},{"key":"e_1_3_2_1_67_1","volume-title":"Proc. of ICLR.","author":"Velickovic Petar","unstructured":"Petar Velickovic, William Fedus, William L. Hamilton, Pietro Li\u00f2, Yoshua Bengio, and R. Devon Hjelm. 2019. Deep Graph Infomax. In Proc. of ICLR."},{"key":"e_1_3_2_1_68_1","volume-title":"Microsoft academic graph: When experts are not enough. Quantitative Science Studies","author":"Wang Kuansan","year":"2020","unstructured":"Kuansan Wang, Zhihong Shen, Chiyuan Huang, Chieh-Han Wu, Yuxiao Dong, and Anshul Kanakia. 2020. Microsoft academic graph: When experts are not enough. Quantitative Science Studies (2020)."},{"key":"e_1_3_2_1_69_1","volume-title":"Text embeddings by weakly-supervised contrastive pre-training. arXiv preprint arXiv:2212.03533","author":"Wang Liang","year":"2022","unstructured":"Liang Wang, Nan Yang, Xiaolong Huang, Binxing Jiao, Linjun Yang, Daxin Jiang, Rangan Majumder, and Furu Wei. 2022. Text embeddings by weakly-supervised contrastive pre-training. arXiv preprint arXiv:2212.03533 (2022)."},{"key":"e_1_3_2_1_70_1","volume-title":"Proc. of ICML.","author":"Phillip Isola TongzhouWang","year":"2020","unstructured":"TongzhouWang and Phillip Isola. 2020. Understanding contrastive representation learning through alignment and uniformity on the hypersphere. In Proc. of ICML."},{"key":"e_1_3_2_1_71_1","first-page":"5776","article-title":"Minilm: Deep self-attention distillation for task-agnostic compression of pre-trained transformers","volume":"33","author":"Wang Wenhui","year":"2020","unstructured":"Wenhui Wang, Furu Wei, Li Dong, Hangbo Bao, Nan Yang, and Ming Zhou. 2020. Minilm: Deep self-attention distillation for task-agnostic compression of pre-trained transformers. Advances in Neural Information Processing Systems 33 (2020), 5776--5788.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_72_1","unstructured":"Xiaotang Wang Yun Zhu Haizhou Shi Yongchao Liu and Chuntao Hong. 2024. Graph Triple Attention Network: A Decoupled Perspective. arXiv:2408.07654 [cs.LG] https:\/\/arxiv.org\/abs\/2408.07654"},{"key":"e_1_3_2_1_73_1","unstructured":"Xiaotang Wang Yun Zhu Haizhou Shi Yongchao Liu and Chuntao Hong. 2024. UniGAP: A Universal and Adaptive Graph Upsampling Approach to Mitigate Over-Smoothing in Node Classification Tasks. arXiv:2407.19420 [cs.LG] https:\/\/arxiv.org\/abs\/2407.19420"},{"key":"e_1_3_2_1_74_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.emnlp-main.823"},{"key":"e_1_3_2_1_75_1","doi-asserted-by":"publisher","DOI":"10.1145\/3539618.3591641"},{"key":"e_1_3_2_1_76_1","doi-asserted-by":"publisher","DOI":"10.1145\/3366423.3380219"},{"key":"e_1_3_2_1_77_1","volume-title":"Proc. of NeurIPS","author":"Wu Qitian","year":"2022","unstructured":"Qitian Wu, Wentao Zhao, Zenan Li, David P Wipf, and Junchi Yan. 2022. Nodeformer: A scalable graph structure learning transformer for node classification. Proc. of NeurIPS (2022)."},{"key":"e_1_3_2_1_78_1","volume-title":"Lpml: Llm-prompting markup language for mathematical reasoning. arXiv preprint arXiv:2309.13078","author":"Yamauchi Ryutaro","year":"2023","unstructured":"Ryutaro Yamauchi, Sho Sonoda, Akiyoshi Sannai, and Wataru Kumagai. 2023. Lpml: Llm-prompting markup language for mathematical reasoning. arXiv preprint arXiv:2309.13078 (2023)."},{"key":"e_1_3_2_1_79_1","volume-title":"Proc. of NeurIPS.","author":"Yan Hao","year":"2023","unstructured":"Hao Yan, Chaozhuo Li, Ruosong Long, Chao Yan, Jianan Zhao, Wenwen Zhuang, Jun Yin, Peiyan Zhang,Weihao Han, Hao Sun, et al. 2023. A Comprehensive Study on Text-attributed Graphs: Benchmarking and Rethinking. In Proc. of NeurIPS."},{"key":"e_1_3_2_1_80_1","unstructured":"An Yang Baosong Yang Binyuan Hui Bo Zheng Bowen Yu Chang Zhou Chengpeng Li Chengyuan Li Dayiheng Liu Fei Huang et al. 2024. Qwen2 technical report. arXiv preprint arXiv:2407.10671 (2024)."},{"key":"e_1_3_2_1_81_1","volume-title":"Proc. of NeurIPS","author":"Yang Junhan","year":"2021","unstructured":"Junhan Yang, Zheng Liu, Shitao Xiao, Chaozhuo Li, Defu Lian, Sanjay Agrawal, Amit Singh, Guangzhong Sun, and Xing Xie. 2021. GraphFormers: GNN-nested transformers for representation learning on textual graph. Proc. of NeurIPS (2021)."},{"key":"e_1_3_2_1_82_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2024.112155"},{"key":"e_1_3_2_1_83_1","volume-title":"Exploiting Large Language Models Capabilities for Question Answer-Driven Knowledge Graph Completion Across Static and Temporal Domains. arXiv preprint arXiv:2408.10819","author":"Yang Rui","year":"2024","unstructured":"Rui Yang, Jiahao Zhu, Jianping Man, Li Fang, and Yi Zhou. 2024. Exploiting Large Language Models Capabilities for Question Answer-Driven Knowledge Graph Completion Across Static and Temporal Domains. arXiv preprint arXiv:2408.10819 (2024)."},{"key":"e_1_3_2_1_84_1","volume-title":"Proc. of ICML.","author":"Yang Zhilin","year":"2016","unstructured":"Zhilin Yang, William Cohen, and Ruslan Salakhudinov. 2016. Revisiting semisupervised learning with graph embeddings. In Proc. of ICML."},{"key":"e_1_3_2_1_85_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2024.3369699"},{"key":"e_1_3_2_1_86_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.01585"},{"key":"e_1_3_2_1_87_1","volume-title":"Proc. of ICLR.","author":"Zhao Jianan","year":"2022","unstructured":"Jianan Zhao, Meng Qu, Chaozhuo Li, Hao Yan, Qian Liu, Rui Li, Xing Xie, and Jian Tang. 2022. Learning on Large-scale Text-attributed Graphs via Variational Inference. In Proc. of ICLR."},{"key":"e_1_3_2_1_88_1","unstructured":"Jianan Zhao Le Zhuo Yikang Shen Meng Qu Kai Liu Michael Bronstein Zhaocheng Zhu and Jian Tang. 2023. GraphText: Graph Reasoning in Text Space. arXiv:2310.01089 [cs.CL]"},{"key":"e_1_3_2_1_89_1","unstructured":"Wayne Xin Zhao Kun Zhou Junyi Li Tianyi Tang Xiaolei Wang Yupeng Hou Yingqian Min Beichen Zhang Junjie Zhang Zican Dong et al. 2023. A survey of large language models. arXiv preprint arXiv:2303.18223 (2023)."},{"key":"e_1_3_2_1_90_1","volume-title":"Proc. of ICLR.","author":"Zhao Xuyang","year":"2022","unstructured":"Xuyang Zhao, Tianqi Du, Yisen Wang, Jun Yao, and Weiran Huang. 2022. ArCL: Enhancing Contrastive Learning with Augmentation-Robust Representations. In Proc. of ICLR."},{"key":"e_1_3_2_1_91_1","volume-title":"Proc. of ICML.","author":"Zhou Xiao","year":"2022","unstructured":"Xiao Zhou, Yong Lin, Weizhong Zhang, and Tong Zhang. 2022. Sparse invariant risk minimization. In Proc. of ICML."},{"key":"e_1_3_2_1_92_1","volume-title":"SGL-PT: A Strong Graph Learner with Graph Prompt Tuning. arXiv preprint arXiv:2302.12449","author":"Zhu Yun","year":"2023","unstructured":"Yun Zhu, Jianhao Guo, and Siliang Tang. 2023. SGL-PT: A Strong Graph Learner with Graph Prompt Tuning. arXiv preprint arXiv:2302.12449 (2023)."},{"key":"e_1_3_2_1_93_1","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2022\/527"},{"key":"e_1_3_2_1_94_1","doi-asserted-by":"publisher","DOI":"10.1145\/3589334.3645322"},{"key":"e_1_3_2_1_95_1","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2024\/634"},{"key":"e_1_3_2_1_96_1","doi-asserted-by":"publisher","DOI":"10.1145\/3589334.3645439"},{"key":"e_1_3_2_1_97_1","volume-title":"Deep Graph Contrastive Representation Learning. In ICML Workshop on Graph Representation Learning and Beyond.","author":"Zhu Yanqiao","year":"2020","unstructured":"Yanqiao Zhu, Yichen Xu, Feng Yu, Qiang Liu, Shu Wu, and Liang Wang. 2020. Deep Graph Contrastive Representation Learning. In ICML Workshop on Graph Representation Learning and Beyond."}],"event":{"name":"WWW '25: The ACM Web Conference 2025","location":"Sydney NSW Australia","acronym":"WWW '25","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"]},"container-title":["Proceedings of the ACM on Web Conference 2025"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3696410.3714801","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3696410.3714801","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:18:42Z","timestamp":1750295922000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3696410.3714801"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4,22]]},"references-count":96,"alternative-id":["10.1145\/3696410.3714801","10.1145\/3696410"],"URL":"https:\/\/doi.org\/10.1145\/3696410.3714801","relation":{},"subject":[],"published":{"date-parts":[[2025,4,22]]},"assertion":[{"value":"2025-04-22","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}