{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,18]],"date-time":"2025-12-18T14:30:06Z","timestamp":1766068206007,"version":"3.48.0"},"publisher-location":"New York, NY, USA","reference-count":43,"publisher":"ACM","license":[{"start":{"date-parts":[[2025,11,10]],"date-time":"2025-11-10T00:00:00Z","timestamp":1762732800000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/100000001","name":"NSF (National Science Foundation)","doi-asserted-by":"publisher","award":["2007716, 2007976, 1942594, 1907805"],"award-info":[{"award-number":["2007716, 2007976, 1942594, 1907805"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Cisco Faculty Research Award"},{"DOI":"10.13039\/100000002","name":"NIH (National Institutes of Health)","doi-asserted-by":"publisher","award":["R01AG089806"],"award-info":[{"award-number":["R01AG089806"]}],"id":[{"id":"10.13039\/100000002","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,11,10]]},"DOI":"10.1145\/3746252.3761357","type":"proceedings-article","created":{"date-parts":[[2025,11,8]],"date-time":"2025-11-08T00:29:28Z","timestamp":1762561768000},"page":"4263-4272","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["TAGA: Text-Attributed Graph Self-Supervised Learning by Synergizing Graph and Text Mutual Transformations"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0008-9808-6020","authenticated-orcid":false,"given":"Zheng","family":"Zhang","sequence":"first","affiliation":[{"name":"Emory University, Atlanta, GA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3802-9039","authenticated-orcid":false,"given":"Yuntong","family":"Hu","sequence":"additional","affiliation":[{"name":"Emory University, Atlanta, GA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-7501-7581","authenticated-orcid":false,"given":"Bo","family":"Pan","sequence":"additional","affiliation":[{"name":"Emory University, Atlanta, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8044-6026","authenticated-orcid":false,"given":"Chen","family":"Ling","sequence":"additional","affiliation":[{"name":"Emory University, Atlanta, GA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2648-9989","authenticated-orcid":false,"given":"Liang","family":"Zhao","sequence":"additional","affiliation":[{"name":"Emory University, Atlanta, GA, USA"}]}],"member":"320","published-online":{"date-parts":[[2025,11,10]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Fastgcn: fast learning with graph convolutional networks via importance sampling. arXiv preprint arXiv:1801.10247","author":"Chen Jie","year":"2018","unstructured":"Jie Chen, Tengfei Ma, and Cao Xiao. 2018. Fastgcn: fast learning with graph convolutional networks via importance sampling. arXiv preprint arXiv:1801.10247 (2018)."},{"key":"e_1_3_2_1_2_1","volume-title":"Extending context window of large language models via positional interpolation. arXiv preprint arXiv:2306.15595","author":"Chen Shouyuan","year":"2023","unstructured":"Shouyuan Chen, Sherman Wong, Liangjian Chen, and Yuandong Tian. 2023b. Extending context window of large language models via positional interpolation. arXiv preprint arXiv:2306.15595 (2023)."},{"key":"e_1_3_2_1_3_1","volume-title":"Longlora: Efficient fine-tuning of long-context large language models. arXiv preprint arXiv:2309.12307","author":"Chen Yukang","year":"2023","unstructured":"Yukang Chen, Shengju Qian, Haotian Tang, Xin Lai, Zhijian Liu, Song Han, and Jiaya Jia. 2023a. Longlora: Efficient fine-tuning of long-context large language models. arXiv preprint arXiv:2309.12307 (2023)."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/3655103.3655110"},{"key":"e_1_3_2_1_5_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)."},{"key":"e_1_3_2_1_6_1","volume-title":"Longnet: Scaling transformers to 1,000,000,000 tokens. arXiv preprint arXiv:2307.02486","author":"Ding Jiayu","year":"2023","unstructured":"Jiayu Ding, Shuming Ma, Li Dong, Xingxing Zhang, Shaohan Huang, Wenhui Wang, Nanning Zheng, and Furu Wei. 2023a. Longnet: Scaling transformers to 1,000,000,000 tokens. arXiv preprint arXiv:2307.02486 (2023)."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i6.25898"},{"key":"e_1_3_2_1_8_1","volume-title":"The Twelfth International Conference on Learning Representations.","author":"Fatemi Bahare","year":"2023","unstructured":"Bahare Fatemi, Jonathan Halcrow, and Bryan Perozzi. 2023. Talk like a Graph: Encoding Graphs for Large Language Models. In The Twelfth International Conference on Learning Representations."},{"key":"e_1_3_2_1_9_1","volume-title":"Inductive representation learning on large graphs. Advances in neural information processing systems","author":"Hamilton Will","year":"2017","unstructured":"Will Hamilton, Zhitao Ying, and Jure Leskovec. 2017. Inductive representation learning on large graphs. Advances in neural information processing systems, Vol. 30 (2017)."},{"key":"e_1_3_2_1_10_1","volume-title":"Lm-infinite: Simple on-the-fly length generalization for large language models. arXiv preprint arXiv:2308.16137","author":"Han Chi","year":"2023","unstructured":"Chi Han, Qifan Wang, Wenhan Xiong, Yu Chen, Heng Ji, and Sinong Wang. 2023. Lm-infinite: Simple on-the-fly length generalization for large language models. arXiv preprint arXiv:2308.16137 (2023)."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3534678.3539321"},{"key":"e_1_3_2_1_12_1","volume-title":"Open Graph Benchmark: Datasets for Machine Learning on Graphs. arXiv preprint arXiv:2005.00687","author":"Hu Weihua","year":"2020","unstructured":"Weihua Hu, Matthias Fey, Marinka Zitnik, Yuxiao Dong, Hongyu Ren, Bowen Liu, Michele Catasta, and Jure Leskovec. 2020. Open Graph Benchmark: Datasets for Machine Learning on Graphs. arXiv preprint arXiv:2005.00687 (2020)."},{"key":"e_1_3_2_1_13_1","volume-title":"Beyond Text: A Deep Dive into Large Language Models' Ability on Understanding Graph Data. arXiv preprint arXiv:2310.04944","author":"Hu Yuntong","year":"2023","unstructured":"Yuntong Hu, Zheng Zhang, and Liang Zhao. 2023. Beyond Text: A Deep Dive into Large Language Models' Ability on Understanding Graph Data. arXiv preprint arXiv:2310.04944 (2023)."},{"key":"e_1_3_2_1_14_1","volume-title":"Can llms effectively leverage graph structural information: when and why. arXiv preprint arXiv:2309.16595","author":"Huang Jin","year":"2023","unstructured":"Jin Huang, Xingjian Zhang, Qiaozhu Mei, and Jiaqi Ma. 2023. Can llms effectively leverage graph structural information: when and why. arXiv preprint arXiv:2309.16595 (2023)."},{"key":"e_1_3_2_1_15_1","volume-title":"Longllmlingua: Accelerating and enhancing llms in long context scenarios via prompt compression. arXiv preprint arXiv:2310.06839","author":"Jiang Huiqiang","year":"2023","unstructured":"Huiqiang Jiang, Qianhui Wu, Xufang Luo, Dongsheng Li, Chin-Yew Lin, Yuqing Yang, and Lili Qiu. 2023. Longllmlingua: Accelerating and enhancing llms in long context scenarios via prompt compression. arXiv preprint arXiv:2310.06839 (2023)."},{"key":"e_1_3_2_1_16_1","volume-title":"Large language models on graphs: A comprehensive survey. arXiv preprint arXiv:2312.02783","author":"Jin Bowen","year":"2023","unstructured":"Bowen Jin, Gang Liu, Chi Han, Meng Jiang, Heng Ji, and Jiawei Han. 2023a. Large language models on graphs: A comprehensive survey. arXiv preprint arXiv:2312.02783 (2023)."},{"key":"e_1_3_2_1_17_1","volume-title":"Patton: Language model pretraining on text-rich networks. arXiv preprint arXiv:2305.12268","author":"Jin Bowen","year":"2023","unstructured":"Bowen Jin, Wentao Zhang, Yu Zhang, Yu Meng, Xinyang Zhang, Qi Zhu, and Jiawei Han. 2023c. Patton: Language model pretraining on text-rich networks. arXiv preprint arXiv:2305.12268 (2023)."},{"key":"e_1_3_2_1_18_1","volume-title":"Edgeformers: Graph-Empowered Transformers for Representation Learning on Textual-Edge Networks. In The Eleventh International Conference on Learning Representations,{ICLR}","author":"Jin Bowen","year":"2023","unstructured":"Bowen Jin, Yu Zhang, Yu Meng, and Jiawei Han. 2023b. Edgeformers: Graph-Empowered Transformers for Representation Learning on Textual-Edge Networks. In The Eleventh International Conference on Learning Representations,{ICLR} 2023. OpenReview. net."},{"key":"e_1_3_2_1_19_1","volume-title":"Variational graph auto-encoders. arXiv preprint arXiv:1611.07308","author":"Kipf Thomas N","year":"2016","unstructured":"Thomas N Kipf and Max Welling. 2016. Variational graph auto-encoders. arXiv preprint arXiv:1611.07308 (2016)."},{"volume-title":"Proceedings of the 5th International Conference on Learning Representations (Palais des Congr\u00e8s Neptune","author":"Thomas","key":"e_1_3_2_1_20_1","unstructured":"Thomas N. Kipf and Max Welling. 2017. Semi-Supervised Classification with Graph Convolutional Networks. In Proceedings of the 5th International Conference on Learning Representations (Palais des Congr\u00e8s Neptune, Toulon, France)."},{"key":"e_1_3_2_1_21_1","volume-title":"AnglE-optimized Text Embeddings. arXiv preprint arXiv:2309.12871","author":"Li Xianming","year":"2023","unstructured":"Xianming Li and Jing Li. 2023. AnglE-optimized Text Embeddings. arXiv preprint arXiv:2309.12871 (2023)."},{"key":"e_1_3_2_1_22_1","volume-title":"GRENADE: Graph-Centric Language Model for Self-Supervised Representation Learning on Text-Attributed Graphs. arXiv preprint arXiv:2310.15109","author":"Li Yichuan","year":"2023","unstructured":"Yichuan Li, Kaize Ding, and Kyumin Lee. 2023. GRENADE: Graph-Centric Language Model for Self-Supervised Representation Learning on Text-Attributed Graphs. arXiv preprint arXiv:2310.15109 (2023)."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1002\/asi.22883"},{"key":"e_1_3_2_1_24_1","unstructured":"OpenAI. 2023. Text-embedding-3-small Model. https:\/\/openai.com"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"crossref","unstructured":"Dmitry Paranyushkin. 2019. InfraNodus: Generating insight using text network analysis. In The world wide web conference. 3584-3589.","DOI":"10.1145\/3308558.3314123"},{"key":"e_1_3_2_1_26_1","volume-title":"Yarn: Efficient context window extension of large language models. arXiv preprint arXiv:2309.00071","author":"Peng Bowen","year":"2023","unstructured":"Bowen Peng, Jeffrey Quesnelle, Honglu Fan, and Enrico Shippole. 2023. Yarn: Efficient context window extension of large language models. arXiv preprint arXiv:2309.00071 (2023)."},{"key":"e_1_3_2_1_27_1","volume-title":"Pitfalls of graph neural network evaluation. arXiv preprint arXiv:1811.05868","author":"Shchur Oleksandr","year":"2018","unstructured":"Oleksandr Shchur, Maximilian Mumme, Aleksandar Bojchevski, and Stephan G\u00fcnnemann. 2018. Pitfalls of graph neural network evaluation. arXiv preprint arXiv:1811.05868 (2018)."},{"key":"e_1_3_2_1_28_1","volume-title":"Graphgpt: Graph instruction tuning for large language models. arXiv preprint arXiv:2310.13023","author":"Tang Jiabin","year":"2023","unstructured":"Jiabin Tang, Yuhao Yang, Wei Wei, Lei Shi, Lixin Su, Suqi Cheng, Dawei Yin, and Chao Huang. 2023. Graphgpt: Graph instruction tuning for large language models. arXiv preprint arXiv:2310.13023 (2023)."},{"key":"e_1_3_2_1_29_1","volume-title":"Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971","author":"Touvron Hugo","year":"2023","unstructured":"Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timoth\u00e9e Lacroix, Baptiste Rozi\u00e8re, Naman Goyal, Eric Hambro, Faisal Azhar, et al., 2023. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023)."},{"key":"e_1_3_2_1_30_1","volume-title":"Deep graph infomax. arXiv preprint arXiv:1809.10341","author":"Veli\u010dkovi\u0107 Petar","year":"2018","unstructured":"Petar Veli\u010dkovi\u0107, William Fedus, William L Hamilton, Pietro Li\u00f2, Yoshua Bengio, and R Devon Hjelm. 2018. Deep graph infomax. arXiv preprint arXiv:1809.10341 (2018)."},{"key":"e_1_3_2_1_31_1","volume-title":"Advances in Neural Information Processing Systems","volume":"36","author":"Wang Heng","year":"2024","unstructured":"Heng Wang, Shangbin Feng, Tianxing He, Zhaoxuan Tan, Xiaochuang Han, and Yulia Tsvetkov. 2024. Can language models solve graph problems in natural language? Advances in Neural Information Processing Systems, Vol. 36 (2024)."},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/3539618.3591641"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/3494523"},{"key":"e_1_3_2_1_34_1","volume-title":"International Conference on Learning Representations.","author":"Xu Keyulu","year":"2018","unstructured":"Keyulu Xu, Weihua Hu, Jure Leskovec, and Stefanie Jegelka. 2018. How Powerful are Graph Neural Networks?. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_35_1","first-page":"17238","article-title":"A comprehensive study on text-attributed graphs: Benchmarking and rethinking","volume":"36","author":"Yan Hao","year":"2023","unstructured":"Hao Yan, Chaozhuo Li, Ruosong Long, Chao Yan, Jianan Zhao, Wenwen Zhuang, Jun Yin, Peiyan Zhang, Weihao Han, Hao Sun, et al., 2023. A comprehensive study on text-attributed graphs: Benchmarking and rethinking. Advances in Neural Information Processing Systems, Vol. 36 (2023), 17238-17264.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_36_1","first-page":"28798","article-title":"Graphformers: Gnn-nested transformers for representation learning on textual graph","volume":"34","author":"Yang Junhan","year":"2021","unstructured":"Junhan Yang, Zheng Liu, Shitao Xiao, Chaozhuo Li, Defu Lian, Sanjay Agrawal, Amit Singh, Guangzhong Sun, and Xing Xie. 2021. Graphformers: Gnn-nested transformers for representation learning on textual graph. Advances in Neural Information Processing Systems, Vol. 34 (2021), 28798-28810.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_37_1","volume-title":"International conference on machine learning. PMLR, 40-48","author":"Yang Zhilin","year":"2016","unstructured":"Zhilin Yang, William Cohen, and Ruslan Salakhudinov. 2016. Revisiting semi-supervised learning with graph embeddings. In International conference on machine learning. PMLR, 40-48."},{"key":"e_1_3_2_1_38_1","volume-title":"Natural language is all a graph needs. arXiv preprint arXiv:2308.07134","author":"Ye Ruosong","year":"2023","unstructured":"Ruosong Ye, Caiqi Zhang, Runhui Wang, Shuyuan Xu, and Yongfeng Zhang. 2023. Natural language is all a graph needs. arXiv preprint arXiv:2308.07134 (2023)."},{"key":"e_1_3_2_1_39_1","volume-title":"Graph contrastive learning with augmentations. Advances in neural information processing systems","author":"You Yuning","year":"2020","unstructured":"Yuning You, Tianlong Chen, Yongduo Sui, Ting Chen, Zhangyang Wang, and Yang Shen. 2020. Graph contrastive learning with augmentations. Advances in neural information processing systems, Vol. 33 (2020), 5812-5823."},{"key":"e_1_3_2_1_40_1","volume-title":"Graphsaint: Graph sampling based inductive learning method. arXiv preprint arXiv:1907.04931","author":"Zeng Hanqing","year":"2019","unstructured":"Hanqing Zeng, Hongkuan Zhou, Ajitesh Srivastava, Rajgopal Kannan, and Viktor Prasanna. 2019. Graphsaint: Graph sampling based inductive learning method. arXiv preprint arXiv:1907.04931 (2019)."},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1145\/3589335.3641255"},{"key":"e_1_3_2_1_42_1","volume-title":"Learning on large-scale text-attributed graphs via variational inference. arXiv preprint arXiv:2210.14709","author":"Zhao Jianan","year":"2022","unstructured":"Jianan Zhao, Meng Qu, Chaozhuo Li, Hao Yan, Qian Liu, Rui Li, Xing Xie, and Jian Tang. 2022. Learning on large-scale text-attributed graphs via variational inference. arXiv preprint arXiv:2210.14709 (2022)."},{"key":"e_1_3_2_1_43_1","volume-title":"Deep graph contrastive representation learning. arXiv preprint arXiv:2006.04131","author":"Zhu Yanqiao","year":"2020","unstructured":"Yanqiao Zhu, Yichen Xu, Feng Yu, Qiang Liu, Shu Wu, and Liang Wang. 2020. Deep graph contrastive representation learning. arXiv preprint arXiv:2006.04131 (2020)."}],"event":{"name":"CIKM '25: The 34th ACM International Conference on Information and Knowledge Management","sponsor":["SIGIR ACM Special Interest Group on Information Retrieval","SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"],"location":"Seoul Republic of Korea","acronym":"CIKM '25"},"container-title":["Proceedings of the 34th ACM International Conference on Information and Knowledge Management"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/abs\/10.1145\/3746252.3761357","content-type":"text\/html","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3746252.3761357","content-type":"application\/pdf","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3746252.3761357","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,12]],"date-time":"2025-12-12T01:47:44Z","timestamp":1765504064000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3746252.3761357"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,10]]},"references-count":43,"alternative-id":["10.1145\/3746252.3761357","10.1145\/3746252"],"URL":"https:\/\/doi.org\/10.1145\/3746252.3761357","relation":{},"subject":[],"published":{"date-parts":[[2025,11,10]]},"assertion":[{"value":"2025-11-10","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}