{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,10]],"date-time":"2026-04-10T10:00:23Z","timestamp":1775815223390,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":60,"publisher":"ACM","license":[{"start":{"date-parts":[[2025,7,20]],"date-time":"2025-07-20T00:00:00Z","timestamp":1752969600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"Ministry of Education, Singapore, under the Academic Research Fund Tier 2 (FY2025)","award":["MOE-T2EP20124-0009"],"award-info":[{"award-number":["MOE-T2EP20124-0009"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,7,20]]},"DOI":"10.1145\/3690624.3709277","type":"proceedings-article","created":{"date-parts":[[2025,4,4]],"date-time":"2025-04-04T18:42:22Z","timestamp":1743792142000},"page":"448-459","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":7,"title":["UniGraph: Learning a Unified Cross-Domain Foundation Model for Text-Attributed Graphs"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8918-6734","authenticated-orcid":false,"given":"Yufei","family":"He","sequence":"first","affiliation":[{"name":"National University of Singapore, Singapore, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8559-831X","authenticated-orcid":false,"given":"Yuan","family":"Sui","sequence":"additional","affiliation":[{"name":"National University of Singapore, Singapore, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8281-8070","authenticated-orcid":false,"given":"Xiaoxin","family":"He","sequence":"additional","affiliation":[{"name":"National University of Singapore, Singapore, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5645-1754","authenticated-orcid":false,"given":"Bryan","family":"Hooi","sequence":"additional","affiliation":[{"name":"National University of Singapore, Singapore, Singapore"}]}],"member":"320","published-online":{"date-parts":[[2025,7,20]]},"reference":[{"key":"e_1_3_2_2_1_1","doi-asserted-by":"publisher","DOI":"10.1145\/1052934.1052938"},{"key":"e_1_3_2_2_2_1","unstructured":"Rishi Bommasani Drew A Hudson Ehsan Adeli Russ Altman Simran Arora Sydney von Arx Michael S Bernstein Jeannette Bohg Antoine Bosselut Emma Brunskill et al. 2021. On the opportunities and risks of foundation models. arXiv preprint arXiv:2108.07258 (2021)."},{"key":"e_1_3_2_2_3_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00951"},{"key":"e_1_3_2_2_4_1","volume-title":"LLaGA: Large Language and Graph Assistant. arXiv preprint arXiv:2402.08170","author":"Chen Runjin","year":"2024","unstructured":"Runjin Chen, Tong Zhao, Ajay Jaiswal, Neil Shah, and Zhangyang Wang. 2024. LLaGA: Large Language and Graph Assistant. arXiv preprint arXiv:2402.08170 (2024)."},{"key":"e_1_3_2_2_5_1","volume-title":"International Journal of Computer Vision","author":"Chen Xiaokang","year":"2023","unstructured":"Xiaokang Chen, Mingyu Ding, Xiaodi Wang, Ying Xin, Shentong Mo, Yunhao Wang, Shumin Han, Ping Luo, Gang Zeng, and Jingdong Wang. 2023. Context autoencoder for self-supervised representation learning. International Journal of Computer Vision (2023), 1--16."},{"key":"e_1_3_2_2_6_1","volume-title":"Xing","author":"Chiang Wei-Lin","year":"2023","unstructured":"Wei-Lin Chiang, Zhuohan Li, Zi Lin, Ying Sheng, Zhanghao Wu, Hao Zhang, Lianmin Zheng, Siyuan Zhuang, Yonghao Zhuang, Joseph E. Gonzalez, Ion Stoica, and Eric P. Xing. 2023. Vicuna: An Open-Source Chatbot Impressing GPT-4 with 90%* ChatGPT Quality. https:\/\/lmsys.org\/blog\/2023-03--30-vicuna\/"},{"key":"e_1_3_2_2_7_1","first-page":"4908","article-title":"Graph transfer learning via adversarial domain adaptation with graph convolution","volume":"35","author":"Dai Quanyu","year":"2022","unstructured":"Quanyu Dai, Xiao-Ming Wu, Jiaren Xiao, Xiao Shen, and Dan Wang. 2022. Graph transfer learning via adversarial domain adaptation with graph convolution. IEEE Transactions on Knowledge and Data Engineering, Vol. 35, 5 (2022), 4908--4922.","journal-title":"IEEE Transactions on Knowledge and Data Engineering"},{"key":"e_1_3_2_2_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3340531.3411922"},{"key":"e_1_3_2_2_9_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-20056-4_15"},{"key":"e_1_3_2_2_10_1","volume-title":"Advances in Neural Information Processing Systems","volume":"36","author":"Fang Taoran","year":"2024","unstructured":"Taoran Fang, Yunchao Zhang, Yang Yang, Chunping Wang, and Lei Chen. 2024. Universal prompt tuning for graph neural networks. Advances in Neural Information Processing Systems, Vol. 36 (2024)."},{"key":"e_1_3_2_2_11_1","volume-title":"Towards Foundation Models for Knowledge Graph Reasoning. In The Twelfth International Conference on Learning Representations.","author":"Galkin Mikhail","year":"2023","unstructured":"Mikhail Galkin, Xinyu Yuan, Hesham Mostafa, Jian Tang, and Zhaocheng Zhu. 2023. Towards Foundation Models for Knowledge Graph Reasoning. In The Twelfth International Conference on Learning Representations."},{"key":"e_1_3_2_2_12_1","volume-title":"International Conference on Learning Representations","author":"Gasteiger Johannes","year":"2018","unstructured":"Johannes Gasteiger, Aleksandar Bojchevski, and Stephan G\u00fcnnemann. 2018. Predict then propagate: Graph neural networks meet personalized pagerank. International Conference on Learning Representations (2018)."},{"key":"e_1_3_2_2_13_1","volume-title":"Zhaohan Guo, Mohammad Gheshlaghi Azar, et al.","author":"Grill Jean-Bastien","year":"2020","unstructured":"Jean-Bastien Grill, Florian Strub, Florent Altch\u00e9, Corentin Tallec, Pierre Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Guo, Mohammad Gheshlaghi Azar, et al. 2020. Bootstrap your own latent-a new approach to self-supervised learning. Advances in neural information processing systems, Vol. 33 (2020), 21271--21284."},{"key":"e_1_3_2_2_14_1","volume-title":"GPT4Graph: Can Large Language Models Understand Graph Structured Data? An Empirical Evaluation and Benchmarking. arXiv preprint arXiv:2305.15066","author":"Guo Jiayan","year":"2023","unstructured":"Jiayan Guo, Lun Du, and Hengyu Liu. 2023. GPT4Graph: Can Large Language Models Understand Graph Structured Data? An Empirical Evaluation and Benchmarking. arXiv preprint arXiv:2305.15066 (2023)."},{"key":"e_1_3_2_2_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01553"},{"key":"e_1_3_2_2_16_1","volume-title":"Deberta: Decoding-enhanced bert with disentangled attention. arXiv preprint arXiv:2006.03654","author":"He Pengcheng","year":"2020","unstructured":"Pengcheng He, Xiaodong Liu, Jianfeng Gao, and Weizhu Chen. 2020. Deberta: Decoding-enhanced bert with disentangled attention. arXiv preprint arXiv:2006.03654 (2020)."},{"key":"e_1_3_2_2_17_1","volume-title":"Harnessing Explanations: LLM-to-LM Interpreter for Enhanced Text-Attributed Graph Representation Learning. International Conference on Learning Representations","author":"He Xiaoxin","year":"2024","unstructured":"Xiaoxin He, Xavier Bresson, Thomas Laurent, Adam Perold, Yann LeCun, and Bryan Hooi. 2024a. Harnessing Explanations: LLM-to-LM Interpreter for Enhanced Text-Attributed Graph Representation Learning. International Conference on Learning Representations (2024)."},{"key":"e_1_3_2_2_18_1","volume-title":"Generalizing Graph Transformers Across Diverse Graphs and Tasks via Pre-Training on Industrial-Scale Data. arXiv preprint arXiv:2407.03953","author":"He Yufei","year":"2024","unstructured":"Yufei He, Zhenyu Hou, Yukuo Cen, Feng He, Xu Cheng, and Bryan Hooi. 2024b. Generalizing Graph Transformers Across Diverse Graphs and Tasks via Pre-Training on Industrial-Scale Data. arXiv preprint arXiv:2407.03953 (2024)."},{"key":"e_1_3_2_2_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDMW58026.2022.00091"},{"key":"e_1_3_2_2_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3543507.3583379"},{"key":"e_1_3_2_2_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/3534678.3539321"},{"key":"e_1_3_2_2_22_1","volume-title":"LoRA: Low-Rank Adaptation of Large Language Models. In International Conference on Learning Representations.","author":"Hu Edward J","year":"2021","unstructured":"Edward J Hu, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, Weizhu Chen, et al. 2021b. LoRA: Low-Rank Adaptation of Large Language Models. In International Conference on Learning Representations."},{"key":"e_1_3_2_2_23_1","volume-title":"NeurIPS","volume":"34","author":"Hu Weihua","year":"2021","unstructured":"Weihua Hu, Matthias Fey, Hongyu Ren, Maho Nakata, Yuxiao Dong, and Jure Leskovec. 2021a. OGB-LSC: A Large-Scale Challenge for Machine Learning on Graphs. NeurIPS, Vol. 34 (2021)."},{"key":"e_1_3_2_2_24_1","volume-title":"Open graph benchmark: Datasets for machine learning on graphs. Advances in neural information processing systems","author":"Hu Weihua","year":"2020","unstructured":"Weihua Hu, Matthias Fey, Marinka Zitnik, Yuxiao Dong, Hongyu Ren, Bowen Liu, Michele Catasta, and Jure Leskovec. 2020. Open graph benchmark: Datasets for machine learning on graphs. Advances in neural information processing systems, Vol. 33 (2020), 22118--22133."},{"key":"e_1_3_2_2_25_1","volume-title":"PRODIGY: Enabling In-context Learning Over Graphs. NeurIPS","author":"Huang Qian","year":"2023","unstructured":"Qian Huang, Hongyu Ren, Peng Chen, Gregor Kr\u017emanc, Daniel Zeng, Percy Liang, and Jure Leskovec. 2023. PRODIGY: Enabling In-context Learning Over Graphs. NeurIPS (2023)."},{"key":"e_1_3_2_2_26_1","volume-title":"Proceedings of NAACL-HLT. 4171--4186","author":"Ming-Wei Chang Jacob Devlin","year":"2019","unstructured":"Jacob Devlin Ming-Wei Chang Kenton and Lee Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of NAACL-HLT. 4171--4186."},{"key":"e_1_3_2_2_27_1","volume-title":"Semi-Supervised Classification with Graph Convolutional Networks. In International Conference on Learning Representations.","author":"Kipf Thomas N","year":"2016","unstructured":"Thomas N Kipf and Max Welling. 2016a. Semi-Supervised Classification with Graph Convolutional Networks. In International Conference on Learning Representations."},{"key":"e_1_3_2_2_28_1","volume-title":"Variational graph auto-encoders. arXiv preprint arXiv:1611.07308","author":"Kipf Thomas N","year":"2016","unstructured":"Thomas N Kipf and Max Welling. 2016b. Variational graph auto-encoders. arXiv preprint arXiv:1611.07308 (2016)."},{"key":"e_1_3_2_2_29_1","volume-title":"Jeffrey Xu Yu, and Jia Li","author":"Li Yuhan","year":"2024","unstructured":"Yuhan Li, Peisong Wang, Zhixun Li, Jeffrey Xu Yu, and Jia Li. 2024. ZeroG: Investigating Cross-dataset Zero-shot Transferability in Graphs. arXiv preprint arXiv:2402.11235 (2024)."},{"key":"e_1_3_2_2_30_1","volume-title":"Evaluating large language models on graphs: Performance insights and comparative analysis. arXiv preprint arXiv:2308.11224","author":"Liu Chang","year":"2023","unstructured":"Chang Liu and Bo Wu. 2023. Evaluating large language models on graphs: Performance insights and comparative analysis. arXiv preprint arXiv:2308.11224 (2023)."},{"key":"e_1_3_2_2_31_1","volume-title":"One for All: Towards Training One Graph Model for All Classification Tasks. arXiv preprint arXiv:2310.00149","author":"Liu Hao","year":"2023","unstructured":"Hao Liu, Jiarui Feng, Lecheng Kong, Ningyue Liang, Dacheng Tao, Yixin Chen, and Muhan Zhang. 2023a. One for All: Towards Training One Graph Model for All Classification Tasks. arXiv preprint arXiv:2310.00149 (2023)."},{"key":"e_1_3_2_2_32_1","volume-title":"International Conference on Learning Representations.","author":"Liu Shengchao","year":"2021","unstructured":"Shengchao Liu, Hanchen Wang, Weiyang Liu, Joan Lasenby, Hongyu Guo, and Jian Tang. 2021. Pre-training Molecular Graph Representation with 3D Geometry. In International Conference on Learning Representations."},{"key":"e_1_3_2_2_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/3543507.3583386"},{"key":"e_1_3_2_2_34_1","doi-asserted-by":"publisher","DOI":"10.1145\/2835776.2835823"},{"key":"e_1_3_2_2_35_1","volume-title":"Wiki-cs: A wikipedia-based benchmark for graph neural networks. arXiv preprint arXiv:2007.02901","author":"Mernyei P\u00e9ter","year":"2020","unstructured":"P\u00e9ter Mernyei and C\u0103t\u0103lina Cangea. 2020. Wiki-cs: A wikipedia-based benchmark for graph neural networks. arXiv preprint arXiv:2007.02901 (2020)."},{"key":"e_1_3_2_2_36_1","volume-title":"Distributed representations of words and phrases and their compositionality. Advances in neural information processing systems","author":"Mikolov Tomas","year":"2013","unstructured":"Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. 2013. Distributed representations of words and phrases and their compositionality. Advances in neural information processing systems, Vol. 26 (2013)."},{"key":"e_1_3_2_2_37_1","volume-title":"Networks","author":"Newman Mark","unstructured":"Mark Newman. 2018. Networks. Oxford university press."},{"key":"e_1_3_2_2_38_1","doi-asserted-by":"publisher","DOI":"10.1145\/3394486.3403168"},{"key":"e_1_3_2_2_39_1","volume-title":"International conference on machine learning. PMLR, 8748--8763","author":"Radford Alec","year":"2021","unstructured":"Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. 2021. Learning transferable visual models from natural language supervision. In International conference on machine learning. PMLR, 8748--8763."},{"key":"e_1_3_2_2_40_1","unstructured":"Yuan Sui Yufei He Zifeng Ding and Bryan Hooi. 2024a. Can Knowledge Graphs Make Large Language Models More Trustworthy? An Empirical Study over Open-ended Question Answering. arxiv: 2410.08085 [cs.CL] https:\/\/arxiv.org\/abs\/2410.08085"},{"key":"e_1_3_2_2_41_1","volume-title":"FiDeLiS: Faithful Reasoning in Large Language Model for Knowledge Graph Question Answering. arXiv preprint arXiv:2405.13873","author":"Sui Yuan","year":"2024","unstructured":"Yuan Sui, Yufei He, Nian Liu, Xiaoxin He, Kun Wang, and Bryan Hooi. 2024b. FiDeLiS: Faithful Reasoning in Large Language Model for Knowledge Graph Question Answering. arXiv preprint arXiv:2405.13873 (2024)."},{"key":"e_1_3_2_2_42_1","unstructured":"Xiangguo Sun Hong Cheng Jia Li Bo Liu and Jihong Guan. 2023. All in One: Multi-Task Prompting for Graph Neural Networks. (2023)."},{"key":"e_1_3_2_2_43_1","volume-title":"Graphgpt: Graph instruction tuning for large language models. arXiv preprint arXiv:2310.13023","author":"Tang Jiabin","year":"2023","unstructured":"Jiabin Tang, Yuhao Yang, Wei Wei, Lei Shi, Lixin Su, Suqi Cheng, Dawei Yin, and Chao Huang. 2023. Graphgpt: Graph instruction tuning for large language models. arXiv preprint arXiv:2310.13023 (2023)."},{"key":"e_1_3_2_2_44_1","volume-title":"ICLR 2021 Workshop on Geometrical and Topological Representation Learning.","author":"Thakoor Shantanu","year":"2021","unstructured":"Shantanu Thakoor, Corentin Tallec, Mohammad Gheshlaghi Azar, R\u00e9mi Munos, Petar Veli\u010dkovi\u0107, and Michal Valko. 2021. Bootstrapped representation learning on graphs. In ICLR 2021 Workshop on Geometrical and Topological Representation Learning."},{"key":"e_1_3_2_2_45_1","volume-title":"Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971","author":"Touvron Hugo","year":"2023","unstructured":"Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timoth\u00e9e Lacroix, Baptiste Rozi\u00e8re, Naman Goyal, Eric Hambro, Faisal Azhar, et al. 2023a. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023)."},{"key":"e_1_3_2_2_46_1","unstructured":"Hugo Touvron Louis Martin Kevin Stone Peter Albert Amjad Almahairi Yasmine Babaei Nikolay Bashlykov Soumya Batra Prajjwal Bhargava Shruti Bhosale et al. 2023b. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288 (2023)."},{"key":"e_1_3_2_2_47_1","volume-title":"Graph Attention Networks. In International Conference on Learning Representations.","author":"Veli\u010dkovi\u0107 Petar","year":"2018","unstructured":"Petar Veli\u010dkovi\u0107, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Li\u00f2, and Yoshua Bengio. 2018a. Graph Attention Networks. In International Conference on Learning Representations."},{"key":"e_1_3_2_2_48_1","volume-title":"Deep Graph Infomax. In International Conference on Learning Representations.","author":"Veli\u010dkovi\u0107 Petar","year":"2018","unstructured":"Petar Veli\u010dkovi\u0107, William Fedus, William L Hamilton, Pietro Li\u00f2, Yoshua Bengio, and R Devon Hjelm. 2018b. Deep Graph Infomax. In International Conference on Learning Representations."},{"key":"e_1_3_2_2_49_1","volume-title":"Can Language Models Solve Graph Problems in Natural Language? arXiv preprint arXiv:2305.10037","author":"Wang Heng","year":"2023","unstructured":"Heng Wang, Shangbin Feng, Tianxing He, Zhaoxuan Tan, Xiaochuang Han, and Yulia Tsvetkov. 2023. Can Language Models Solve Graph Problems in Natural Language? arXiv preprint arXiv:2305.10037 (2023)."},{"key":"e_1_3_2_2_50_1","first-page":"38925","article-title":"Graph few-shot learning with task-specific structures","volume":"35","author":"Wang Song","year":"2022","unstructured":"Song Wang, Chen Chen, and Jundong Li. 2022a. Graph few-shot learning with task-specific structures. Advances in Neural Information Processing Systems, Vol. 35 (2022), 38925--38936.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_2_51_1","doi-asserted-by":"publisher","DOI":"10.1145\/3534678.3539265"},{"key":"e_1_3_2_2_52_1","doi-asserted-by":"publisher","DOI":"10.1145\/3539618.3591641"},{"key":"e_1_3_2_2_53_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00393"},{"key":"e_1_3_2_2_54_1","volume-title":"Large language models can learn temporal reasoning. arXiv preprint arXiv:2401.06853","author":"Xiong Siheng","year":"2024","unstructured":"Siheng Xiong, Ali Payani, Ramana Kompella, and Faramarz Fekri. 2024. Large language models can learn temporal reasoning. arXiv preprint arXiv:2401.06853 (2024)."},{"key":"e_1_3_2_2_55_1","volume-title":"Harnessing the power of large language models for natural language to first-order logic translation. arXiv preprint arXiv:2305.15541","author":"Yang Yuan","year":"2023","unstructured":"Yuan Yang, Siheng Xiong, Ali Payani, Ehsan Shareghi, and Faramarz Fekri. 2023. Harnessing the power of large language models for natural language to first-order logic translation. arXiv preprint arXiv:2305.15541 (2023)."},{"key":"e_1_3_2_2_56_1","volume-title":"Natural language is all a graph needs. arXiv preprint arXiv:2308.07134","author":"Ye Ruosong","year":"2023","unstructured":"Ruosong Ye, Caiqi Zhang, Runhui Wang, Shuyuan Xu, and Yongfeng Zhang. 2023. Natural language is all a graph needs. arXiv preprint arXiv:2308.07134 (2023)."},{"key":"e_1_3_2_2_57_1","volume-title":"SCR: Training Graph Neural Networks with Consistency Regularization. arXiv preprint arXiv:2112.04319","author":"Zhang Chenhui","year":"2021","unstructured":"Chenhui Zhang, Yufei He, Yukuo Cen, Zhenyu Hou, Wenzheng Feng, Yuxiao Dong, Xu Cheng, Hongyun Cai, Feng He, and Jie Tang. 2021b. SCR: Training Graph Neural Networks with Consistency Regularization. arXiv preprint arXiv:2112.04319 (2021)."},{"key":"e_1_3_2_2_58_1","volume-title":"Improving the training of graph neural networks with consistency regularization. arXiv preprint arXiv:2112.04319","author":"Zhang Chenhui","year":"2021","unstructured":"Chenhui Zhang, Yufei He, Yukuo Cen, Zhenyu Hou, and Jie Tang. 2021a. Improving the training of graph neural networks with consistency regularization. arXiv preprint arXiv:2112.04319 (2021)."},{"key":"e_1_3_2_2_59_1","doi-asserted-by":"publisher","DOI":"10.1145\/3637528.3671952"},{"key":"e_1_3_2_2_60_1","volume-title":"Learning on large-scale text-attributed graphs via variational inference. arXiv preprint arXiv:2210.14709","author":"Zhao Jianan","year":"2022","unstructured":"Jianan Zhao, Meng Qu, Chaozhuo Li, Hao Yan, Qian Liu, Rui Li, Xing Xie, and Jian Tang. 2022. Learning on large-scale text-attributed graphs via variational inference. arXiv preprint arXiv:2210.14709 (2022)."}],"event":{"name":"KDD '25: The 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining","location":"Toronto ON Canada","acronym":"KDD '25","sponsor":["SIGMOD ACM Special Interest Group on Management of Data","SIGKDD ACM Special Interest Group on Knowledge Discovery in Data"]},"container-title":["Proceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining V.1"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3690624.3709277","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3690624.3709277","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,16]],"date-time":"2025-08-16T15:37:11Z","timestamp":1755358631000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3690624.3709277"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,20]]},"references-count":60,"alternative-id":["10.1145\/3690624.3709277","10.1145\/3690624"],"URL":"https:\/\/doi.org\/10.1145\/3690624.3709277","relation":{},"subject":[],"published":{"date-parts":[[2025,7,20]]},"assertion":[{"value":"2025-07-20","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}