{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,11]],"date-time":"2026-04-11T13:08:24Z","timestamp":1775912904707,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":46,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,8,24]],"date-time":"2024-08-24T00:00:00Z","timestamp":1724457600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"Guangzhou Municipal Nansha District Science and Technology Bureau","award":["2022ZD012"],"award-info":[{"award-number":["2022ZD012"]}]},{"DOI":"10.13039\/https:\/\/doi.org\/10.13039\/501100012166","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","award":["23IAA02114"],"award-info":[{"award-number":["23IAA02114"]}],"id":[{"id":"10.13039\/https:\/\/doi.org\/10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/https:\/\/doi.org\/10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U23B2030, U22B2057"],"award-info":[{"award-number":["U23B2030, U22B2057"]}],"id":[{"id":"10.13039\/https:\/\/doi.org\/10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,8,25]]},"DOI":"10.1145\/3637528.3671965","type":"proceedings-article","created":{"date-parts":[[2024,8,25]],"date-time":"2024-08-25T04:55:12Z","timestamp":1724561712000},"page":"307-318","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":13,"title":["Large Language Model-driven Meta-structure Discovery in Heterogeneous Information Network"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2605-749X","authenticated-orcid":false,"given":"Lin","family":"Chen","sequence":"first","affiliation":[{"name":"Hong Kong University of Science and Technology, Hong Kong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5720-4026","authenticated-orcid":false,"given":"Fengli","family":"Xu","sequence":"additional","affiliation":[{"name":"BNRist, Department of Electronic Engineering, Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4689-2289","authenticated-orcid":false,"given":"Nian","family":"Li","sequence":"additional","affiliation":[{"name":"Shenzhen International Graduate School, Tsinghua University, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9634-7962","authenticated-orcid":false,"given":"Zhenyu","family":"Han","sequence":"additional","affiliation":[{"name":"BNRist, Department of Electronic Engineering, Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3094-7735","authenticated-orcid":false,"given":"Meng","family":"Wang","sequence":"additional","affiliation":[{"name":"Hefei University of Technology, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5617-1659","authenticated-orcid":false,"given":"Yong","family":"Li","sequence":"additional","affiliation":[{"name":"BNRist, Department of Electronic Engineering, Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6026-1083","authenticated-orcid":false,"given":"Pan","family":"Hui","sequence":"additional","affiliation":[{"name":"Hong Kong University of Science and Technology (Guangzhou) &amp; Hong Kong University of Science and Technology, Guangzhou, China"}]}],"member":"320","published-online":{"date-parts":[[2024,8,24]]},"reference":[{"key":"e_1_3_2_2_1_1","unstructured":"Zhikai Chen Haitao Mao Hang Li Wei Jin Hongzhi Wen Xiaochi Wei Shuaiqiang Wang Dawei Yin Wenqi Fan Hui Liu et al. 2023. Exploring the potential of large language models (llms) in learning on graphs. arXiv preprint arXiv:2307.03393 (2023)."},{"key":"e_1_3_2_2_2_1","unstructured":"Jingtao Ding Chang Liu Yu Zheng Yunke Zhang Zihan Yu Ruikun Li Hongyi Chen Jinghua Piao Huandong Wang Jiazhen Liu et al. 2024. Artificial Intelligence for Complex Network: Potential Methodology and Application. arXiv preprint arXiv:2402.16887 (2024)."},{"key":"e_1_3_2_2_3_1","doi-asserted-by":"publisher","DOI":"10.1145\/3447548.3467447"},{"key":"e_1_3_2_2_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/3097983.3098036"},{"key":"e_1_3_2_2_5_1","volume-title":"Talk like a graph: Encoding graphs for large language models. arXiv preprint arXiv:2310.04560","author":"Fatemi Bahare","year":"2023","unstructured":"Bahare Fatemi, Jonathan Halcrow, and Bryan Perozzi. 2023. Talk like a graph: Encoding graphs for large language models. arXiv preprint arXiv:2310.04560 (2023)."},{"key":"e_1_3_2_2_6_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11023-020-09548-1"},{"key":"e_1_3_2_2_7_1","volume-title":"Proceedings of the 2017 ACM on Conference on Information and Knowledge Management. 1797--1806","author":"Lee Wang-Chien","year":"2017","unstructured":"Tao-yang Fu, Wang-Chien Lee, and Zhen Lei. 2017. Hin2vec: Explore meta-paths in heterogeneous information networks for representation learning. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management. 1797--1806."},{"key":"e_1_3_2_2_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3366423.3380297"},{"key":"e_1_3_2_2_9_1","volume-title":"GPT4Graph: Can Large Language Models Understand Graph Structured Data? An Empirical Evaluation and Benchmarking. arXiv preprint arXiv:2305.15066","author":"Guo Jiayan","year":"2023","unstructured":"Jiayan Guo, Lun Du, and Hengyu Liu. 2023. GPT4Graph: Can Large Language Models Understand Graph Structured Data? An Empirical Evaluation and Benchmarking. arXiv preprint arXiv:2305.15066 (2023)."},{"key":"e_1_3_2_2_10_1","volume-title":"Connecting large language models with evolutionary algorithms yields powerful prompt optimizers. arXiv preprint arXiv:2309.08532","author":"Guo Qingyan","year":"2023","unstructured":"Qingyan Guo, Rui Wang, Junliang Guo, Bei Li, Kaitao Song, Xu Tan, Guoqing Liu, Jiang Bian, and Yujiu Yang. 2023. Connecting large language models with evolutionary algorithms yields powerful prompt optimizers. arXiv preprint arXiv:2309.08532 (2023)."},{"key":"e_1_3_2_2_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3340531.3412015"},{"key":"e_1_3_2_2_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/3366423.3380027"},{"key":"e_1_3_2_2_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/2939672.2939815"},{"key":"e_1_3_2_2_14_1","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2017\/270"},{"key":"e_1_3_2_2_15_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i7.26026"},{"key":"e_1_3_2_2_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDE51399.2021.00084"},{"key":"e_1_3_2_2_17_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i5.16544"},{"key":"e_1_3_2_2_18_1","unstructured":"Jiawei Liu Cheng Yang Zhiyuan Lu Junze Chen Yibo Li Mengmei Zhang Ting Bai Yuan Fang Lichao Sun Philip S Yu et al. 2023. Towards graph foundation models: A survey and beyond. arXiv preprint arXiv:2310.11829 (2023)."},{"key":"e_1_3_2_2_19_1","unstructured":"Siyi Liu Chen Gao and Yong Li. 2024. Large Language Model Agent for Hyper-Parameter Optimization. arxiv: 2402.01881 [cs.LG]"},{"key":"e_1_3_2_2_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/2736277.2741123"},{"key":"e_1_3_2_2_21_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-981-97-2262-4_6"},{"key":"e_1_3_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/3511808.3557244"},{"key":"e_1_3_2_2_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/3490181"},{"key":"e_1_3_2_2_24_1","doi-asserted-by":"publisher","DOI":"10.1111\/j.1365-2702.2006.01573.x"},{"key":"e_1_3_2_2_25_1","volume-title":"Jordan S Ellenberg, Pengming Wang, Omar Fawzi, et al.","author":"Romera-Paredes Bernardino","year":"2023","unstructured":"Bernardino Romera-Paredes, Mohammadamin Barekatain, Alexander Novikov, Matej Balog, M Pawan Kumar, Emilien Dupont, Francisco JR Ruiz, Jordan S Ellenberg, Pengming Wang, Omar Fawzi, et al. 2023. Mathematical discoveries from program search with large language models. Nature (2023), 1--3."},{"key":"e_1_3_2_2_26_1","volume-title":"Wayne Xin Zhao, and S Yu Philip","author":"Shi Chuan","year":"2018","unstructured":"Chuan Shi, Binbin Hu, Wayne Xin Zhao, and S Yu Philip. 2018. Heterogeneous information network embedding for recommendation. IEEE transactions on knowledge and data engineering, Vol. 31, 2 (2018), 357--370."},{"key":"e_1_3_2_2_27_1","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2016.2598561"},{"key":"e_1_3_2_2_28_1","volume-title":"Thirty-seventh Conference on Neural Information Processing Systems. NeurIPS.","author":"Shinn Noah","year":"2023","unstructured":"Noah Shinn, Federico Cassano, Ashwin Gopinath, Karthik R Narasimhan, and Shunyu Yao. 2023. Reflexion: Language agents with verbal reinforcement learning. In Thirty-seventh Conference on Neural Information Processing Systems. NeurIPS."},{"key":"e_1_3_2_2_29_1","volume-title":"Evaluating the zero-shot robustness of instruction-tuned language models. arXiv preprint arXiv:2306.11270","author":"Sun Jiuding","year":"2023","unstructured":"Jiuding Sun, Chantal Shaib, and Byron C Wallace. 2023. Evaluating the zero-shot robustness of instruction-tuned language models. arXiv preprint arXiv:2306.11270 (2023)."},{"key":"e_1_3_2_2_30_1","volume-title":"Think-on-graph: Deep and responsible reasoning of large language model with knowledge graph. arXiv preprint arXiv:2307.07697","author":"Sun Jiashuo","year":"2023","unstructured":"Jiashuo Sun, Chengjin Xu, Lumingyuan Tang, Saizhuo Wang, Chen Lin, Yeyun Gong, Heung-Yeung Shum, and Jian Guo. 2023. Think-on-graph: Deep and responsible reasoning of large language model with knowledge graph. arXiv preprint arXiv:2307.07697 (2023)."},{"key":"e_1_3_2_2_31_1","doi-asserted-by":"publisher","DOI":"10.14778\/3402707.3402736"},{"key":"e_1_3_2_2_32_1","volume-title":"Graphgpt: Graph instruction tuning for large language models. arXiv preprint arXiv:2310.13023","author":"Tang Jiabin","year":"2023","unstructured":"Jiabin Tang, Yuhao Yang, Wei Wei, Lei Shi, Lixin Su, Suqi Cheng, Dawei Yin, and Chao Huang. 2023. Graphgpt: Graph instruction tuning for large language models. arXiv preprint arXiv:2310.13023 (2023)."},{"key":"e_1_3_2_2_33_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i04.6073"},{"key":"e_1_3_2_2_34_1","volume-title":"Can Language Models Solve Graph Problems in Natural Language? arXiv preprint arXiv:2305.10037","author":"Wang Heng","year":"2023","unstructured":"Heng Wang, Shangbin Feng, Tianxing He, Zhaoxuan Tan, Xiaochuang Han, and Yulia Tsvetkov. 2023. Can Language Models Solve Graph Problems in Natural Language? arXiv preprint arXiv:2305.10037 (2023)."},{"key":"e_1_3_2_2_35_1","volume-title":"ACM","author":"Wang Xiao","year":"2019","unstructured":"Xiao Wang, Houye Ji, Chuan Shi, Bai Wang, Yanfang Ye, Peng Cui, and Philip S Yu. 2019. Heterogeneous graph attention network. In The world wide web conference. ACM, 2022--2032."},{"key":"e_1_3_2_2_36_1","unstructured":"Jason Wei Yi Tay Rishi Bommasani Colin Raffel Barret Zoph Sebastian Borgeaud Dani Yogatama Maarten Bosma Denny Zhou Donald Metzler et al. 2022. Emergent abilities of large language models. arXiv preprint arXiv:2206.07682 (2022)."},{"key":"e_1_3_2_2_37_1","first-page":"24824","article-title":"Chain-of-thought prompting elicits reasoning in large language models","volume":"35","author":"Wei Jason","year":"2022","unstructured":"Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Fei Xia, Ed Chi, Quoc V Le, Denny Zhou, et al. 2022. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, Vol. 35 (2022), 24824--24837.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_2_38_1","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2020.2978386"},{"key":"e_1_3_2_2_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3357384.3357924"},{"key":"e_1_3_2_2_40_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-10928-8_3"},{"key":"e_1_3_2_2_41_1","volume-title":"Tree of thoughts: Deliberate problem solving with large language models. arXiv preprint arXiv:2305.10601","author":"Yao Shunyu","year":"2023","unstructured":"Shunyu Yao, Dian Yu, Jeffrey Zhao, Izhak Shafran, Thomas L Griffiths, Yuan Cao, and Karthik Narasimhan. 2023. Tree of thoughts: Deliberate problem solving with large language models. arXiv preprint arXiv:2305.10601 (2023)."},{"key":"e_1_3_2_2_42_1","volume-title":"Natural language is all a graph needs. arXiv preprint arXiv:2308.07134","author":"Ye Ruosong","year":"2023","unstructured":"Ruosong Ye, Caiqi Zhang, Runhui Wang, Shuyuan Xu, and Yongfeng Zhang. 2023. Natural language is all a graph needs. arXiv preprint arXiv:2308.07134 (2023)."},{"key":"e_1_3_2_2_43_1","volume-title":"Graph transformer networks. Advances in neural information processing systems","author":"Yun Seongjun","year":"2019","unstructured":"Seongjun Yun, Minbyul Jeong, Raehyun Kim, Jaewoo Kang, and Hyunwoo J Kim. 2019. Graph transformer networks. Advances in neural information processing systems, Vol. 32 (2019)."},{"key":"e_1_3_2_2_44_1","volume-title":"PPTC-R benchmark: Towards Evaluating the Robustness of Large Language Models for PowerPoint Task Completion. arXiv preprint arXiv:2403.03788","author":"Zhang Zekai","year":"2024","unstructured":"Zekai Zhang, Yiduo Guo, Yaobo Liang, Dongyan Zhao, and Nan Duan. 2024. PPTC-R benchmark: Towards Evaluating the Robustness of Large Language Models for PowerPoint Task Completion. arXiv preprint arXiv:2403.03788 (2024)."},{"key":"e_1_3_2_2_45_1","volume-title":"Graph Meets LLMs: Towards Large Graph Models. In NeurIPS 2023 Workshop: New Frontiers in Graph Learning. NeurIPS.","author":"Zhang Ziwei","year":"2023","unstructured":"Ziwei Zhang, Haoyang Li, Zeyang Zhang, Yijian Qin, Xin Wang, and Wenwu Zhu. 2023. Graph Meets LLMs: Towards Large Graph Models. In NeurIPS 2023 Workshop: New Frontiers in Graph Learning. NeurIPS."},{"key":"e_1_3_2_2_46_1","volume-title":"Graphtext: Graph reasoning in text space. arXiv preprint arXiv:2310.01089","author":"Zhao Jianan","year":"2023","unstructured":"Jianan Zhao, Le Zhuo, Yikang Shen, Meng Qu, Kai Liu, Michael Bronstein, Zhaocheng Zhu, and Jian Tang. 2023. Graphtext: Graph reasoning in text space. arXiv preprint arXiv:2310.01089 (2023)."}],"event":{"name":"KDD '24: The 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining","location":"Barcelona Spain","acronym":"KDD '24","sponsor":["SIGMOD ACM Special Interest Group on Management of Data","SIGKDD ACM Special Interest Group on Knowledge Discovery in Data"]},"container-title":["Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3637528.3671965","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3637528.3671965","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T00:06:05Z","timestamp":1750291565000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3637528.3671965"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,8,24]]},"references-count":46,"alternative-id":["10.1145\/3637528.3671965","10.1145\/3637528"],"URL":"https:\/\/doi.org\/10.1145\/3637528.3671965","relation":{},"subject":[],"published":{"date-parts":[[2024,8,24]]},"assertion":[{"value":"2024-08-24","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}