{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,10]],"date-time":"2026-03-10T03:50:51Z","timestamp":1773114651218,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":71,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,4,30]],"date-time":"2023-04-30T00:00:00Z","timestamp":1682812800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"NSFC","award":["91846204"],"award-info":[{"award-number":["91846204"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,4,30]]},"DOI":"10.1145\/3543507.3583301","type":"proceedings-article","created":{"date-parts":[[2023,4,26]],"date-time":"2023-04-26T23:30:25Z","timestamp":1682551825000},"page":"2581-2590","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":19,"title":["Structure Pretraining and Prompt Tuning for Knowledge Graph Transfer"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8429-9326","authenticated-orcid":false,"given":"Wen","family":"Zhang","sequence":"first","affiliation":[{"name":"Zhejiang University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9809-9269","authenticated-orcid":false,"given":"Yushan","family":"Zhu","sequence":"additional","affiliation":[{"name":"Zhejiang University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6080-3559","authenticated-orcid":false,"given":"Mingyang","family":"Chen","sequence":"additional","affiliation":[{"name":"Zhejiang University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2461-2613","authenticated-orcid":false,"given":"Yuxia","family":"Geng","sequence":"additional","affiliation":[{"name":"Zhejiang University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7110-9298","authenticated-orcid":false,"given":"Yufeng","family":"Huang","sequence":"additional","affiliation":[{"name":"Zhejiang University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0887-4445","authenticated-orcid":false,"given":"Yajing","family":"Xu","sequence":"additional","affiliation":[{"name":"Zhejiang University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3163-2399","authenticated-orcid":false,"given":"Wenting","family":"Song","sequence":"additional","affiliation":[{"name":"Huawei Technologies Co., Ltd, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5496-7442","authenticated-orcid":false,"given":"Huajun","family":"Chen","sequence":"additional","affiliation":[{"name":"Zhejiang University, China"}]}],"member":"320","published-online":{"date-parts":[[2023,4,30]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"ViViT: A Video Vision Transformer","author":"Arnab Anurag","unstructured":"Anurag Arnab, Mostafa Dehghani, Georg Heigold, Chen Sun, Mario Lucic, and Cordelia Schmid. 2021. ViViT: A Video Vision Transformer. In ICCV. IEEE, 6816\u20136826."},{"key":"e_1_3_2_1_2_1","unstructured":"Hangbo Bao Li Dong Songhao Piao and Furu Wei. 2022. BEiT: BERT Pre-Training of Image Transformers. In ICLR. OpenReview.net."},{"key":"e_1_3_2_1_3_1","unstructured":"Antoine Bordes Nicolas Usunier Alberto Garc\u00eda-Dur\u00e1n Jason Weston and Oksana Yakhnenko. 2013. Translating Embeddings for Modeling Multi-relational Data. In NIPS. 2787\u20132795."},{"key":"e_1_3_2_1_4_1","volume-title":"Toward an Architecture for Never-Ending Language Learning","author":"Carlson Andrew","unstructured":"Andrew Carlson, Justin Betteridge, Bryan Kisiel, Burr Settles, Estevam R.\u00a0Hruschka Jr., and Tom\u00a0M. Mitchell. 2010. Toward an Architecture for Never-Ending Language Learning. In AAAI. AAAI Press."},{"key":"e_1_3_2_1_5_1","volume-title":"EMNLP (1)","author":"Chen Sanxing","unstructured":"Sanxing Chen, Xiaodong Liu, Jianfeng Gao, Jian Jiao, Ruofei Zhang, and Yangfeng Ji. 2021. HittER: Hierarchical Transformers for Knowledge Graph Embeddings. In EMNLP (1). Association for Computational Linguistics, 10395\u201310407."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"crossref","unstructured":"Xiang Chen Ningyu Zhang Lei Li Shumin Deng Chuanqi Tan Changliang Xu Fei Huang Luo Si and Huajun Chen. 2022. Hybrid Transformer with Multi-level Fusion for Multimodal Knowledge Graph Completion. In SIGIR. ACM 904\u2013915.","DOI":"10.1145\/3477495.3531992"},{"key":"e_1_3_2_1_7_1","volume-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL-HLT (1)","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL-HLT (1). Association for Computational Linguistics, 4171\u20134186."},{"key":"e_1_3_2_1_8_1","unstructured":"Alexey Dosovitskiy Lucas Beyer Alexander Kolesnikov Dirk Weissenborn Xiaohua Zhai Thomas Unterthiner Mostafa Dehghani Matthias Minderer Georg Heigold Sylvain Gelly Jakob Uszkoreit and Neil Houlsby. 2021. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In ICLR. OpenReview.net."},{"key":"e_1_3_2_1_9_1","volume-title":"EMNLP (1)","author":"Feng Yanlin","unstructured":"Yanlin Feng, Xinyue Chen, Bill\u00a0Yuchen Lin, Peifeng Wang, Jun Yan, and Xiang Ren. 2020. Scalable Multi-Hop Relational Reasoning for Knowledge-Aware Question Answering. In EMNLP (1). Association for Computational Linguistics, 1295\u20131309."},{"key":"e_1_3_2_1_10_1","unstructured":"Andrea Frome Gregory\u00a0S. Corrado Jonathon Shlens Samy Bengio Jeffrey Dean Marc\u2019Aurelio Ranzato and Tom\u00e1s Mikolov. 2013. DeViSE: A Deep Visual-Semantic Embedding Model. In NIPS. 2121\u20132129."},{"key":"e_1_3_2_1_11_1","unstructured":"Dehong Gao Linbo Jin Ben Chen Minghui Qiu Peng Li Yi Wei Yi Hu and Hao Wang. 2020. FashionBERT: Text and Image Matching with Adaptive Loss for Cross-modal Retrieval. In SIGIR. ACM 2251\u20132260."},{"key":"e_1_3_2_1_12_1","volume-title":"ACL\/IJCNLP (1)","author":"Gao Tianyu","unstructured":"Tianyu Gao, Adam Fisch, and Danqi Chen. 2021. Making Pre-trained Language Models Better Few-shot Learners. In ACL\/IJCNLP (1). Association for Computational Linguistics, 3816\u20133830."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.44"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"crossref","unstructured":"Yuxia Geng Jiaoyan Chen Zhuo Chen Jeff\u00a0Z. Pan Zhiquan Ye Zonggang Yuan Yantao Jia and Huajun Chen. 2021. OntoZSL: Ontology-enhanced Zero-shot Learning. In WWW. ACM \/ IW3C2 3325\u20133336.","DOI":"10.1145\/3442381.3450042"},{"key":"e_1_3_2_1_15_1","volume-title":"K-ZSL: resources for knowledge-driven zero-shot learning. arXiv preprint arXiv:2106.15047","author":"Geng Yuxia","year":"2021","unstructured":"Yuxia Geng, Jiaoyan Chen, Zhuo Chen, Jeff\u00a0Z Pan, Zonggang Yuan, and Huajun Chen. 2021. K-ZSL: resources for knowledge-driven zero-shot learning. arXiv preprint arXiv:2106.15047 (2021)."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"crossref","unstructured":"Yuxia Geng Jiaoyan Chen Wen Zhang Yajing Xu Zhuo Chen Jeff\u00a0Z. Pan Yufeng Huang Feiyu Xiong and Huajun Chen. 2022. Disentangled Ontology Embedding for Zero-shot Learning. In KDD. ACM 443\u2013453.","DOI":"10.1145\/3534678.3539453"},{"key":"e_1_3_2_1_17_1","volume-title":"Deep Residual Learning for Image Recognition","author":"He Kaiming","unstructured":"Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep Residual Learning for Image Recognition. In CVPR. IEEE Computer Society, 770\u2013778."},{"key":"e_1_3_2_1_18_1","volume-title":"TaPas: Weakly Supervised Table Parsing via Pre-training","author":"Herzig Jonathan","unstructured":"Jonathan Herzig, Pawel\u00a0Krzysztof Nowak, Thomas M\u00fcller, Francesco Piccinno, and Julian\u00a0Martin Eisenschlos. 2020. TaPas: Weakly Supervised Table Parsing via Pre-training. In ACL. Association for Computational Linguistics, 4320\u20134333."},{"key":"e_1_3_2_1_19_1","volume-title":"Transformer-based Entity Typing in Knowledge Graphs","author":"Hu Zhiwei","unstructured":"Zhiwei Hu, V\u00edctor Guti\u00e9rrez-Basulto, Zhiliang Xiang, Ru Li, and Jeff\u00a0Z. Pan. 2022. Transformer-based Entity Typing in Knowledge Graphs. In EMNLP. Association for Computational Linguistics, 5988\u20136001."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"crossref","unstructured":"Xiao Huang Jingyuan Zhang Dingcheng Li and Ping Li. 2019. Knowledge Graph Embedding Based Question Answering. In WSDM. ACM 105\u2013113.","DOI":"10.1145\/3289600.3290956"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"crossref","unstructured":"Bosung Kim Taesuk Hong Youngjoong Ko and Jungyun Seo. 2020. Multi-Task Learning for Knowledge Graph Completion with Pre-trained Language Models. In COLING. International Committee on Computational Linguistics 1737\u20131743.","DOI":"10.18653\/v1\/2020.coling-main.153"},{"key":"e_1_3_2_1_22_1","volume-title":"Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980","author":"Kingma P","year":"2014","unstructured":"Diederik\u00a0P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)."},{"key":"e_1_3_2_1_23_1","volume-title":"NAACL-HLT (1)","author":"Koncel-Kedziorski Rik","unstructured":"Rik Koncel-Kedziorski, Dhanush Bekal, Yi Luan, Mirella Lapata, and Hannaneh Hajishirzi. 2019. Text Generation from Knowledge Graphs with Graph Transformers. In NAACL-HLT (1). Association for Computational Linguistics, 2284\u20132293."},{"key":"e_1_3_2_1_24_1","volume-title":"EMNLP (1)","author":"Lester Brian","unstructured":"Brian Lester, Rami Al-Rfou, and Noah Constant. 2021. The Power of Scale for Parameter-Efficient Prompt Tuning. In EMNLP (1). Association for Computational Linguistics, 3045\u20133059."},{"key":"e_1_3_2_1_25_1","volume-title":"BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension","author":"Lewis Mike","year":"2020","unstructured":"Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Veselin Stoyanov, and Luke Zettlemoyer. 2020. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension. In ACL. Association for Computational Linguistics, 7871\u20137880."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1145\/3534678.3539426"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2021.3055147"},{"key":"e_1_3_2_1_28_1","volume-title":"EMNLP\/IJCNLP (1)","author":"Lin Bill\u00a0Yuchen","unstructured":"Bill\u00a0Yuchen Lin, Xinyue Chen, Jamin Chen, and Xiang Ren. 2019. KagNet: Knowledge-Aware Graph Networks for Commonsense Reasoning. In EMNLP\/IJCNLP (1). Association for Computational Linguistics, 2829\u20132839."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"crossref","unstructured":"Qika Lin Jun Liu Fangzhi Xu Yudai Pan Yifan Zhu Lingling Zhang and Tianzhe Zhao. 2022. Incorporating Context Graph with Logical Reasoning for Inductive Relation Prediction. In SIGIR. ACM 893\u2013903.","DOI":"10.1145\/3477495.3531996"},{"key":"e_1_3_2_1_30_1","volume-title":"Learning Entity and Relation Embeddings for Knowledge Graph Completion","author":"Lin Yankai","unstructured":"Yankai Lin, Zhiyuan Liu, Maosong Sun, Yang Liu, and Xuan Zhu. 2015. Learning Entity and Relation Embeddings for Knowledge Graph Completion. In AAAI. AAAI Press, 2181\u20132187."},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"crossref","unstructured":"Lihui Liu Boxin Du Jiejun Xu Yinglong Xia and Hanghang Tong. 2022. Joint Knowledge Graph Completion and Question Answering. In KDD. ACM 1098\u20131108.","DOI":"10.1145\/3534678.3539289"},{"key":"e_1_3_2_1_32_1","volume-title":"prompt, and predict: A systematic survey of prompting methods in natural language processing. arXiv preprint arXiv:2107.13586","author":"Liu Pengfei","year":"2021","unstructured":"Pengfei Liu, Weizhe Yuan, Jinlan Fu, Zhengbao Jiang, Hiroaki Hayashi, and Graham Neubig. 2021. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. arXiv preprint arXiv:2107.13586 (2021)."},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"crossref","unstructured":"Xiao Liu Shiyu Zhao Kai Su Yukuo Cen Jiezhong Qiu Mengdi Zhang Wei Wu Yuxiao Dong and Jie Tang. 2022. Mask and Reason: Pre-Training Knowledge Graph Transformers for Complex Logical Queries. In KDD. ACM 1120\u20131130.","DOI":"10.1145\/3534678.3539472"},{"key":"e_1_3_2_1_34_1","volume-title":"RoBERTa: A Robustly Optimized BERT Pretraining Approach. CoRR abs\/1907.11692","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. RoBERTa: A Robustly Optimized BERT Pretraining Approach. CoRR abs\/1907.11692 (2019)."},{"key":"e_1_3_2_1_35_1","unstructured":"Jiasen Lu Dhruv Batra Devi Parikh and Stefan Lee. 2019. ViLBERT: Pretraining Task-Agnostic Visiolinguistic Representations for Vision-and-Language Tasks. In NeurIPS. 13\u201323."},{"key":"e_1_3_2_1_36_1","volume-title":"Communicative Message Passing for Inductive Relation Reasoning","author":"Mai Sijie","unstructured":"Sijie Mai, Shuangjia Zheng, Yuedong Yang, and Haifeng Hu. 2021. Communicative Message Passing for Inductive Relation Reasoning. In AAAI. AAAI Press, 4294\u20134302."},{"key":"e_1_3_2_1_37_1","unstructured":"Alec Radford Karthik Narasimhan Tim Salimans and Ilya Sutskever. 2018. Improving language understanding by generative pre-training. (2018)."},{"key":"e_1_3_2_1_38_1","article-title":"Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer","volume":"21","author":"Raffel Colin","year":"2020","unstructured":"Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter\u00a0J. Liu. 2020. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. J. Mach. Learn. Res. 21 (2020), 140:1\u2013140:67.","journal-title":"J. Mach. Learn. Res."},{"key":"e_1_3_2_1_39_1","volume-title":"YAGO: A Multilingual Knowledge Base from Wikipedia, Wordnet, and Geonames. In ISWC (2)(Lecture Notes in Computer Science, Vol.\u00a09982). 177\u2013185.","author":"Rebele Thomas","year":"2016","unstructured":"Thomas Rebele, Fabian\u00a0M. Suchanek, Johannes Hoffart, Joanna Biega, Erdal Kuzey, and Gerhard Weikum. 2016. YAGO: A Multilingual Knowledge Base from Wikipedia, Wordnet, and Geonames. In ISWC (2)(Lecture Notes in Computer Science, Vol.\u00a09982). 177\u2013185."},{"key":"e_1_3_2_1_40_1","volume-title":"ACL (1)","author":"Saxena Apoorv","unstructured":"Apoorv Saxena, Adrian Kochsiek, and Rainer Gemulla. 2022. Sequence-to-Sequence Knowledge Graph Completion and Question Answering. In ACL (1). Association for Computational Linguistics, 2814\u20132828."},{"key":"e_1_3_2_1_41_1","volume-title":"Improving Multi-hop Question Answering over Knowledge Graphs using Knowledge Base Embeddings","author":"Saxena Apoorv","unstructured":"Apoorv Saxena, Aditay Tripathi, and Partha\u00a0P. Talukdar. 2020. Improving Multi-hop Question Answering over Knowledge Graphs using Knowledge Base Embeddings. In ACL. Association for Computational Linguistics, 4498\u20134507."},{"key":"e_1_3_2_1_42_1","volume-title":"Ivan Titov, and Max Welling.","author":"Schlichtkrull Michael\u00a0Sejr","year":"2018","unstructured":"Michael\u00a0Sejr Schlichtkrull, Thomas\u00a0N. Kipf, Peter Bloem, Rianne van\u00a0den Berg, Ivan Titov, and Max Welling. 2018. Modeling Relational Data with Graph Convolutional Networks. In ESWC(Lecture Notes in Computer Science, Vol.\u00a010843). Springer, 593\u2013607."},{"key":"e_1_3_2_1_43_1","volume-title":"An Open Multilingual Graph of General Knowledge","author":"Speer Robyn","unstructured":"Robyn Speer, Joshua Chin, and Catherine Havasi. 2017. ConceptNet 5.5: An Open Multilingual Graph of General Knowledge. In AAAI. AAAI Press, 4444\u20134451."},{"key":"e_1_3_2_1_44_1","unstructured":"Weijie Su Xizhou Zhu Yue Cao Bin Li Lewei Lu Furu Wei and Jifeng Dai. 2020. VL-BERT: Pre-training of Generic Visual-Linguistic Representations. In ICLR. OpenReview.net."},{"key":"e_1_3_2_1_45_1","volume-title":"Graph Hawkes Transformer for Extrapolated Reasoning on Temporal Knowledge Graphs","author":"Sun Haohai","unstructured":"Haohai Sun, Shangyi Geng, Jialun Zhong, Han Hu, and Kun He. 2022. Graph Hawkes Transformer for Extrapolated Reasoning on Temporal Knowledge Graphs. In EMNLP. Association for Computational Linguistics, 7481\u20137493."},{"key":"e_1_3_2_1_46_1","unstructured":"Zhiqing Sun Zhi-Hong Deng Jian-Yun Nie and Jian Tang. 2019. RotatE: Knowledge Graph Embedding by Relational Rotation in Complex Space. In ICLR (Poster). OpenReview.net."},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"crossref","unstructured":"Zequn Sun Wei Hu Qingheng Zhang and Yuzhong Qu. 2018. Bootstrapping Entity Alignment with Knowledge Graph Embedding. In IJCAI. ijcai.org 4396\u20134402.","DOI":"10.24963\/ijcai.2018\/611"},{"key":"e_1_3_2_1_48_1","volume-title":"NAACL-HLT (1)","author":"Talmor Alon","unstructured":"Alon Talmor, Jonathan Herzig, Nicholas Lourie, and Jonathan Berant. 2019. CommonsenseQA: A Question Answering Challenge Targeting Commonsense Knowledge. In NAACL-HLT (1). Association for Computational Linguistics, 4149\u20134158."},{"key":"e_1_3_2_1_49_1","unstructured":"Komal\u00a0K. Teru Etienne\u00a0G. Denis and William\u00a0L. Hamilton. 2020. Inductive Relation Prediction by Subgraph Reasoning. In ICML(Proceedings of Machine Learning Research Vol.\u00a0119). PMLR 9448\u20139457."},{"key":"e_1_3_2_1_50_1","volume-title":"Complex Embeddings for Simple Link Prediction. In ICML(JMLR Workshop and Conference Proceedings, Vol.\u00a048)","author":"Trouillon Th\u00e9o","year":"2016","unstructured":"Th\u00e9o Trouillon, Johannes Welbl, Sebastian Riedel, \u00c9ric Gaussier, and Guillaume Bouchard. 2016. Complex Embeddings for Simple Link Prediction. In ICML(JMLR Workshop and Conference Proceedings, Vol.\u00a048). JMLR.org, 2071\u20132080."},{"key":"e_1_3_2_1_51_1","unstructured":"Shikhar Vashishth Soumya Sanyal Vikram Nitin and Partha\u00a0P. Talukdar. 2020. Composition-based Multi-Relational Graph Convolutional Networks. In ICLR. OpenReview.net."},{"key":"e_1_3_2_1_52_1","unstructured":"Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan\u00a0N. Gomez Lukasz Kaiser and Illia Polosukhin. 2017. Attention is All you Need. In NIPS. 5998\u20136008."},{"key":"e_1_3_2_1_53_1","volume-title":"Graph attention networks. ICLR","author":"Veli\u010dkovi\u0107 Petar","year":"2018","unstructured":"Petar Veli\u010dkovi\u0107, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Lio, and Yoshua Bengio. 2018. Graph attention networks. ICLR (2018)."},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1145\/2629489"},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"crossref","unstructured":"Bo Wang Tao Shen Guodong Long Tianyi Zhou Ying Wang and Yi Chang. 2021. Structure-Augmented Text Representation Learning for Efficient Knowledge Graph Completion. In WWW. ACM \/ IW3C2 1737\u20131748.","DOI":"10.1145\/3442381.3450043"},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"crossref","unstructured":"Hongwei Wang Fuzheng Zhang Jialin Wang Miao Zhao Wenjie Li Xing Xie and Minyi Guo. 2018. RippleNet: Propagating User Preferences on the Knowledge Graph for Recommender Systems. In CIKM. ACM 417\u2013426.","DOI":"10.1145\/3269206.3271739"},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00360"},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1145\/3292500.3330989"},{"key":"e_1_3_2_1_59_1","volume-title":"Improving Natural Language Inference Using External Knowledge in the Science Questions Domain","author":"Wang Xiaoyan","unstructured":"Xiaoyan Wang, Pavan Kapanipathi, Ryan Musa, Mo Yu, Kartik Talamadupula, Ibrahim Abdelaziz, Maria Chang, Achille Fokoue, Bassem Makni, Nicholas Mattei, and Michael Witbrock. 2019. Improving Natural Language Inference Using External Knowledge in the Science Questions Domain. In AAAI. AAAI Press, 7208\u20137215."},{"key":"e_1_3_2_1_60_1","volume-title":"Zero-Shot Recognition via Semantic Embeddings and Knowledge Graphs","author":"Wang Xiaolong","unstructured":"Xiaolong Wang, Yufei Ye, and Abhinav Gupta. 2018. Zero-Shot Recognition via Semantic Embeddings and Knowledge Graphs. In CVPR. Computer Vision Foundation \/ IEEE Computer Society, 6857\u20136866."},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1145\/3447548.3467434"},{"key":"e_1_3_2_1_62_1","doi-asserted-by":"crossref","unstructured":"Zihan Wang Zhaochun Ren Chunyu He Peng Zhang and Yue Hu. 2019. Robust Embedding with Multi-Level Structures for Link Prediction. In IJCAI. ijcai.org 5240\u20135246.","DOI":"10.24963\/ijcai.2019\/728"},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2018.2857768"},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"crossref","unstructured":"Xiaohan Xu Peng Zhang Yongquan He Chengpeng Chao and Chaoyang Yan. 2022. Subgraph Neighboring Relations Infomax for Inductive Link Prediction on Knowledge Graphs. In IJCAI. ijcai.org 2341\u20132347.","DOI":"10.24963\/ijcai.2022\/325"},{"key":"e_1_3_2_1_65_1","volume-title":"Neural-Symbolic Entangled Framework for Complex Query Answering. CoRR abs\/2209.08779","author":"Xu Zezhong","year":"2022","unstructured":"Zezhong Xu, Wen Zhang, Peng Ye, Hui Chen, and Huajun Chen. 2022. Neural-Symbolic Entangled Framework for Complex Query Answering. CoRR abs\/2209.08779 (2022)."},{"key":"e_1_3_2_1_66_1","volume-title":"KG-BERT: BERT for Knowledge Graph Completion. CoRR abs\/1909.03193","author":"Yao Liang","year":"2019","unstructured":"Liang Yao, Chengsheng Mao, and Yuan Luo. 2019. KG-BERT: BERT for Knowledge Graph Completion. CoRR abs\/1909.03193 (2019)."},{"key":"e_1_3_2_1_67_1","volume-title":"QA-GNN: Reasoning with Language Models and Knowledge Graphs for Question Answering","author":"Yasunaga Michihiro","unstructured":"Michihiro Yasunaga, Hongyu Ren, Antoine Bosselut, Percy Liang, and Jure Leskovec. 2021. QA-GNN: Reasoning with Language Models and Knowledge Graphs for Question Answering. In NAACL-HLT. Association for Computational Linguistics, 535\u2013546."},{"key":"e_1_3_2_1_68_1","unstructured":"Chengxuan Ying Tianle Cai Shengjie Luo Shuxin Zheng Guolin Ke Di He Yanming Shen and Tie-Yan Liu. 2021. Do Transformers Really Perform Badly for Graph Representation?. In NeurIPS. 28877\u201328888."},{"key":"e_1_3_2_1_69_1","doi-asserted-by":"crossref","unstructured":"Wen Zhang Bibek Paudel Liang Wang Jiaoyan Chen Hai Zhu Wei Zhang Abraham Bernstein and Huajun Chen. 2019. Iteratively Learning Embeddings and Rules for Knowledge Graph Reasoning. In WWW. ACM 2366\u20132377.","DOI":"10.1145\/3308558.3313612"},{"key":"e_1_3_2_1_70_1","volume-title":"Billion-scale Pre-trained E-commerce Product Knowledge Graph Model","author":"Zhang Wen","unstructured":"Wen Zhang, Chi\u00a0Man Wong, Ganqiang Ye, Bo Wen, Wei Zhang, and Huajun Chen. 2021. Billion-scale Pre-trained E-commerce Product Knowledge Graph Model. In ICDE. IEEE, 2476\u20132487."},{"key":"e_1_3_2_1_71_1","volume-title":"Learning Hierarchy-Aware Knowledge Graph Embeddings for Link Prediction","author":"Zhang Zhanqiu","unstructured":"Zhanqiu Zhang, Jianyu Cai, Yongdong Zhang, and Jie Wang. 2020. Learning Hierarchy-Aware Knowledge Graph Embeddings for Link Prediction. In AAAI. AAAI Press, 3065\u20133072."}],"event":{"name":"WWW '23: The ACM Web Conference 2023","location":"Austin TX USA","acronym":"WWW '23","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"]},"container-title":["Proceedings of the ACM Web Conference 2023"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3543507.3583301","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3543507.3583301","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:37:22Z","timestamp":1750178242000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3543507.3583301"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,4,30]]},"references-count":71,"alternative-id":["10.1145\/3543507.3583301","10.1145\/3543507"],"URL":"https:\/\/doi.org\/10.1145\/3543507.3583301","relation":{},"subject":[],"published":{"date-parts":[[2023,4,30]]},"assertion":[{"value":"2023-04-30","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}