{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,19]],"date-time":"2026-02-19T17:02:25Z","timestamp":1771520545287,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":48,"publisher":"ACM","license":[{"start":{"date-parts":[[2025,4,22]],"date-time":"2025-04-22T00:00:00Z","timestamp":1745280000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"NSFC","award":["62322606"],"award-info":[{"award-number":["62322606"]}]},{"name":"Zhejiang NSF","award":["LR22F020005"],"award-info":[{"award-number":["LR22F020005"]}]},{"name":"CCF-Zhipu Large Model Fund"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,4,22]]},"DOI":"10.1145\/3696410.3714792","type":"proceedings-article","created":{"date-parts":[[2025,4,22]],"date-time":"2025-04-22T23:08:29Z","timestamp":1745363309000},"page":"2158-2171","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Enhancing Cross-domain Link Prediction via Evolution Process Modeling"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-4668-4570","authenticated-orcid":false,"given":"Xuanwen","family":"Huang","sequence":"first","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-6075-6749","authenticated-orcid":false,"given":"Wei","family":"Chow","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-6756-6049","authenticated-orcid":false,"given":"Yize","family":"Zhu","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-1805-5346","authenticated-orcid":false,"given":"Yang","family":"Wang","sequence":"additional","affiliation":[{"name":"Finvolution Group, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1376-5101","authenticated-orcid":false,"given":"Ziwei","family":"Chai","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1854-8667","authenticated-orcid":false,"given":"Chunping","family":"Wang","sequence":"additional","affiliation":[{"name":"Finvolution Group, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4912-3293","authenticated-orcid":false,"given":"Lei","family":"Chen","sequence":"additional","affiliation":[{"name":"Finvolution Group, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5058-4417","authenticated-orcid":false,"given":"Yang","family":"Yang","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]}],"member":"320","published-online":{"date-parts":[[2025,4,22]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-04180-8_25"},{"key":"e_1_3_2_1_2_1","unstructured":"Rishi Bommasani Drew A Hudson Ehsan Adeli Russ Altman Simran Arora Sydney von Arx Michael S Bernstein Jeannette Bohg Antoine Bosselut Emma Brunskill et al. 2021. On the opportunities and risks of foundation models. In arXiv."},{"key":"e_1_3_2_1_3_1","volume-title":"NeurIPS'20","volume":"33","author":"Brown Tom","year":"2020","unstructured":"Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. 2020. Language models are few-shot learners. In NeurIPS'20, Vol. 33. 1877--1901."},{"key":"e_1_3_2_1_4_1","volume-title":"ICML'10","author":"Cao Bin","year":"2010","unstructured":"Bin Cao, Nathan N Liu, and Qiang Yang. 2010. Transfer learning for collective link prediction in multiple heterogenous domains. In ICML'10. 159--166."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/1242572.1242650"},{"key":"e_1_3_2_1_6_1","unstructured":"Wei Chow Juncheng Li Qifan Yu Kaihang Pan Hao Fei Zhiqi Ge Shuai Yang Siliang Tang Hanwang Zhang and Qianru Sun. 2024. Unified Generative and Discriminative Training for Multi-modal Large Language Models. In arXiv."},{"key":"e_1_3_2_1_7_1","unstructured":"Weilin Cong Si Zhang Jian Kang Baichuan Yuan Hao Wu Xin Zhou Hanghang Tong and Mehrdad Mahdavi. 2023. Do We Really Need Complicated Model Architectures For Temporal Networks?. In arXiv."},{"key":"e_1_3_2_1_8_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. In arXiv.","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. In arXiv."},{"key":"e_1_3_2_1_9_1","volume-title":"KAA: Kolmogorov-Arnold Attention for Enhancing Attentive Graph Neural Networks. In arXiv.","author":"Fang Taoran","year":"2025","unstructured":"Taoran Fang, Tianhong Gao, Chunping Wang, Yihao Shang, Wei Chow, Lei Chen, and Yang Yang. 2025. KAA: Kolmogorov-Arnold Attention for Enhancing Attentive Graph Neural Networks. In arXiv."},{"key":"e_1_3_2_1_10_1","unstructured":"Taoran Fang Wei Zhou Yifei Sun Kaiqiao Han Lvbin Ma and Yang Yang. 2024. Exploring Correlations of Self-supervised Tasks for Graphs. In arXiv."},{"key":"e_1_3_2_1_11_1","unstructured":"Ying Guo and Xi Chen. 2013. Cross-domain scientific collaborations prediction using citation. In ASONAM. 765--770."},{"key":"e_1_3_2_1_12_1","volume-title":"NeurIPS'17","volume":"30","author":"Hamilton Will","year":"2017","unstructured":"Will Hamilton, Zhitao Ying, and Jure Leskovec. 2017. Inductive representation learning on large graphs. In NeurIPS'17, Vol. 30."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/2567948.2576940"},{"key":"e_1_3_2_1_14_1","unstructured":"Shenyang Huang Farimah Poursafaei Jacob Danovitch Matthias Fey Weihua Hu Emanuele Rossi Jure Leskovec Michael Bronstein Guillaume Rabusseau and Reihaneh Rabbany. 2023. Temporal graph benchmark for machine learning on temporal graphs. In arXiv."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.5555\/3455716.3455786"},{"key":"e_1_3_2_1_16_1","unstructured":"Shah Noor Khan Eelco Herder and Diba Kaya. 2024. Experiences of Non-Mainstream and Minority Users with Music Recommendation Systems. In Mensch und Computer 2024-Workshopband. 10--18420."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"crossref","unstructured":"Alexander Kirillov Eric Mintun Nikhila Ravi Hanzi Mao Chloe Rolland Laura Gustafson Tete Xiao Spencer Whitehead Alexander C Berg Wan-Yen Lo et al. 2023. Segment anything. In arXiv.","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1126\/science.1116869"},{"key":"e_1_3_2_1_19_1","volume-title":"Kuldeep Singh, and Bhaskar Biswas.","author":"Kumar Ajay","year":"2020","unstructured":"Ajay Kumar, Shashank Sheshar Singh, Kuldeep Singh, and Bhaskar Biswas. 2020. Link prediction techniques, applications, and performance: A survey. In Physica A: Statistical Mechanics and its Applications, Vol. 553."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3292500.3330895"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1038\/nature03204"},{"key":"e_1_3_2_1_22_1","unstructured":"Hao Liu Jiarui Feng Lecheng Kong Ningyue Liang Dacheng Tao Yixin Chen and Muhan Zhang. 2023a. One for all: Towards training one graph model for all classification tasks. In arXiv."},{"key":"e_1_3_2_1_23_1","unstructured":"Jiawei Liu Cheng Yang Zhiyuan Lu Junze Chen Yibo Li Mengmei Zhang Ting Bai Yuan Fang Lichao Sun Philip S Yu et al. 2023b. Towards graph foundation models: A survey and beyond. In arXiv."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/3560815"},{"key":"e_1_3_2_1_25_1","unstructured":"Erfan Loghmani and MohammadAmin Fazli. 2023. Effect of Choosing Loss Function when Using T-batching for Representation Learning on Dynamic Networks. In arXiv."},{"key":"e_1_3_2_1_26_1","volume-title":"Physical review E","author":"Newman Mark EJ","unstructured":"Mark EJ Newman. 2001. Clustering and preferential attachment in growing networks. In Physical review E, Vol. 64."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.5555\/1543767.1543769"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3018661.3018731"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i04.5984"},{"key":"e_1_3_2_1_30_1","first-page":"2001","article-title":"Linguistic inquiry and word count: LIWC 2001","volume":"71","author":"Pennebaker James W","year":"2001","unstructured":"James W Pennebaker, Martha E Francis, and Roger J Booth. 2001. Linguistic inquiry and word count: LIWC 2001. In Mahway, Vol. 71. 2001.","journal-title":"Mahway"},{"key":"e_1_3_2_1_31_1","volume-title":"Proceedings of Machine Learning and Systems","volume":"5","author":"Pope Reiner","year":"2023","unstructured":"Reiner Pope, Sholto Douglas, Aakanksha Chowdhery, Jacob Devlin, James Bradbury, Jonathan Heek, Kefan Xiao, Shivani Agrawal, and Jeff Dean. 2023. Efficiently scaling transformer inference. In Proceedings of Machine Learning and Systems, Vol. 5."},{"key":"e_1_3_2_1_32_1","volume-title":"NeurIPS'22","volume":"35","author":"Poursafaei Farimah","year":"2022","unstructured":"Farimah Poursafaei, Shenyang Huang, Kellin Pelrine, and Reihaneh Rabbany. 2022. Towards better evaluation for dynamic link prediction. In NeurIPS'22, Vol. 35. 32928--32941."},{"key":"e_1_3_2_1_33_1","volume-title":"ICML'21","author":"Radford Alec","year":"2021","unstructured":"Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. 2021. Learning transferable visual models from natural language supervision. In ICML'21. PMLR, 8748--8763."},{"key":"e_1_3_2_1_34_1","unstructured":"Alec Radford Jeffrey Wu Rewon Child David Luan Dario Amodei Ilya Sutskever et al. 2019. Language models are unsupervised multitask learners. In OpenAI blog Vol. 1. 9."},{"key":"e_1_3_2_1_35_1","unstructured":"Emanuele Rossi Ben Chamberlain Fabrizio Frasca Davide Eynard Federico Monti and Michael Bronstein. 2020. Temporal graph networks for deep learning on dynamic graphs. In arXiv."},{"key":"e_1_3_2_1_36_1","volume-title":"Annual review of sociology","author":"Snijders Tom AB","unstructured":"Tom AB Snijders. 2011. Statistical models for social networks. In Annual review of sociology, Vol. 37. 131--153."},{"key":"e_1_3_2_1_37_1","unstructured":"Hugo Touvron Louis Martin Kevin Stone Peter Albert Amjad Almahairi Yasmine Babaei Nikolay Bashlykov Soumya Batra Prajjwal Bhargava Shruti Bhosale et al. 2023. Llama 2: Open foundation and fine-tuned chat models. In arXiv."},{"key":"e_1_3_2_1_38_1","volume-title":"ICLR'19","author":"Trivedi Rakshit","year":"2019","unstructured":"Rakshit Trivedi, Mehrdad Farajtabar, Prasenjeet Biswal, and Hongyuan Zha. 2019. Dyrep: Learning representations over dynamic graphs. In ICLR'19."},{"key":"e_1_3_2_1_39_1","unstructured":"Petar Velickovic Guillem Cucurull Arantxa Casanova Adriana Romero Pietro Lio and Yoshua Bengio. 2017. Graph attention networks. In arXiv."},{"key":"e_1_3_2_1_40_1","volume-title":"Tcl: Transformer-based dynamic graph modelling via contrastive learning. In arXiv.","author":"Wang Lu","year":"2021","unstructured":"Lu Wang, Xiaofu Chang, Shuang Li, Yunfei Chu, Hui Li, Wei Zhang, Xiaofeng He, Le Song, Jingren Zhou, and Hongxia Yang. 2021a. Tcl: Transformer-based dynamic graph modelling via contrastive learning. In arXiv."},{"key":"e_1_3_2_1_41_1","unstructured":"Yanbang Wang Yen-Yu Chang Yunyu Liu Jure Leskovec and Pan Li. 2021b. Inductive representation learning in temporal networks via causal anonymous walks. In arXiv."},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1109\/TMC.2014.2322373"},{"key":"e_1_3_2_1_43_1","volume-title":"Denny Zhou, et al.","author":"Wei Jason","year":"2022","unstructured":"Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Fei Xia, Ed Chi, Quoc V Le, Denny Zhou, et al. 2022. Chain-of-thought prompting elicits reasoning in large language models. In NeurIPS'22, Vol. 35. 24824--24837."},{"key":"e_1_3_2_1_44_1","unstructured":"Da Xu Chuanwei Ruan Evren Korpeoglu Sushant Kumar and Kannan Achan. 2020. Inductive representation learning on temporal graphs. In arXiv."},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1145\/3460231.3478845"},{"key":"e_1_3_2_1_46_1","unstructured":"Le Yu Leilei Sun Bowen Du and Weifeng Lv. 2023. Towards Better Dynamic Graph Learning: New Architecture and Unified Library. In arXiv."},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.14569\/IJACSA.2016.070273"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11257"}],"event":{"name":"WWW '25: The ACM Web Conference 2025","location":"Sydney NSW Australia","acronym":"WWW '25","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"]},"container-title":["Proceedings of the ACM on Web Conference 2025"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3696410.3714792","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3696410.3714792","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:18:41Z","timestamp":1750295921000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3696410.3714792"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4,22]]},"references-count":48,"alternative-id":["10.1145\/3696410.3714792","10.1145\/3696410"],"URL":"https:\/\/doi.org\/10.1145\/3696410.3714792","relation":{},"subject":[],"published":{"date-parts":[[2025,4,22]]},"assertion":[{"value":"2025-04-22","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}