{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,27]],"date-time":"2026-04-27T13:38:53Z","timestamp":1777297133848,"version":"3.51.4"},"publisher-location":"New York, NY, USA","reference-count":56,"publisher":"ACM","funder":[{"name":"National Science and Technology Major Project","award":["2023ZD0121103"],"award-info":[{"award-number":["2023ZD0121103"]}]},{"name":"National Natural Science Foundation of China","award":["62402159"],"award-info":[{"award-number":["62402159"]}]},{"name":"National Natural Science Foundation of China","award":["U23B2031"],"award-info":[{"award-number":["U23B2031"]}]},{"name":"National Natural Science Foundation of China","award":["72188101"],"award-info":[{"award-number":["72188101"]}]},{"name":"Fundamental Research Funds for the Central Universities","award":["PA2025IISL0105"],"award-info":[{"award-number":["PA2025IISL0105"]}]},{"name":"Key Laboratory of Knowledge Engineering with Big Data &#x28;the Ministry of Education of China&#x29;","award":["BigKEOpen2025-02"],"award-info":[{"award-number":["BigKEOpen2025-02"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2026,4,13]]},"DOI":"10.1145\/3774904.3792365","type":"proceedings-article","created":{"date-parts":[[2026,4,9]],"date-time":"2026-04-09T21:54:34Z","timestamp":1775771674000},"page":"6342-6353","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["WeaveRec: An LLM-Based Cross-Domain Sequential Recommendation Framework with Model Merging"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0524-6806","authenticated-orcid":false,"given":"Min","family":"Hou","sequence":"first","affiliation":[{"name":"The Key Laboratory of Knowledge Engineering with Big Data, Hefei University of Technology, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-9038-9132","authenticated-orcid":false,"given":"Xin","family":"Liu","sequence":"additional","affiliation":[{"name":"Hefei University of Technology, Anhui Provincial Industry Innovation Center of Humanoid Robots, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4556-0581","authenticated-orcid":false,"given":"Le","family":"Wu","sequence":"additional","affiliation":[{"name":"Innovation School of Artificial Intelligence, Hefei University of Technology, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-9141-3712","authenticated-orcid":false,"given":"Chenyi","family":"He","sequence":"additional","affiliation":[{"name":"Hefei University of Technology, Anhui Provincial Key Laboratory of Humanoid Robots, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-4646-500X","authenticated-orcid":false,"given":"Hao","family":"Liu","sequence":"additional","affiliation":[{"name":"Hefei University of Technology, Jianghuai Advance Technology Center, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8061-7486","authenticated-orcid":false,"given":"Zhi","family":"Li","sequence":"additional","affiliation":[{"name":"Shenzhen International Graduate School, Tsinghua University, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7333-5114","authenticated-orcid":false,"given":"Xin","family":"Li","sequence":"additional","affiliation":[{"name":"School of Information Science and Technology, University of Science and Technology of China, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-6872-3355","authenticated-orcid":false,"given":"Si","family":"Wei","sequence":"additional","affiliation":[{"name":"Artificial Intelligence Research Institute, iFLYTEK Company Ltd., Hefei, China"}]}],"member":"320","published-online":{"date-parts":[[2026,4,12]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"Honghui Bao Wenjie Wang Xinyu Lin Fengbin Zhu Teng Sun Fuli Feng and Tat Seng Chua. 2025. Heterogeneous User Modeling for LLM-based Recommendation. In RecSys."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/3604915.3608857"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10994-009--5152--4"},{"key":"e_1_3_2_1_4_1","volume-title":"European Conference on Computer Vision. Springer, 257--274","author":"Biggs Benjamin","year":"2024","unstructured":"Benjamin Biggs, Arjun Seshadri, Yang Zou, Achin Jain, Aditya Golatkar, Yusheng Xie, Alessandro Achille, Ashwin Swaminathan, and Stefano Soatto. 2024. Diffusion soup: Model merging for text-to-image diffusion models. In European Conference on Computer Vision. Springer, 257--274."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/3539597.3570366"},{"key":"e_1_3_2_1_6_1","unstructured":"Zeyu Cui Jianxin Ma Chang Zhou Jingren Zhou and Hongxia Yang. 2022. M6-Rec: Generative Pretrained Language Models are Open-Ended Recommender Systems. arXiv:2205.08084 [cs.IR] https:\/\/arxiv.org\/abs\/2205.08084"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/3604915.3610646"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.emnlp-main.64"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11704-019--8208-z"},{"key":"e_1_3_2_1_10_1","volume-title":"Proceedings of the 37th International Conference on Machine Learning (Proceedings of Machine Learning Research","volume":"3269","author":"Frankle Jonathan","year":"2020","unstructured":"Jonathan Frankle, Gintare Karolina Dziugaite, Daniel M Roy, and Michael Carbin. 2020. Linear Mode Connectivity and the Lottery Ticket Hypothesis. In Proceedings of the 37th International Conference on Machine Learning (Proceedings of Machine Learning Research, Vol. 119). PMLR, 3259--3269."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.5555\/3327546.3327556"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/3523227.3546767"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/3523227.3546767"},{"key":"e_1_3_2_1_14_1","volume-title":"Automated prompting for non-overlapping cross-domain sequential recommendation. arXiv preprint arXiv:2304.04218","author":"Guo Lei","year":"2023","unstructured":"Lei Guo, Chunxiao Wang, Xinhua Wang, Lei Zhu, and Hongzhi Yin. 2023. Automated prompting for non-overlapping cross-domain sequential recommendation. arXiv preprint arXiv:2304.04218 (2023)."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2022.3185101"},{"key":"e_1_3_2_1_16_1","volume-title":"4th International Conference on Learning Representations, ICLR","author":"Hidasi Bal\u00e1zs","year":"2016","unstructured":"Bal\u00e1zs Hidasi, Alexandros Karatzoglou, Linas Baltrunas, and Domonkos Tikk. 2016. Session-based Recommendations with Recurrent Neural Networks. In 4th International Conference on Learning Representations, ICLR 2016, San Juan, Puerto Rico, May 2--4, 2016, Conference Track Proceedings, Yoshua Bengio and Yann LeCun (Eds.). http:\/\/arxiv.org\/abs\/1511.06939"},{"key":"e_1_3_2_1_17_1","unstructured":"Min Hou Chenxi Bai Le Wu Hao Liu Kai Zhang Weiwen Liu Richang Hong Ruiming Tang and Meng Wang. 2025. RecCocktail: A Generalizable and Efficient Framework for LLM-Based Recommendation. arXiv:2502.08271 [cs.IR] https:\/\/arxiv.org\/abs\/2502.08271"},{"key":"e_1_3_2_1_18_1","volume-title":"A survey on generative recommendation: Data, model, and tasks. arXiv preprint arXiv:2510.27157","author":"Hou Min","year":"2025","unstructured":"Min Hou, Le Wu, Yuxin Liao, Yonghui Yang, Zhen Zhang, Changlong Zheng, Han Wu, and Richang Hong. 2025. A survey on generative recommendation: Data, model, and tasks. arXiv preprint arXiv:2510.27157 (2025)."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"crossref","unstructured":"Yupeng Hou Zhankui He Julian McAuley and Wayne Xin Zhao. 2023. Learning Vector-Quantized Item Representation for Transferable Sequential Recommenders. In The WebConf.","DOI":"10.1145\/3543507.3583434"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3534678.3539381"},{"key":"e_1_3_2_1_21_1","first-page":"3","article-title":"Lora: Low-rank adaptation of large language models","volume":"1","author":"Hu Edward J","year":"2022","unstructured":"Edward J Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, Weizhu Chen, et al. 2022. Lora: Low-rank adaptation of large language models. ICLR 1, 2 (2022), 3.","journal-title":"ICLR"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","unstructured":"Xinshuo Hu Dongfang Li Baotian Hu Zihao Zheng Zhenyu Liu and Min Zhang. 2024. Separate the wheat from the chaff: model deficiency unlearning via parameter-efficient module operation. In Proceedings of the Thirty-Eighth AAAI Conference on Artificial Intelligence and Thirty-Sixth Conference on Innovative Applications of Artificial Intelligence and Fourteenth Symposium on Educational Advances in Artificial Intelligence (AAAI'24\/IAAI'24\/EAAI'24). AAAI Press Article 2036 9 pages. doi:10.1609\/aaai.v38i16.29784","DOI":"10.1609\/aaai.v38i16.29784"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-acl.67"},{"key":"e_1_3_2_1_24_1","volume-title":"Mitchell Wortsman, Suchin Gururangan, Ludwig Schmidt, Hannaneh Hajishirzi, and Ali Farhadi.","author":"Ilharco Gabriel","year":"2023","unstructured":"Gabriel Ilharco, Marco Tulio Ribeiro, Mitchell Wortsman, Suchin Gururangan, Ludwig Schmidt, Hannaneh Hajishirzi, and Ali Farhadi. 2023. Editing Models with Task Arithmetic. arXiv:2212.04089 [cs.LG] https:\/\/arxiv.org\/abs\/2212.04089"},{"key":"e_1_3_2_1_25_1","volume-title":"Averaging weights leads to wider optima and better generalization. arXiv preprint arXiv:1803.05407","author":"Izmailov Pavel","year":"2018","unstructured":"Pavel Izmailov, Dmitrii Podoprikhin, Timur Garipov, Dmitry Vetrov, and Andrew Gordon Wilson. 2018. Averaging weights leads to wider optima and better generalization. arXiv preprint arXiv:1803.05407 (2018)."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2018.00035"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599519"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3678004"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/3637528.3671884"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1145\/3641860"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","unstructured":"Jing Liu Lele Sun Weizhi Nie Peiguang Jing and Yuting Su. 2024. Graph disentangled contrastive learning with personalized transfer for cross-domain recommendation. In Proceedings of the Thirty-Eighth AAAI Conference on Artificial Intelligence and Thirty-Sixth Conference on Innovative Applications of Artificial Intelligence and Fourteenth Symposium on Educational Advances in Artificial Intelligence (AAAI'24\/IAAI'24\/EAAI'24). AAAI Press Article 975 9 pages. doi:10.1609\/aaai.v38i8.28723","DOI":"10.1609\/aaai.v38i8.28723"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/3701716.3717850"},{"key":"e_1_3_2_1_33_1","volume-title":"High-quality Instruction Data. In Forty-first International Conference on Machine Learning. https:\/\/openreview.net\/forum?id=LWRI4uPG2X","author":"Peng Bo","year":"2024","unstructured":"Bo Peng, Xinyi Ling, Ziru Chen, Huan Sun, and Xia Ning. 2024. eCeLLM: Generalizing Large Language Models for E-commerce from Large-scale, High-quality Instruction Data. In Forty-first International Conference on Machine Learning. https:\/\/openreview.net\/forum?id=LWRI4uPG2X"},{"key":"e_1_3_2_1_34_1","volume-title":"Proceedings of the 41st International Conference on Machine Learning","author":"Peng Bo","year":"2024","unstructured":"Bo Peng, Xinyi Ling, Ziru Chen, Huan Sun, and Xia Ning. 2024. eCeLLM: generalizing large language models for E-commerce from large-scale, high-quality instruction data. In Proceedings of the 41st International Conference on Machine Learning (Vienna, Austria) (ICML'24). JMLR.org, Article 1632, 43 pages."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/3604915.3608845"},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1145\/3357384.3357895"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3705727"},{"key":"e_1_3_2_1_38_1","volume-title":"Weight Averaging for Neural Networks and Local Resampling Schemes. In AAAI-96 Workshop on Integrating Multiple Learned Models.","author":"Utans Joachim","year":"1996","unstructured":"Joachim Utans. 1996. Weight Averaging for Neural Networks and Local Resampling Schemes. In AAAI-96 Workshop on Integrating Multiple Learned Models."},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"crossref","unstructured":"Xiaolei Wang Xinyu Tang Xin Zhao Jingyuan Wang and Ji-Rong Wen. 2023. Rethinking the Evaluation for Conversational Recommendation in the Era of Large Language Models. In The 2023 Conference on Empirical Methods in Natural Language Processing. https:\/\/openreview.net\/forum?id=O4kDO3yS9B","DOI":"10.18653\/v1\/2023.emnlp-main.621"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2024.3511602"},{"key":"e_1_3_2_1_41_1","volume-title":"Proceedings of the 39th International Conference on Machine Learning (Proceedings of Machine Learning Research","volume":"23998","author":"Wortsman Mitchell","year":"2022","unstructured":"Mitchell Wortsman, Gabriel Ilharco, Samir Ya Gadre, Rebecca Roelofs, Raphael Gontijo-Lopes, Ari S Morcos, Hongseok Namkoong, Ali Farhadi, Yair Carmon, Simon Kornblith, and Ludwig Schmidt. 2022. Model soups: averaging weights of multiple fine-tuned models improves accuracy without increasing inference time. In Proceedings of the 39th International Conference on Machine Learning (Proceedings of Machine Learning Research, Vol. 162). PMLR, 23965--23998."},{"key":"e_1_3_2_1_42_1","volume-title":"Proceedings of the 39th International Conference on Machine Learning (Proceedings of Machine Learning Research","volume":"23998","author":"Wortsman Mitchell","year":"2022","unstructured":"Mitchell Wortsman, Gabriel Ilharco, Samir Ya Gadre, Rebecca Roelofs, Raphael Gontijo-Lopes, Ari S Morcos, Hongseok Namkoong, Ali Farhadi, Yair Carmon, Simon Kornblith, and Ludwig Schmidt. 2022. Model soups: averaging weights of multiple fine-tuned models improves accuracy without increasing inference time. In Proceedings of the 39th International Conference on Machine Learning (Proceedings of Machine Learning Research, Vol. 162), Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvari, Gang Niu, and Sivan Sabato (Eds.). PMLR, 23965--23998. https:\/\/proceedings.mlr.press\/v162\/wortsman22a.html"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11280-024-01291--2"},{"key":"e_1_3_2_1_44_1","first-page":"7093","article-title":"Ties-merging: Resolving interference when merging models","volume":"36","author":"Yadav Prateek","year":"2023","unstructured":"Prateek Yadav, Derek Tam, Leshem Choshen, Colin A Raffel, and Mohit Bansal. 2023. Ties-merging: Resolving interference when merging models. Advances in Neural Information Processing Systems 36 (2023), 7093--7115.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_45_1","volume-title":"Theories, Applications and Opportunities. arXiv preprint arXiv:2408.07666","author":"Yang Enneng","year":"2024","unstructured":"Enneng Yang, Li Shen, Guibing Guo, Xingwei Wang, Xiaochun Cao, Jie Zhang, and Dacheng Tao. 2024. Model Merging in LLMs, MLLMs, and Beyond: Methods, Theories, Applications and Opportunities. arXiv preprint arXiv:2408.07666 (2024)."},{"key":"e_1_3_2_1_46_1","volume-title":"AdaMerging: Adaptive Model Merging for Multi-Task Learning. The Twelfth International Conference on Learning Representations","author":"Yang Enneng","year":"2024","unstructured":"Enneng Yang, Zhenyi Wang, Li Shen, Shiwei Liu, Guibing Guo, Xingwei Wang, and Dacheng Tao. 2024. AdaMerging: Adaptive Model Merging for Multi-Task Learning. The Twelfth International Conference on Learning Representations (2024)."},{"key":"e_1_3_2_1_47_1","volume-title":"International Conference on Machine Learning. PMLR.","author":"Yu Le","year":"2024","unstructured":"Le Yu, Bowen Yu, Haiyang Yu, Fei Huang, and Yongbin Li. 2024. Language Models are Super Mario: Absorbing Abilities from Homologous Models as a Free Lunch. In International Conference on Machine Learning. PMLR."},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"crossref","unstructured":"Kerem Zaman Leshem Choshen and Shashank Srivastava. 2024. Fuse to Forget: Bias Reduction and Selective Memorization through Model Fusion. arXiv:2311.07682 [cs.CL] https:\/\/arxiv.org\/abs\/2311.07682","DOI":"10.18653\/v1\/2024.emnlp-main.1045"},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1145\/3548455"},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"crossref","unstructured":"Hao Zhang Mingyue Cheng Qi Liu Junzhe Jiang Xianquan Wang Rujiao Zhang Chenyi Lei and Enhong Chen. 2025. A Comprehensive Survey on Cross-Domain Recommendation: Taxonomy Progress and Prospects. arXiv:2503.14110 [cs.IR] https:\/\/arxiv.org\/abs\/2503.14110","DOI":"10.2139\/ssrn.5262197"},{"key":"e_1_3_2_1_51_1","volume-title":"Proceedings of the 37th International Conference on Neural Information Processing Systems","author":"Zhang Jinghan","year":"2023","unstructured":"Jinghan Zhang, Shiqi Chen, Junteng Liu, and Junxian He. 2023. Composing parameter-efficient modules with arithmetic operations. In Proceedings of the 37th International Conference on Neural Information Processing Systems (New Orleans, LA, USA) (NIPS '23). Curran Associates Inc., Red Hook, NY, USA, Article 552, 22 pages."},{"key":"e_1_3_2_1_52_1","volume-title":"SimCDR: Preserving Intra-Domain Similarities of Users for Cross-Domain Recommendation. ACM Transactions on Information Systems","author":"Zhang Kai","year":"2025","unstructured":"Kai Zhang, Zhihong Pan, Lei Yu, Qi Liu, Hongke Zhao, Chaochao Chen, and Enhong Chen. 2025. SimCDR: Preserving Intra-Domain Similarities of Users for Cross-Domain Recommendation. ACM Transactions on Information Systems (2025)."},{"key":"e_1_3_2_1_53_1","unstructured":"Wayne Xin Zhao Kun Zhou Junyi Li Tianyi Tang Xiaolei Wang Yupeng Hou Yingqian Min Beichen Zhang Junjie Zhang Zican Dong et al. 2023. A survey of large language models. arXiv preprint arXiv:2303.18223 (2023)."},{"key":"e_1_3_2_1_54_1","volume-title":"Merging loras like playing lego: Pushing the modularity of lora to extremes through rank-wise clustering. arXiv preprint arXiv:2409.16167","author":"Zhao Ziyu","year":"2024","unstructured":"Ziyu Zhao, Tao Shen, Didi Zhu, Zexi Li, Jing Su, Xuwu Wang, Kun Kuang, and Fei Wu. 2024. Merging loras like playing lego: Pushing the modularity of lora to extremes through rank-wise clustering. arXiv preprint arXiv:2409.16167 (2024)."},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1145\/3485447.3512111"},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2021\/639"}],"event":{"name":"WWW '26: The ACM Web Conference 2026","location":"Dubai United Arab Emirates","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"]},"container-title":["Proceedings of the ACM Web Conference 2026"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3774904.3792365","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,27]],"date-time":"2026-04-27T12:48:09Z","timestamp":1777294089000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3774904.3792365"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,4,12]]},"references-count":56,"alternative-id":["10.1145\/3774904.3792365","10.1145\/3774904"],"URL":"https:\/\/doi.org\/10.1145\/3774904.3792365","relation":{},"subject":[],"published":{"date-parts":[[2026,4,12]]},"assertion":[{"value":"2026-04-12","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}