{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,21]],"date-time":"2025-11-21T12:46:50Z","timestamp":1763729210568,"version":"3.45.0"},"reference-count":106,"publisher":"Association for Computing Machinery (ACM)","issue":"2","funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["624B2096, 62177033"],"award-info":[{"award-number":["624B2096, 62177033"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Science and Technology Development Fund of Macao","award":["0004\/2024\/E1B1"],"award-info":[{"award-number":["0004\/2024\/E1B1"]}]},{"name":"Huawei Innovation Research Program"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Recomm. Syst."],"published-print":{"date-parts":[[2026,6,30]]},"abstract":"<jats:p>\n                    As large language models (LLMs) achieve remarkable success in natural language processing (NLP) domains, LLM-enhanced recommender systems have received much attention and are being actively explored currently. In this article, we focus on adapting and enhancing large language models for recommendation tasks. First and foremost, we identify and formulate\n                    <jats:italic toggle=\"yes\">the lifelong sequential behavior incomprehension problem<\/jats:italic>\n                    for LLMs in recommendation realms, i.e., LLMs fail to effectively extract useful information from a pure textual context of long user behavior sequence, even if the length of context is well below the context limitation of LLMs. To address such an issue and improve the recommendation performance of LLMs, we propose a novel framework, namely,\n                    <jats:underline>\n                      <jats:bold>R<\/jats:bold>\n                    <\/jats:underline>\n                    etrieval-\n                    <jats:underline>\n                      <jats:bold>e<\/jats:bold>\n                    <\/jats:underline>\n                    nhanced\n                    <jats:underline>\n                      <jats:bold>L<\/jats:bold>\n                    <\/jats:underline>\n                    arge\n                    <jats:underline>\n                      <jats:bold>La<\/jats:bold>\n                    <\/jats:underline>\n                    nguage models\n                    <jats:bold>\n                      <jats:underline>Plus<\/jats:underline>\n                    <\/jats:bold>\n                    (ReLLaX), which provides full-stack optimization from three perspectives, i.e., data, prompt, and parameter. For data-level enhancement, we design semantic user behavior retrieval (SUBR) to reduce the heterogeneity of the behavior sequence, thus lowering the difficulty for LLMs to extract the essential information from user behavior sequences. Although SUBR can improve the data quality, further increase in the sequence length will still raise its heterogeneity to a level where LLMs can no longer comprehend it. Hence, we further propose to perform prompt-level and parameter-level enhancement, with the integration of conventional recommendation models (CRMs). As for prompt-level enhancement, we apply soft prompt augmentation (SPA) to explicitly inject collaborative knowledge from CRMs into the prompt. The item representations of LLMs are thus more aligned with recommendation, helping LLMs better explore the item relationships in the sequence and facilitating comprehension. Finally, for parameter-level enhancement, we propose component fully-interactive LoRA (CFLoRA). By enabling sufficient interaction between the LoRA atom components, the expressive ability of LoRA is extended, making the parameters effectively capture more sequence information. Moreover, we present new perspectives to compare current LoRA-based LLM4Rec methods, i.e., from both a composite and a decomposed view. We theoretically demonstrate that the ways they employ LoRA for recommendation are degraded versions of our CFLoRA, with different constraints on atom component interactions. Extensive experiments are conducted on three real-world public datasets to demonstrate the superiority of ReLLaX compared with existing baseline models, as well as its capability to alleviate lifelong sequential behavior incomprehension. Our code is available.\n                    <jats:xref ref-type=\"fn\">\n                      <jats:sup>1<\/jats:sup>\n                    <\/jats:xref>\n                  <\/jats:p>","DOI":"10.1145\/3766553","type":"journal-article","created":{"date-parts":[[2025,9,8]],"date-time":"2025-09-08T12:42:36Z","timestamp":1757335356000},"page":"1-33","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Full-Stack Optimized Large Language Models for Lifelong Sequential Behavior Comprehension in Recommendation"],"prefix":"10.1145","volume":"4","author":[{"ORCID":"https:\/\/orcid.org\/0009-0006-8905-1817","authenticated-orcid":false,"given":"Rong","family":"Shan","sequence":"first","affiliation":[{"name":"Computer Science, Shanghai Jiao Tong University","place":["Shanghai, China"]}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9140-1429","authenticated-orcid":false,"given":"Jiachen","family":"Zhu","sequence":"additional","affiliation":[{"name":"Computer Science, Shanghai Jiao Tong University","place":["Shanghai, China"]}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8953-3203","authenticated-orcid":false,"given":"Jianghao","family":"Lin","sequence":"additional","affiliation":[{"name":"Computer Science, Shanghai Jiao Tong University","place":["Shanghai, China"]}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8320-6845","authenticated-orcid":false,"given":"Chenxu","family":"Zhu","sequence":"additional","affiliation":[{"name":"Huawei Noah's Ark Lab, Huawei Technologies Co Ltd","place":["Shanghai, China"]}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3750-2533","authenticated-orcid":false,"given":"Bo","family":"Chen","sequence":"additional","affiliation":[{"name":"Huawei Noah's Ark Lab, Huawei Technologies Co Ltd","place":["Shenzhen, China"]}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9224-2431","authenticated-orcid":false,"given":"Ruiming","family":"Tang","sequence":"additional","affiliation":[{"name":"Huawei Noah's Ark Lab, Huawei Technologies Co Ltd","place":["Shenzhen, China"]}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4457-2820","authenticated-orcid":false,"given":"Yong","family":"Yu","sequence":"additional","affiliation":[{"name":"Computer Science, Shanghai Jiao Tong University","place":["Shanghai, China"]}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0127-2425","authenticated-orcid":false,"given":"Weinan","family":"Zhang","sequence":"additional","affiliation":[{"name":"Computer Science, Shanghai Jiao Tong University","place":["Shanghai, China"]}]}],"member":"320","published-online":{"date-parts":[[2025,11,21]]},"reference":[{"key":"e_1_3_3_2_2","unstructured":"Keqin Bao Ming Yan Yang Zhang Jizhi Zhang Wenjie Wang Fuli Feng and Xiangnan He. 2024. Real-time personalization for LLM-based recommendation with customized in-context learning. arXiv:2410.23136. Retrieved from https:\/\/arxiv.org\/abs\/2410.23136"},{"key":"e_1_3_3_3_2","doi-asserted-by":"crossref","unstructured":"Keqin Bao Jizhi Zhang Yang Zhang Wenjie Wang Fuli Feng and Xiangnan He. 2023. Tallrec: An effective and efficient tuning framework to align large language model with recommendation. arXiv:2305.00447. Retrieved from https:\/\/arxiv.org\/abs\/2305.00447","DOI":"10.1145\/3604915.3608857"},{"key":"e_1_3_3_4_2","unstructured":"Vadim Borisov Kathrin Se\u00dfler Tobias Leemann Martin Pawelczyk and Gjergji Kasneci. 2022. Language models are realistic tabular data generators. arXiv:2210.06280. Retrieved from https:\/\/arxiv.org\/abs\/2210.06280"},{"key":"e_1_3_3_5_2","first-page":"1877","article-title":"Language models are few-shot learners","volume":"33","author":"Brown Tom","year":"2020","unstructured":"Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D. Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et\u00a0al. 2020. Language models are few-shot learners. Advances in Neural Information Processing Systems 33 (2020), 1877\u20131901.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_3_6_2","unstructured":"Aldo Gael Carranza Rezsa Farahani Natalia Ponomareva Alex Kurakin Matthew Jagielski and Milad Nasr. 2023. Privacy-preserving recommender systems with synthetic query generation using differentially private large language models. arXiv:2305.05973. Retrieved from https:\/\/arxiv.org\/abs\/2305.05973"},{"issue":"4","key":"e_1_3_3_7_2","article-title":"Training and testing low-degree polynomial data mappings via linear SVM.","volume":"11","author":"Chang Yin-Wen","year":"2010","unstructured":"Yin-Wen Chang, Cho-Jui Hsieh, Kai-Wei Chang, Michael Ringgaard, and Chih-Jen Lin. 2010. Training and testing low-degree polynomial data mappings via linear SVM. Journal of Machine Learning Research 11, 4 (2010).","journal-title":"Journal of Machine Learning Research"},{"key":"e_1_3_3_8_2","unstructured":"Qiwei Chen Changhua Pei Shanshan Lv Chao Li Junfeng Ge and Wenwu Ou. 2021. End-to-end user behavior retrieval in click-through rateprediction model. arXiv:2108.04468. Retrieved from https:\/\/arxiv.org\/abs\/2108.04468"},{"key":"e_1_3_3_9_2","doi-asserted-by":"publisher","DOI":"10.1145\/3159652.3159668"},{"key":"e_1_3_3_10_2","unstructured":"Zheng Chen. 2023. PALR: Personalization aware LLMs for recommendation. arXiv:2305.07622. Retrieved from https:\/\/arxiv.org\/abs\/2305.07622"},{"key":"e_1_3_3_11_2","unstructured":"Zhenyi Lu Chenghao Fan and Jie Tian. 2023. Chinese-vicuna: A chinese instruction-following LLaMA-based model. Retrieved from https:\/\/github.com\/Facico\/Chinese-Vicuna"},{"key":"e_1_3_3_12_2","unstructured":"Wei-Lin Chiang Zhuohan Li Zi Lin Ying Sheng Zhanghao Wu Hao Zhang Lianmin Zheng Siyuan Zhuang Yonghao Zhuang Joseph E. Gonzalez Ion Stoica and Eric P. Xing. 2023. Vicuna: An Open-Source Chatbot Impressing GPT-4 with 90%* ChatGPT Quality. Retrieved from https:\/\/lmsys.org\/blog\/2023-03-30-vicuna\/"},{"key":"e_1_3_3_13_2","unstructured":"Konstantina Christakopoulou Alberto Lalama Cj Adams Iris Qu Yifat Amir Samer Chucri Pierce Vollucci Fabio Soldo Dina Bseiso Sarah Scodel et\u00a0al. 2023. Large language models for user interest journeys. arXiv:2305.15498. Retrieved from https:\/\/arxiv.org\/abs\/2305.15498"},{"key":"e_1_3_3_14_2","unstructured":"Junyoung Chung Caglar Gulcehre KyungHyun Cho and Yoshua Bengio. 2014. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv:1412.3555. Retrieved from https:\/\/arxiv.org\/abs\/1412.3555"},{"key":"e_1_3_3_15_2","unstructured":"Zeyu Cui Jianxin Ma Chang Zhou Jingren Zhou and Hongxia Yang. 2022. M6-rec: Generative pretrained language models are open-ended recommender systems. arXiv:2205.08084. Retrieved from https:\/\/arxiv.org\/abs\/2205.08084"},{"key":"e_1_3_3_16_2","unstructured":"Sunhao Dai Ninglu Shao Haiyuan Zhao Weijie Yu Zihua Si Chen Xu Zhongxiang Sun Xiao Zhang and Jun Xu. 2023. Uncovering ChatGPT\u2019s capabilities in recommender systems. arXiv:2305.02182. Retrieved from https:\/\/arxiv.org\/abs\/2305.02182"},{"key":"e_1_3_3_17_2","doi-asserted-by":"publisher","DOI":"10.1145\/3442381.3449913"},{"key":"e_1_3_3_18_2","unstructured":"Jacob Devlin Ming-Wei Chang Kenton Lee and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv:1810.04805. Retrieved from https:\/\/arxiv.org\/abs\/1810.04805"},{"key":"e_1_3_3_19_2","unstructured":"Junchen Fu Fajie Yuan Yu Song Zheng Yuan Mingyue Cheng Shenghui Cheng Jiaqi Zhang Jie Wang and Yunzhu Pan. 2023. Exploring adapter-based transfer learning for recommender systems: Empirical studies and practical insights. arXiv:2305.15036. Retrieved from https:\/\/arxiv.org\/abs\/2305.15036"},{"key":"e_1_3_3_20_2","doi-asserted-by":"publisher","DOI":"10.1145\/3539597.3570365"},{"key":"e_1_3_3_21_2","doi-asserted-by":"publisher","DOI":"10.1145\/3523227.3546767"},{"key":"e_1_3_3_22_2","doi-asserted-by":"publisher","DOI":"10.1145\/3523227.3546767"},{"key":"e_1_3_3_23_2","doi-asserted-by":"crossref","unstructured":"Shijie Geng Juntao Tan Shuchang Liu Zuohui Fu and Yongfeng Zhang. 2023. VIP5: Towards multimodal foundation models for recommendation. arXiv:2305.14302. Retrieved from https:\/\/arxiv.org\/abs\/2305.14302","DOI":"10.18653\/v1\/2023.findings-emnlp.644"},{"key":"e_1_3_3_24_2","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2017\/239"},{"key":"e_1_3_3_25_2","article-title":"Recurrent neural networks with top-k gains for session-based recommendations","author":"Hidasi Bal\u00e1zs","year":"2018","unstructured":"Bal\u00e1zs Hidasi and Alexandros Karatzoglou. 2018. Recurrent neural networks with top-k gains for session-based recommendations. CIKM (2018).","journal-title":"CIKM"},{"key":"e_1_3_3_26_2","volume-title":"Proceedings of the ICLR","author":"Hidasi Bal\u00e1zs","year":"2016","unstructured":"Bal\u00e1zs Hidasi, Alexandros Karatzoglou, Linas Baltrunas, and Domonkos Tikk. 2016. Session-based recommendations with recurrent neural networks. In Proceedings of the ICLR."},{"key":"e_1_3_3_27_2","doi-asserted-by":"publisher","DOI":"10.1145\/3543507.3583434"},{"key":"e_1_3_3_28_2","doi-asserted-by":"publisher","DOI":"10.1145\/3534678.3539381"},{"key":"e_1_3_3_29_2","unstructured":"Yupeng Hou Junjie Zhang Zihan Lin Hongyu Lu Ruobing Xie Julian McAuley and Wayne Xin Zhao. 2023. Large language models are zero-shot rankers for recommender systems. arXiv:2305.08845. Retrieved from https:\/\/arxiv.org\/abs\/2305.08845"},{"key":"e_1_3_3_30_2","unstructured":"Edward J. Hu Yelong Shen Phillip Wallis Zeyuan Allen-Zhu Yuanzhi Li Shean Wang Lu Wang and Weizhu Chen. 2021. Lora: Low-rank adaptation of large language models. arXiv:2106.09685. Retrieved from https:\/\/arxiv.org\/abs\/2106.09685"},{"key":"e_1_3_3_31_2","unstructured":"Wenyue Hua Yingqiang Ge Shuyuan Xu Jianchao Ji and Yongfeng Zhang. 2023. UP5: Unbiased foundation model for fairness-aware recommendation. arXiv:2305.12090. Retrieved from https:\/\/arxiv.org\/abs\/2305.12090"},{"key":"e_1_3_3_32_2","unstructured":"Wenyue Hua Shuyuan Xu Yingqiang Ge and Yongfeng Zhang. 2023. How to index item IDs for recommendation foundation models. arXiv:2305.06569. Retrieved from https:\/\/arxiv.org\/abs\/2305.06569"},{"key":"e_1_3_3_33_2","volume-title":"Proceedings of the SIGIR","author":"Huang Jin","year":"2018","unstructured":"Jin Huang, Wayne Xin Zhao, Hongjian Dou, Ji-Rong Wen, and Edward Y. Chang. 2018. Improving sequential recommendation with knowledge-enhanced memory networks. In Proceedings of the SIGIR."},{"key":"e_1_3_3_34_2","unstructured":"Albert Q. Jiang Alexandre Sablayrolles Arthur Mensch Chris Bamford Devendra Singh Chaplot Diego de las Casas Florian Bressand Gianna Lengyel Guillaume Lample Lucile Saulnier et\u00a0al. 2023. Mistral 7B. arXiv:2310.06825. Retrieved from https:\/\/arxiv.org\/abs\/2310.06825"},{"key":"e_1_3_3_35_2","article-title":"Self-attentive sequential recommendation","author":"Kang Wang-Cheng","year":"2018","unstructured":"Wang-Cheng Kang and Julian McAuley. 2018. Self-attentive sequential recommendation. ICDM (2018).","journal-title":"ICDM"},{"key":"e_1_3_3_36_2","unstructured":"Wang-Cheng Kang Jianmo Ni Nikhil Mehta Maheswaran Sathiamoorthy Lichan Hong Ed Chi and Derek Zhiyuan Cheng. 2023. Do LLMs understand user preferences? Evaluating LLMs on user rating prediction. arXiv:2305.06474. Retrieved from https:\/\/arxiv.org\/abs\/2305.06474"},{"key":"e_1_3_3_37_2","unstructured":"Xiaoyu Kong Jiancan Wu An Zhang Leheng Sheng Hui Lin Xiang Wang and Xiangnan He. 2024. Customizing language models with instance-wise LoRA for sequential recommendation. arXiv:2408.10159. Retrieved from https:\/\/arxiv.org\/abs\/2408.10159"},{"key":"e_1_3_3_38_2","unstructured":"Chen Li Yixiao Ge Jiayong Mao Dian Li and Ying Shan. 2023. TagGPT: Large language models are zero-shot multimodal taggers. arXiv:2304.03022. Retrieved from https:\/\/arxiv.org\/abs\/2304.03022"},{"key":"e_1_3_3_39_2","unstructured":"Jiacheng Li Ming Wang Jin Li Jinmiao Fu Xin Shen Jingbo Shang and Julian McAuley. 2023. Text is all you need: Learning language representations for sequential recommendation. arXiv:2305.13731. Retrieved from https:\/\/arxiv.org\/abs\/2305.13731"},{"key":"e_1_3_3_40_2","unstructured":"Ruyu Li Wenhao Deng Yu Cheng Zheng Yuan Jiaqi Zhang and Fajie Yuan. 2023. Exploring the upper limits of text-based collaborative filtering using large language models: Discoveries and insights. arXiv:2305.11700. Retrieved from https:\/\/arxiv.org\/abs\/2305.11700"},{"key":"e_1_3_3_41_2","unstructured":"Xinyi Li Yongfeng Zhang and Edward C. Malthouse. 2023. PBNR: Prompt-based news recommender system. arXiv:2304.07862. Retrieved from https:\/\/arxiv.org\/abs\/2304.07862"},{"key":"e_1_3_3_42_2","doi-asserted-by":"publisher","DOI":"10.1145\/3336191.3371785"},{"key":"e_1_3_3_43_2","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3220023"},{"key":"e_1_3_3_44_2","doi-asserted-by":"publisher","DOI":"10.1145\/3626772.3657690"},{"key":"e_1_3_3_45_2","doi-asserted-by":"publisher","DOI":"10.1145\/3589334.3645396"},{"key":"e_1_3_3_46_2","doi-asserted-by":"publisher","DOI":"10.1145\/3678004"},{"key":"e_1_3_3_47_2","doi-asserted-by":"publisher","DOI":"10.1145\/3404835.3462895"},{"key":"e_1_3_3_48_2","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599422"},{"key":"e_1_3_3_49_2","doi-asserted-by":"publisher","DOI":"10.1145\/3589334.3645467"},{"key":"e_1_3_3_50_2","first-page":"1119","volume-title":"Proceedings of the WWW","author":"Liu Bin","year":"2019","unstructured":"Bin Liu, Ruiming Tang, Yingzhi Chen, Jinkai Yu, Huifeng Guo, and Yuzhou Zhang. 2019. Feature generation by convolutional neural network for click-through rate prediction. In Proceedings of the WWW. 1119\u20131129."},{"key":"e_1_3_3_51_2","unstructured":"Guang Liu Jie Yang and Ledell Wu. 2022. PTab: Using the pre-trained language model for modeling tabular data. arXiv:2209.08060. Retrieved from https:\/\/arxiv.org\/abs\/2209.08060"},{"key":"e_1_3_3_52_2","unstructured":"Junling Liu Chao Liu Renjie Lv Kang Zhou and Yan Zhang. 2023. Is chatgpt a good recommender? A preliminary study. arXiv:2304.10149. Retrieved from https:\/\/arxiv.org\/abs\/2304.10149"},{"key":"e_1_3_3_53_2","unstructured":"Qijiong Liu Nuo Chen Tetsuya Sakai and Xiao-Ming Wu. 2023. A first look at LLM-powered generative news recommendation. arXiv:2305.06566. Retrieved from https:\/\/arxiv.org\/abs\/2305.06566"},{"key":"e_1_3_3_54_2","unstructured":"Ilya Loshchilov and Frank Hutter. 2017. Decoupled weight decay regularization. arXiv:1711.05101. Retrieved from https:\/\/arxiv.org\/abs\/1711.05101"},{"key":"e_1_3_3_55_2","unstructured":"Zhiming Mao Huimin Wang Yiming Du and Kam-fai Wong. 2023. UniTRec: A unified text-to-text transformer and joint contrastive learning framework for text-based recommendation. arXiv:2305.15756. Retrieved from https:\/\/arxiv.org\/abs\/2305.15756"},{"key":"e_1_3_3_56_2","volume-title":"Proceedings of the NeurIPS Efficient Natural Language and Speech Processing Workshop","author":"Muhamed Aashiq","year":"2021","unstructured":"Aashiq Muhamed, Iman Keivanloo, Sujan Perera, James Mracek, Yi Xu, Qingjun Cui, Santosh Rajagopalan, Belinda Zeng, and Trishul Chilimbi. 2021. CTR-BERT: Cost-effective knowledge distillation for billion-parameter teacher models. In Proceedings of the NeurIPS Efficient Natural Language and Speech Processing Workshop."},{"key":"e_1_3_3_57_2","volume-title":"Proceedings of the NeurIPS Efficient Natural Language and Speech Processing Workshop","author":"Muhamed Aashiq","year":"2021","unstructured":"Aashiq Muhamed, Iman Keivanloo, Sujan Perera, James Mracek, Yi Xu, Qingjun Cui, Santosh Rajagopalan, Belinda Zeng, and Trishul Chilimbi. 2021. CTR-BERT: Cost-effective knowledge distillation for billion-parameter teacher models. In Proceedings of the NeurIPS Efficient Natural Language and Speech Processing Workshop."},{"key":"e_1_3_3_58_2","doi-asserted-by":"crossref","unstructured":"Sheshera Mysore Andrew McCallum and Hamed Zamani. 2023. Large language model augmented narrative driven recommendations. arXiv:2306.02250. Retrieved from https:\/\/arxiv.org\/abs\/2306.02250","DOI":"10.1145\/3604915.3608829"},{"key":"e_1_3_3_59_2","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599799"},{"key":"e_1_3_3_60_2","volume-title":"Proceedings of the 13th International Conference on Learning Representations","author":"Pan Renjie","year":"2025","unstructured":"Renjie Pan, Jihao Dong, and Hua Yang. 2025. Discovering clone negatives via adaptive contrastive learning for image-text matching. In Proceedings of the 13th International Conference on Learning Representations."},{"key":"e_1_3_3_61_2","doi-asserted-by":"publisher","DOI":"10.3390\/app14052074"},{"key":"e_1_3_3_62_2","unstructured":"Aleksandr V. Petrov and Craig Macdonald. 2023. Generative sequential recommendation with GPTRec. arXiv:2306.11114. Retrieved from https:\/\/arxiv.org\/abs\/2306.11114"},{"key":"e_1_3_3_63_2","doi-asserted-by":"publisher","DOI":"10.1145\/3292500.3330666"},{"key":"e_1_3_3_64_2","doi-asserted-by":"publisher","DOI":"10.1145\/3340531.3412744"},{"key":"e_1_3_3_65_2","doi-asserted-by":"publisher","DOI":"10.1145\/3340531.3412744"},{"key":"e_1_3_3_66_2","doi-asserted-by":"publisher","DOI":"10.1145\/3447548.3467216"},{"key":"e_1_3_3_67_2","volume-title":"Proceedings of the SIGIR","author":"Qin Jiarui","year":"2020","unstructured":"Jiarui Qin, W. Zhang, Xin Wu, Jiarui Jin, Yuchen Fang, and Y. Yu. 2020. User behavior retrieval for click-through rate prediction. In Proceedings of the SIGIR."},{"key":"e_1_3_3_68_2","unstructured":"Libo Qin Qiguang Chen Xiachong Feng Yang Wu Yongheng Zhang Yinghui Li Min Li Wanxiang Che and Philip S Yu. 2024. Large language models meet nlp: A survey. arXiv:2405.12819. Retrieved from https:\/\/arxiv.org\/abs\/2405.12819"},{"key":"e_1_3_3_69_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i5.16557"},{"key":"e_1_3_3_70_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2016.0151"},{"key":"e_1_3_3_71_2","doi-asserted-by":"publisher","DOI":"10.5555\/3455716.3455856"},{"key":"e_1_3_3_72_2","article-title":"Lifelong sequential modeling with personalized memorization for user response prediction","author":"Ren Kan","year":"2019","unstructured":"Kan Ren, Jiarui Qin, Yuchen Fang, Weinan Zhang, Lei Zheng, Weijie Bian, Guorui Zhou, Jian Xu, Yong Yu, Xiaoqiang Zhu, et\u00a0al. 2019. Lifelong sequential modeling with personalized memorization for user response prediction. SIGIR.","journal-title":"SIGIR"},{"key":"e_1_3_3_73_2","unstructured":"Ruiyang Ren Yuhao Wang Kun Zhou Wayne Xin Zhao Wenjie Wang Jing Liu Ji-Rong Wen and Tat-Seng Chua. 2024. Self-calibrated listwise reranking with large language models. arXiv:2411.04602. Retrieved from https:\/\/arxiv.org\/abs\/2411.04602"},{"key":"e_1_3_3_74_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2010.127"},{"key":"e_1_3_3_75_2","unstructured":"Rong Shan Jianghao Lin Chenxu Zhu Bo Chen Menghui Zhu Kangning Zhang Jieming Zhu Ruiming Tang Yong Yu and Weinan Zhang. 2024. An automatic graph construction framework based on large language models for recommendation. arXiv:2412.18241. Retrieved from https:\/\/arxiv.org\/abs\/2412.18241"},{"key":"e_1_3_3_76_2","unstructured":"Jonathon Shlens. 2014. A tutorial on principal component analysis. arXiv:1404.1100. Retrieved from https:\/\/arxiv.org\/abs\/1404.1100"},{"key":"e_1_3_3_77_2","doi-asserted-by":"publisher","DOI":"10.1145\/3627673.3680030"},{"key":"e_1_3_3_78_2","doi-asserted-by":"publisher","DOI":"10.1145\/3357384.3357925"},{"key":"e_1_3_3_79_2","doi-asserted-by":"crossref","unstructured":"Weiwei Sun Lingyong Yan Xinyu Ma Pengjie Ren Dawei Yin and Zhaochun Ren. 2023. Is ChatGPT good at search? Investigating large language models as re-ranking agent. arXiv:2304.09542. Retrieved from https:\/\/arxiv.org\/abs\/2304.09542","DOI":"10.18653\/v1\/2023.emnlp-main.923"},{"key":"e_1_3_3_80_2","doi-asserted-by":"publisher","DOI":"10.1145\/3159652.3159656"},{"key":"e_1_3_3_81_2","unstructured":"Hugo Touvron Thibaut Lavril Gautier Izacard Xavier Martinet Marie-Anne Lachaux Timoth\u00e9e Lacroix Baptiste Rozi\u00e8re Naman Goyal Eric Hambro Faisal Azhar et\u00a0al. 2023. Llama: Open and efficient foundation language models. arXiv:2302.13971. Retrieved from https:\/\/arxiv.org\/abs\/2302.13971"},{"key":"e_1_3_3_82_2","doi-asserted-by":"publisher","DOI":"10.1145\/3640457.3688106"},{"key":"e_1_3_3_83_2","unstructured":"Jie Wang Fajie Yuan Mingyue Cheng Joemon M. Jose Chenyun Yu Beibei Kong Xiangnan He Zhijin Wang Bo Hu and Zang Li. 2022. TransRec: Learning transferable recommendation from mixture-of-modality feedback. arXiv:2206.06190. Retrieved from https:\/\/arxiv.org\/abs\/2206.06190"},{"key":"e_1_3_3_84_2","unstructured":"Lei Wang and Ee-Peng Lim. 2023. Zero-shot next-item recommendation using large pretrained language models. arXiv:2304.03153. Retrieved from https:\/\/arxiv.org\/abs\/2304.03153"},{"key":"e_1_3_3_85_2","doi-asserted-by":"publisher","DOI":"10.1145\/3442381.3450078"},{"key":"e_1_3_3_86_2","doi-asserted-by":"publisher","DOI":"10.1145\/3627673.3679569"},{"key":"e_1_3_3_87_2","doi-asserted-by":"publisher","DOI":"10.1145\/3437963.3441800"},{"key":"e_1_3_3_88_2","doi-asserted-by":"crossref","unstructured":"Yancheng Wang Ziyan Jiang Zheng Chen Fan Yang Yingxue Zhou Eunah Cho Xing Fan Xiaojiang Huang Yanbin Lu and Yingzhen Yang. 2023. Recmind: Large language model powered agent for recommendation. arXiv:2308.14296. Retrieved from https:\/\/arxiv.org\/abs\/2308.14296","DOI":"10.18653\/v1\/2024.findings-naacl.271"},{"key":"e_1_3_3_89_2","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599324"},{"key":"e_1_3_3_90_2","first-page":"1075","volume-title":"Proceedings of the 16th ACM International Conference on Web Search and Data Mining","author":"Xi Yunjia","year":"2023","unstructured":"Yunjia Xi, Jianghao Lin, Weiwen Liu, Xinyi Dai, Weinan Zhang, Rui Zhang, Ruiming Tang, and Yong Yu. 2023. A bird\u2019s-eye view of reranking: from list level to page level. In Proceedings of the 16th ACM International Conference on Web Search and Data Mining. 1075\u20131083."},{"key":"e_1_3_3_91_2","unstructured":"Yunjia Xi Weiwen Liu Jianghao Lin Jieming Zhu Bo Chen Ruiming Tang Weinan Zhang Rui Zhang and Yong Yu. 2023. Towards open-world recommendation with knowledge augmentation from large language models. arXiv:2306.10933. Retrieved from https:\/\/arxiv.org\/abs\/2306.10933"},{"key":"e_1_3_3_92_2","doi-asserted-by":"publisher","DOI":"10.5555\/3172077.3172324"},{"key":"e_1_3_3_93_2","first-page":"3926","volume-title":"Proceedings of the IJCAI","author":"Xin Xin","year":"2019","unstructured":"Xin Xin, Bo Chen, Xiangnan He, Dong Wang, Yue Ding, and Joemon M. Jose. 2019. CFM: Convolutional factorization machines for context-aware recommendation. In Proceedings of the IJCAI. 3926\u20133932."},{"key":"e_1_3_3_94_2","unstructured":"Jiyuan Yang Yuanzi Li Jingyu Zhao Hanbing Wang Muyang Ma Jun Ma Zhaochun Ren Mengqi Zhang Xin Xin Zhumin Chen et\u00a0al. 2024. Uncovering selective state space model\u2019s capabilities in lifelong sequential recommendation. arXiv:2403.16371. Retrieved from https:\/\/arxiv.org\/abs\/2403.16371"},{"key":"e_1_3_3_95_2","unstructured":"Yang Yu Fangzhao Wu Chuhan Wu Jingwei Yi and Qi Liu. 2021. Tiny-newsrec: Effective and efficient plm-based news recommendation. arXiv:2112.00944. Retrieved from https:\/\/arxiv.org\/abs\/2112.00944"},{"key":"e_1_3_3_96_2","doi-asserted-by":"crossref","unstructured":"Zheng Yuan Fajie Yuan Yu Song Youhua Li Junchen Fu Fei Yang Yunzhu Pan and Yongxin Ni. 2023. Where to go next for recommender systems? id-vs. modality-based recommender models revisited. arXiv:2303.13835. Retrieved from https:\/\/arxiv.org\/abs\/2303.13835","DOI":"10.1145\/3539618.3591932"},{"key":"e_1_3_3_97_2","doi-asserted-by":"crossref","unstructured":"Jizhi Zhang Keqin Bao Yang Zhang Wenjie Wang Fuli Feng and Xiangnan He. 2023. Is chatgpt fair for recommendation? Evaluating fairness in large language model recommendation. arXiv:2305.07609. Retrieved from https:\/\/arxiv.org\/abs\/2305.07609","DOI":"10.1145\/3604915.3608860"},{"key":"e_1_3_3_98_2","unstructured":"Junjie Zhang Ruobing Xie Yupeng Hou Wayne Xin Zhao Leyu Lin and Ji-Rong Wen. 2023. Recommendation as instruction following: A large language model empowered recommendation approach. arXiv:2305.07001. Retrieved from https:\/\/arxiv.org\/abs\/2305.07001"},{"key":"e_1_3_3_99_2","unstructured":"Kai Zhang Fubang Zhao Yangyang Kang and Xiaozhong Liu. 2023. Memory-augmented llm personalization with short-and long-term memory coordination. arXiv:2309.11696. Retrieved from https:\/\/arxiv.org\/abs\/2309.11696"},{"key":"e_1_3_3_100_2","article-title":"Deep learning for click-through rate estimation","author":"Zhang Weinan","year":"2021","unstructured":"Weinan Zhang, Jiarui Qin, Wei Guo, Ruiming Tang, and Xiuqiang He. 2021. Deep learning for click-through rate estimation. IJCAI (2021).","journal-title":"IJCAI"},{"key":"e_1_3_3_101_2","doi-asserted-by":"crossref","unstructured":"Xinyang Zhang Yury Malkov Omar Florez Serim Park Brian McWilliams Jiawei Han and Ahmed El-Kishky. 2022. TwHIN-BERT: A socially-enriched pre-trained language model for multilingual tweet representations. arXiv:2209.07562. Retrieved from https:\/\/arxiv.org\/abs\/2209.07562","DOI":"10.1145\/3580305.3599921"},{"key":"e_1_3_3_102_2","unstructured":"Yuhui Zhang Hao Ding Zeren Shui Yifei Ma James Zou Anoop Deoras and Hao Wang. 2021. Language models as recommender systems: Evaluations and limitations. (2021)."},{"key":"e_1_3_3_103_2","unstructured":"Yang Zhang Fuli Feng Jizhi Zhang Keqin Bao Qifan Wang and Xiangnan He. 2023. Collm: Integrating collaborative embeddings into large language models for recommendation. arXiv:2310.19488. Retrieved from https:\/\/arxiv.org\/abs\/2310.19488"},{"key":"e_1_3_3_104_2","unstructured":"Zizhuo Zhang and Bang Wang. 2023. Prompt learning for news recommendation. arXiv:2304.05263. Retrieved from https:\/\/arxiv.org\/abs\/2304.05263"},{"key":"e_1_3_3_105_2","first-page":"5941","volume-title":"Proceedings of the AAAI","author":"Zhou Guorui","year":"2019","unstructured":"Guorui Zhou, Na Mou, Ying Fan, Qi Pi, Weijie Bian, Chang Zhou, Xiaoqiang Zhu, and Kun Gai. 2019. Deep interest evolution network for click-through rate prediction. In Proceedings of the AAAI. 5941\u20135948."},{"key":"e_1_3_3_106_2","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3219823"},{"key":"e_1_3_3_107_2","unstructured":"Jiachen Zhu Jianghao Lin Xinyi Dai Bo Chen Rong Shan Jieming Zhu Ruiming Tang Yong Yu and Weinan Zhang. 2024. Lifelong personalized low-rank adaptation of large language models for recommendation. arXiv:2408.03533. Retrieved from https:\/\/arxiv.org\/abs\/2408.03533"}],"container-title":["ACM Transactions on Recommender Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3766553","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,21]],"date-time":"2025-11-21T12:33:00Z","timestamp":1763728380000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3766553"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,21]]},"references-count":106,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2026,6,30]]}},"alternative-id":["10.1145\/3766553"],"URL":"https:\/\/doi.org\/10.1145\/3766553","relation":{},"ISSN":["2770-6699"],"issn-type":[{"type":"electronic","value":"2770-6699"}],"subject":[],"published":{"date-parts":[[2025,11,21]]},"assertion":[{"value":"2025-01-14","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2025-08-26","order":2,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2025-11-21","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}