{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,10]],"date-time":"2026-04-10T10:05:52Z","timestamp":1775815552157,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":94,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,5,13]],"date-time":"2024-05-13T00:00:00Z","timestamp":1715558400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"Shanghai Municipal Science and Technology Major Project","award":["2021SHZDZX0102"],"award-info":[{"award-number":["2021SHZDZX0102"]}]},{"name":"National Key R&D Program of China","award":["2022ZD0114804"],"award-info":[{"award-number":["2022ZD0114804"]}]},{"DOI":"10.13039\/501100006374","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62177033, 62322603"],"award-info":[{"award-number":["62177033, 62322603"]}],"id":[{"id":"10.13039\/501100006374","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,5,13]]},"DOI":"10.1145\/3589334.3645467","type":"proceedings-article","created":{"date-parts":[[2024,5,8]],"date-time":"2024-05-08T07:08:13Z","timestamp":1715152093000},"page":"3497-3508","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":44,"title":["ReLLa: Retrieval-enhanced Large Language Models for Lifelong Sequential Behavior Comprehension in Recommendation"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8953-3203","authenticated-orcid":false,"given":"Jianghao","family":"Lin","sequence":"first","affiliation":[{"name":"Shanghai Jiao Tong University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-8905-1817","authenticated-orcid":false,"given":"Rong","family":"Shan","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8320-6845","authenticated-orcid":false,"given":"Chenxu","family":"Zhu","sequence":"additional","affiliation":[{"name":"Huawei Noah's Ark Lab, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2611-5055","authenticated-orcid":false,"given":"Kounianhua","family":"Du","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3750-2533","authenticated-orcid":false,"given":"Bo","family":"Chen","sequence":"additional","affiliation":[{"name":"Huawei Noah's Ark Lab, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-3761-5242","authenticated-orcid":false,"given":"Shigang","family":"Quan","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9224-2431","authenticated-orcid":false,"given":"Ruiming","family":"Tang","sequence":"additional","affiliation":[{"name":"Huawei Noah's Ark Lab, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0281-8271","authenticated-orcid":false,"given":"Yong","family":"Yu","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0127-2425","authenticated-orcid":false,"given":"Weinan","family":"Zhang","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University, Shanghai, China"}]}],"member":"320","published-online":{"date-parts":[[2024,5,13]]},"reference":[{"key":"e_1_3_2_2_1_1","volume-title":"Tallrec: An eective and ecient tuning framework to align large language model with recommendation. arXiv preprint arXiv:2305.00447","author":"Bao Keqin","year":"2023","unstructured":"Keqin Bao, Jizhi Zhang, Yang Zhang, Wenjie Wang, Fuli Feng, and Xiangnan He. 2023. Tallrec: An eective and ecient tuning framework to align large language model with recommendation. arXiv preprint arXiv:2305.00447 (2023)."},{"key":"e_1_3_2_2_2_1","volume-title":"Mixmatch: A holistic approach to semi-supervised learning. Advances in neural information processing systems 32","author":"Berthelot David","year":"2019","unstructured":"David Berthelot, Nicholas Carlini, Ian Goodfellow, Nicolas Papernot, Avital Oliver, and Colin A Rael. 2019. Mixmatch: A holistic approach to semi-supervised learning. Advances in neural information processing systems 32 (2019)."},{"key":"e_1_3_2_2_3_1","volume-title":"Language models are realistic tabular data generators. arXiv preprint arXiv:2210.06280","author":"Borisov Vadim","year":"2022","unstructured":"Vadim Borisov, Kathrin Se\u00dfler, Tobias Leemann, Martin Pawelczyk, and Gjergji Kasneci. 2022. Language models are realistic tabular data generators. arXiv preprint arXiv:2210.06280 (2022)."},{"key":"e_1_3_2_2_4_1","unstructured":"Tom Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared D Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell et al. 2020. Language models are few-shot learners. Advances in neural information processing systems 33 (2020) 1877--1901."},{"key":"e_1_3_2_2_5_1","volume-title":"Privacy-Preserving Recommender Systems with Synthetic Query Generation using Dierentially Private Large Language Models. arXiv preprint arXiv:2305.05973","author":"Carranza Aldo Gael","year":"2023","unstructured":"Aldo Gael Carranza, Rezsa Farahani, Natalia Ponomareva, Alex Kurakin, Matthew Jagielski, and Milad Nasr. 2023. Privacy-Preserving Recommender Systems with Synthetic Query Generation using Dierentially Private Large Language Models. arXiv preprint arXiv:2305.05973 (2023)."},{"key":"e_1_3_2_2_6_1","article-title":"Training and testing low-degree polynomial data mappings via linear SVM","volume":"11","author":"Chang Yin-Wen","year":"2010","unstructured":"Yin-Wen Chang, Cho-Jui Hsieh, Kai-Wei Chang, Michael Ringgaard, and Chih-Jen Lin. 2010. Training and testing low-degree polynomial data mappings via linear SVM. Journal of Machine Learning Research 11, 4 (2010).","journal-title":"Journal of Machine Learning Research"},{"key":"e_1_3_2_2_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/3459637.3481915"},{"key":"e_1_3_2_2_8_1","volume-title":"PALR: Personalization Aware LLMs for Recommendation. arXiv preprint arXiv:2305.07622","author":"Chen Zheng","year":"2023","unstructured":"Zheng Chen. 2023. PALR: Personalization Aware LLMs for Recommendation. arXiv preprint arXiv:2305.07622 (2023)."},{"key":"e_1_3_2_2_9_1","unstructured":"Zhenyi Lu Chenghao Fan and Jie Tian. 2023. Chinese-Vicuna: A Chinese Instruction-following LLaMA-based Model. https:\/\/github.com\/Facico\/Chinese- Vicuna"},{"key":"e_1_3_2_2_10_1","volume-title":"Xing","author":"Chiang Wei-Lin","year":"2023","unstructured":"Wei-Lin Chiang, Zhuohan Li, Zi Lin, Ying Sheng, Zhanghao Wu, Hao Zhang, Lianmin Zheng, Siyuan Zhuang, Yonghao Zhuang, Joseph E. Gonzalez, Ion Stoica, and Eric P. Xing. 2023. Vicuna: An Open-Source Chatbot Impressing GPT-4 with 90%* ChatGPT Quality. https:\/\/lmsys.org\/blog\/2023-03--30-vicuna\/"},{"key":"e_1_3_2_2_11_1","unstructured":"Konstantina Christakopoulou Alberto Lalama Cj Adams Iris Qu Yifat Amir Samer Chucri Pierce Vollucci Fabio Soldo Dina Bseiso Sarah Scodel et al. 2023. Large Language Models for User Interest Journeys. arXiv preprint arXiv:2305.15498 (2023)."},{"key":"e_1_3_2_2_12_1","volume-title":"Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555","author":"Chung Junyoung","year":"2014","unstructured":"Junyoung Chung, Caglar Gulcehre, KyungHyun Cho, and Yoshua Bengio. 2014. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555 (2014)."},{"key":"e_1_3_2_2_13_1","volume-title":"M6-rec: Generative pretrained language models are open-ended recommender systems. arXiv preprint arXiv:2205.08084","author":"Cui Zeyu","year":"2022","unstructured":"Zeyu Cui, Jianxin Ma, Chang Zhou, Jingren Zhou, and Hongxia Yang. 2022. M6-rec: Generative pretrained language models are open-ended recommender systems. arXiv preprint arXiv:2205.08084 (2022)."},{"key":"e_1_3_2_2_14_1","volume-title":"Uncovering ChatGPT's Capabilities in Recommender Systems. arXiv preprint arXiv:2305.02182","author":"Dai Sunhao","year":"2023","unstructured":"Sunhao Dai, Ninglu Shao, Haiyuan Zhao, Weijie Yu, Zihua Si, Chen Xu, Zhongxiang Sun, Xiao Zhang, and Jun Xu. 2023. Uncovering ChatGPT's Capabilities in Recommender Systems. arXiv preprint arXiv:2305.02182 (2023)."},{"key":"e_1_3_2_2_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442381.3449913"},{"key":"e_1_3_2_2_16_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)."},{"key":"e_1_3_2_2_17_1","volume-title":"A survey of data augmentation approaches for NLP. arXiv preprint arXiv:2105.03075","author":"Feng Steven Y","year":"2021","unstructured":"Steven Y Feng, Varun Gangal, Jason Wei, Sarath Chandar, Soroush Vosoughi, Teruko Mitamura, and Eduard Hovy. 2021. A survey of data augmentation approaches for NLP. arXiv preprint arXiv:2105.03075 (2021)."},{"key":"e_1_3_2_2_18_1","volume-title":"Exploring Adapter-based Transfer Learning for Recommender Systems: Empirical Studies and Practical Insights. arXiv preprint arXiv:2305.15036","author":"Fu Junchen","year":"2023","unstructured":"Junchen Fu, Fajie Yuan, Yu Song, Zheng Yuan, Mingyue Cheng, Shenghui Cheng, Jiaqi Zhang, Jie Wang, and Yunzhu Pan. 2023. Exploring Adapter-based Transfer Learning for Recommender Systems: Empirical Studies and Practical Insights. arXiv preprint arXiv:2305.15036 (2023)."},{"key":"e_1_3_2_2_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3539597.3570365"},{"key":"e_1_3_2_2_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3523227.3546767"},{"key":"e_1_3_2_2_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/3523227.3546767"},{"key":"e_1_3_2_2_22_1","volume-title":"VIP5: Towards Multimodal Foundation Models for Recommendation. arXiv preprint arXiv:2305.14302","author":"Geng Shijie","year":"2023","unstructured":"Shijie Geng, Juntao Tan, Shuchang Liu, Zuohui Fu, and Yongfeng Zhang. 2023. VIP5: Towards Multimodal Foundation Models for Recommendation. arXiv preprint arXiv:2305.14302 (2023)."},{"key":"e_1_3_2_2_23_1","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2017\/239"},{"key":"e_1_3_2_2_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/3524610.3527897"},{"key":"e_1_3_2_2_25_1","volume-title":"Recurrent neural networks with top-k gains for session-based recommendations. CIKM","author":"Hidasi Bal\u00e1zs","year":"2018","unstructured":"Bal\u00e1zs Hidasi and Alexandros Karatzoglou. 2018. Recurrent neural networks with top-k gains for session-based recommendations. CIKM (2018)."},{"key":"e_1_3_2_2_26_1","unstructured":"Bal\u00e1zs Hidasi Alexandros Karatzoglou Linas Baltrunas and Domonkos Tikk. 2016. Session-based recommendations with recurrent neural networks. In ICLR."},{"key":"e_1_3_2_2_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3543507.3583434"},{"key":"e_1_3_2_2_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3534678.3539381"},{"key":"e_1_3_2_2_29_1","volume-title":"Large language models are zero-shot rankers for recommender systems. arXiv preprint arXiv:2305.08845","author":"Hou Yupeng","year":"2023","unstructured":"Yupeng Hou, Junjie Zhang, Zihan Lin, Hongyu Lu, Ruobing Xie, Julian McAuley, and Wayne Xin Zhao. 2023. Large language models are zero-shot rankers for recommender systems. arXiv preprint arXiv:2305.08845 (2023)."},{"key":"e_1_3_2_2_30_1","volume-title":"Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685","author":"Hu Edward J","year":"2021","unstructured":"Edward J Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen. 2021. Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685 (2021)."},{"key":"e_1_3_2_2_31_1","volume-title":"UP5: Unbiased Foundation Model for Fairness-aware Recommendation. arXiv preprint arXiv:2305.12090","author":"Hua Wenyue","year":"2023","unstructured":"Wenyue Hua, Yingqiang Ge, Shuyuan Xu, Jianchao Ji, and Yongfeng Zhang. 2023. UP5: Unbiased Foundation Model for Fairness-aware Recommendation. arXiv preprint arXiv:2305.12090 (2023)."},{"key":"e_1_3_2_2_32_1","volume-title":"How to Index Item IDs for Recommendation Foundation Models. arXiv preprint arXiv:2305.06569","author":"Hua Wenyue","year":"2023","unstructured":"Wenyue Hua, Shuyuan Xu, Yingqiang Ge, and Yongfeng Zhang. 2023. How to Index Item IDs for Recommendation Foundation Models. arXiv preprint arXiv:2305.06569 (2023)."},{"key":"e_1_3_2_2_33_1","volume-title":"Self-Attentive Sequential Recommendation. ICDM","author":"Kang Wang-Cheng","year":"2018","unstructured":"Wang-Cheng Kang and Julian McAuley. 2018. Self-Attentive Sequential Recommendation. ICDM (2018)."},{"key":"e_1_3_2_2_34_1","volume-title":"Chi, and Derek Zhiyuan Cheng","author":"Kang Wang-Cheng","year":"2023","unstructured":"Wang-Cheng Kang, Jianmo Ni, Nikhil Mehta, Maheswaran Sathiamoorthy, Lichan Hong, Ed Chi, and Derek Zhiyuan Cheng. 2023. Do LLMs Understand User Preferences? Evaluating LLMs On User Rating Prediction. arXiv preprint arXiv:2305.06474 (2023)."},{"key":"e_1_3_2_2_35_1","volume-title":"International Conference on Machine Learning. PMLR, 11499--11528","author":"Korbak Tomasz","year":"2022","unstructured":"Tomasz Korbak, Hady Elsahar, German Kruszewski, and Marc Dymetman. 2022. Controlling conditional language models without catastrophic forgetting. In International Conference on Machine Learning. PMLR, 11499--11528."},{"key":"e_1_3_2_2_36_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.aiopen.2022.03.001"},{"key":"e_1_3_2_2_37_1","volume-title":"TagGPT: Large Language Models are Zero-shot Multimodal Taggers. arXiv preprint arXiv:2304.03022","author":"Li Chen","year":"2023","unstructured":"Chen Li, Yixiao Ge, Jiayong Mao, Dian Li, and Ying Shan. 2023. TagGPT: Large Language Models are Zero-shot Multimodal Taggers. arXiv preprint arXiv:2304.03022 (2023)."},{"key":"e_1_3_2_2_38_1","volume-title":"Text Is All You Need: Learning Language Representations for Sequential Recommendation. arXiv preprint arXiv:2305.13731","author":"Li Jiacheng","year":"2023","unstructured":"Jiacheng Li, Ming Wang, Jin Li, Jinmiao Fu, Xin Shen, Jingbo Shang, and Julian McAuley. 2023. Text Is All You Need: Learning Language Representations for Sequential Recommendation. arXiv preprint arXiv:2305.13731 (2023)."},{"key":"e_1_3_2_2_39_1","volume-title":"Exploring the Upper Limits of Text-Based Collaborative Filtering Using Large Language Models: Discoveries and Insights. arXiv preprint arXiv:2305.11700","author":"Li Ruyu","year":"2023","unstructured":"Ruyu Li, Wenhao Deng, Yu Cheng, Zheng Yuan, Jiaqi Zhang, and Fajie Yuan. 2023. Exploring the Upper Limits of Text-Based Collaborative Filtering Using Large Language Models: Discoveries and Insights. arXiv preprint arXiv:2305.11700 (2023)."},{"key":"e_1_3_2_2_40_1","volume-title":"CTRL: Connect Tabular and Language Model for CTR Prediction. arXiv preprint arXiv:2306.02841","author":"Li Xiangyang","year":"2023","unstructured":"Xiangyang Li, Bo Chen, Lu Hou, and Ruiming Tang. 2023. CTRL: Connect Tabular and Language Model for CTR Prediction. arXiv preprint arXiv:2306.02841 (2023)."},{"key":"e_1_3_2_2_41_1","volume-title":"PBNR: Prompt-based News Recommender System. arXiv preprint arXiv:2304.07862","author":"Li Xinyi","year":"2023","unstructured":"Xinyi Li, Yongfeng Zhang, and Edward C Malthouse. 2023. PBNR: Prompt-based News Recommender System. arXiv preprint arXiv:2304.07862 (2023)."},{"key":"e_1_3_2_2_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/3336191.3371785"},{"key":"e_1_3_2_2_43_1","doi-asserted-by":"crossref","unstructured":"Jianxun Lian Xiaohuan Zhou Fuzheng Zhang Zhongxia Chen Xing Xie and Guangzhong Sun. 2018. xdeepfm: Combining explicit and implicit feature interactions for recommender systems. In KDD. 1754--1763.","DOI":"10.1145\/3219819.3220023"},{"key":"e_1_3_2_2_44_1","unstructured":"Jianghao Lin Xinyi Dai Yunjia Xi Weiwen Liu Bo Chen Xiangyang Li Chenxu Zhu Huifeng Guo Yong Yu Ruiming Tang et al. 2023. How Can Recommender Systems Benefit from Large Language Models: A Survey. arXiv preprint arXiv:2306.05817 (2023)."},{"key":"e_1_3_2_2_45_1","doi-asserted-by":"publisher","DOI":"10.1145\/3404835.3462895"},{"key":"e_1_3_2_2_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599422"},{"key":"e_1_3_2_2_47_1","doi-asserted-by":"crossref","unstructured":"Bin Liu Ruiming Tang Yingzhi Chen Jinkai Yu Huifeng Guo and Yuzhou Zhang. 2019. Feature generation by convolutional neural network for click-through rate prediction. In WWW. 1119--1129.","DOI":"10.1145\/3308558.3313497"},{"key":"e_1_3_2_2_48_1","volume-title":"PTab: Using the Pre-trained Language Model for Modeling Tabular Data. arXiv preprint arXiv:2209.08060","author":"Liu Guang","year":"2022","unstructured":"Guang Liu, Jie Yang, and Ledell Wu. 2022. PTab: Using the Pre-trained Language Model for Modeling Tabular Data. arXiv preprint arXiv:2209.08060 (2022)."},{"key":"e_1_3_2_2_49_1","volume-title":"Is chatgpt a good recommender? a preliminary study. arXiv preprint arXiv:2304.10149","author":"Liu Junling","year":"2023","unstructured":"Junling Liu, Chao Liu, Renjie Lv, Kang Zhou, and Yan Zhang. 2023. Is chatgpt a good recommender? a preliminary study. arXiv preprint arXiv:2304.10149 (2023)."},{"key":"e_1_3_2_2_50_1","volume-title":"A First Look at LLM-Powered Generative News Recommendation. arXiv preprint arXiv:2305.06566","author":"Liu Qijiong","year":"2023","unstructured":"Qijiong Liu, Nuo Chen, Tetsuya Sakai, and Xiao-Ming Wu. 2023. A First Look at LLM-Powered Generative News Recommendation. arXiv preprint arXiv:2305.06566 (2023)."},{"key":"e_1_3_2_2_51_1","volume-title":"Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101","author":"Loshchilov Ilya","year":"2017","unstructured":"Ilya Loshchilov and Frank Hutter. 2017. Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)."},{"key":"e_1_3_2_2_52_1","volume-title":"UniTRec: A Unified Text-to-Text Transformer and Joint Contrastive Learning Framework for Text-based Recommendation. arXiv preprint arXiv:2305.15756","author":"Mao Zhiming","year":"2023","unstructured":"Zhiming Mao, Huimin Wang, Yiming Du, and Kam-fai Wong. 2023. UniTRec: A Unified Text-to-Text Transformer and Joint Contrastive Learning Framework for Text-based Recommendation. arXiv preprint arXiv:2305.15756 (2023)."},{"key":"e_1_3_2_2_53_1","volume-title":"NeurIPS Ecient Natural Language and Speech Processing Workshop.","author":"Muhamed Aashiq","year":"2021","unstructured":"Aashiq Muhamed, Iman Keivanloo, Sujan Perera, James Mracek, Yi Xu, Qingjun Cui, Santosh Rajagopalan, Belinda Zeng, and Trishul Chilimbi. 2021. CTR-BERT: Cost-eective knowledge distillation for billion-parameter teacher models. In NeurIPS Ecient Natural Language and Speech Processing Workshop."},{"key":"e_1_3_2_2_54_1","volume-title":"NeurIPS Ecient Natural Language and Speech Processing Workshop.","author":"Muhamed Aashiq","year":"2021","unstructured":"Aashiq Muhamed, Iman Keivanloo, Sujan Perera, James Mracek, Yi Xu, Qingjun Cui, Santosh Rajagopalan, Belinda Zeng, and Trishul Chilimbi. 2021. CTR-BERT: Cost-eective knowledge distillation for billion-parameter teacher models. In NeurIPS Ecient Natural Language and Speech Processing Workshop."},{"key":"e_1_3_2_2_55_1","volume-title":"Large Language Model Augmented Narrative Driven Recommendations. arXiv preprint arXiv:2306.02250","author":"Mysore Sheshera","year":"2023","unstructured":"Sheshera Mysore, Andrew McCallum, and Hamed Zamani. 2023. Large Language Model Augmented Narrative Driven Recommendations. arXiv preprint arXiv:2306.02250 (2023)."},{"key":"e_1_3_2_2_56_1","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599799"},{"key":"e_1_3_2_2_57_1","volume-title":"Generative Sequential Recommendation with GPTRec. arXiv preprint arXiv:2306.11114","author":"Petrov Aleksandr V","year":"2023","unstructured":"Aleksandr V Petrov and Craig Macdonald. 2023. Generative Sequential Recommendation with GPTRec. arXiv preprint arXiv:2306.11114 (2023)."},{"key":"e_1_3_2_2_58_1","doi-asserted-by":"crossref","unstructured":"Qi Pi Weijie Bian Guorui Zhou Xiaoqiang Zhu and Kun Gai. 2019. Practice on long sequential user behavior modeling for click-through rate prediction. In KDD. 2671--2679.","DOI":"10.1145\/3292500.3330666"},{"key":"e_1_3_2_2_59_1","doi-asserted-by":"publisher","DOI":"10.1145\/3340531.3412744"},{"key":"e_1_3_2_2_60_1","doi-asserted-by":"publisher","DOI":"10.1145\/3447548.3467216"},{"key":"e_1_3_2_2_61_1","unstructured":"Jiarui Qin W. Zhang Xin Wu Jiarui Jin Yuchen Fang and Y. Yu. 2020. User Behavior Retrieval for Click-Through Rate Prediction. In SIGIR."},{"key":"e_1_3_2_2_62_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i5.16557"},{"key":"e_1_3_2_2_63_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2016.0151"},{"key":"e_1_3_2_2_64_1","first-page":"5485","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume":"21","author":"Rael Colin","year":"2020","unstructured":"Colin Rael, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou,Wei Li, and Peter J Liu. 2020. Exploring the limits of transfer learning with a unified text-to-text transformer. The Journal of Machine Learning Research 21, 1 (2020), 5485--5551.","journal-title":"The Journal of Machine Learning Research"},{"key":"e_1_3_2_2_65_1","unstructured":"Shashank Rajput Nikhil Mehta Anima Singh Raghunandan H Keshavan Trung Vu Lukasz Heldt Lichan Hong Yi Tay Vinh Q Tran Jonah Samost et al. 2023. Recommender Systems with Generative Retrieval. arXiv preprint arXiv:2305.05065 (2023)."},{"key":"e_1_3_2_2_66_1","volume-title":"International Conference on Learning Representations.","author":"Ramasesh Vinay Venkatesh","year":"2021","unstructured":"Vinay Venkatesh Ramasesh, Aitor Lewkowycz, and Ethan Dyer. 2021. Eect of scale on catastrophic forgetting in neural networks. In International Conference on Learning Representations."},{"key":"e_1_3_2_2_67_1","doi-asserted-by":"crossref","unstructured":"Kan Ren Jiarui Qin Yuchen Fang Weinan Zhang Lei Zheng Weijie Bian Guorui Zhou Jian Xu Yong Yu Xiaoqiang Zhu et al. 2019. Lifelong Sequential Modeling with Personalized Memorization for User Response Prediction. SIGIR.","DOI":"10.1145\/3331184.3331230"},{"key":"e_1_3_2_2_68_1","doi-asserted-by":"crossref","unstructured":"Steen Rendle. 2010. Factorization machines. In ICDM.","DOI":"10.1109\/ICDM.2010.127"},{"key":"e_1_3_2_2_69_1","volume-title":"A tutorial on principal component analysis. arXiv preprint arXiv:1404.1100","author":"Shlens Jonathon","year":"2014","unstructured":"Jonathon Shlens. 2014. A tutorial on principal component analysis. arXiv preprint arXiv:1404.1100 (2014)."},{"key":"e_1_3_2_2_70_1","doi-asserted-by":"publisher","DOI":"10.1145\/3357384.3357925"},{"key":"e_1_3_2_2_71_1","volume-title":"Is ChatGPT Good at Search? Investigating Large Language Models as Re-Ranking Agent. arXiv preprint arXiv:2304.09542","author":"Sun Weiwei","year":"2023","unstructured":"Weiwei Sun, Lingyong Yan, Xinyu Ma, Pengjie Ren, Dawei Yin, and Zhaochun Ren. 2023. Is ChatGPT Good at Search? Investigating Large Language Models as Re-Ranking Agent. arXiv preprint arXiv:2304.09542 (2023)."},{"key":"e_1_3_2_2_72_1","doi-asserted-by":"publisher","DOI":"10.1145\/3159652.3159656"},{"key":"e_1_3_2_2_73_1","volume-title":"Llama: Open and ecient foundation language models. arXiv preprint arXiv:2302.13971","author":"Touvron Hugo","year":"2023","unstructured":"Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timoth\u00e9e Lacroix, Baptiste Rozi\u00e8re, Naman Goyal, Eric Hambro, Faisal Azhar, et al. 2023. Llama: Open and ecient foundation language models. arXiv preprint arXiv:2302.13971 (2023)."},{"key":"e_1_3_2_2_74_1","volume-title":"TransRec: Learning Transferable Recommendation from Mixture-of-Modality Feedback. arXiv preprint arXiv:2206.06190","author":"Yuan Fajie","year":"2022","unstructured":"JieWang, Fajie Yuan, Mingyue Cheng, Joemon M Jose, Chenyun Yu, Beibei Kong, Xiangnan He, Zhijin Wang, Bo Hu, and Zang Li. 2022. TransRec: Learning Transferable Recommendation from Mixture-of-Modality Feedback. arXiv preprint arXiv:2206.06190 (2022)."},{"key":"e_1_3_2_2_75_1","volume-title":"Zero-Shot Next-Item Recommendation using Large Pretrained Language Models. arXiv preprint arXiv:2304.03153","author":"Wang Lei","year":"2023","unstructured":"Lei Wang and Ee-Peng Lim. 2023. Zero-Shot Next-Item Recommendation using Large Pretrained Language Models. arXiv preprint arXiv:2304.03153 (2023)."},{"key":"e_1_3_2_2_76_1","volume-title":"Proceedings of the ADKDD'17","author":"Fu Bin","year":"2017","unstructured":"RuoxiWang, Bin Fu, Gang Fu, and MingliangWang. 2017. Deep & cross network for ad click predictions. In Proceedings of the ADKDD'17. 1--7."},{"key":"e_1_3_2_2_77_1","volume-title":"Proceedings of the Web Conference","author":"Shivanna Rakesh","year":"2021","unstructured":"RuoxiWang, Rakesh Shivanna, Derek Cheng, Sagar Jain, Dong Lin, Lichan Hong, and Ed Chi. 2021. Dcn v2: Improved deep & cross network and practical lessons for web-scale learning to rank systems. In Proceedings of the Web Conference 2021. 1785--1797."},{"key":"e_1_3_2_2_78_1","volume-title":"Recmind: Large language model powered agent for recommendation. arXiv preprint arXiv:2308.14296","author":"Wang Yancheng","year":"2023","unstructured":"Yancheng Wang, Ziyan Jiang, Zheng Chen, Fan Yang, Yingxue Zhou, Eunah Cho, Xing Fan, Xiaojiang Huang, Yanbin Lu, and Yingzhen Yang. 2023. Recmind: Large language model powered agent for recommendation. arXiv preprint arXiv:2308.14296 (2023)."},{"key":"e_1_3_2_2_79_1","doi-asserted-by":"publisher","DOI":"10.1145\/3539597.3570399"},{"key":"e_1_3_2_2_80_1","volume-title":"Towards Open-World Recommendation with Knowledge Augmentation from Large Language Models. arXiv preprint arXiv:2306.10933","author":"Xi Yunjia","year":"2023","unstructured":"Yunjia Xi, Weiwen Liu, Jianghao Lin, Jieming Zhu, Bo Chen, Ruiming Tang, Weinan Zhang, Rui Zhang, and Yong Yu. 2023. Towards Open-World Recommendation with Knowledge Augmentation from Large Language Models. arXiv preprint arXiv:2306.10933 (2023)."},{"key":"e_1_3_2_2_81_1","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2017\/435"},{"key":"e_1_3_2_2_82_1","first-page":"3926","article-title":"CFM: Convolutional Factorization Machines for Context-Aware Recommendation","volume":"19","author":"Xin Xin","year":"2019","unstructured":"Xin Xin, Bo Chen, Xiangnan He, Dong Wang, Yue Ding, and Joemon M Jose. 2019. CFM: Convolutional Factorization Machines for Context-Aware Recommendation.. In IJCAI, Vol. 19. 3926--3932.","journal-title":"IJCAI"},{"key":"e_1_3_2_2_83_1","volume-title":"Tinynewsrec: Eective and ecient plm-based news recommendation. arXiv preprint arXiv:2112.00944","author":"Yu Yang","year":"2021","unstructured":"Yang Yu, Fangzhao Wu, Chuhan Wu, Jingwei Yi, and Qi Liu. 2021. Tinynewsrec: Eective and ecient plm-based news recommendation. arXiv preprint arXiv:2112.00944 (2021)."},{"key":"e_1_3_2_2_84_1","volume-title":"Where to go next for recommender systems? id-vs. modality-based recommender models revisited. arXiv preprint arXiv:2303.13835","author":"Yuan Zheng","year":"2023","unstructured":"Zheng Yuan, Fajie Yuan, Yu Song, Youhua Li, Junchen Fu, Fei Yang, Yunzhu Pan, and Yongxin Ni. 2023. Where to go next for recommender systems? id-vs. modality-based recommender models revisited. arXiv preprint arXiv:2303.13835 (2023)."},{"key":"e_1_3_2_2_85_1","volume-title":"mixup: Beyond empirical risk minimization. arXiv preprint arXiv:1710.09412","author":"Zhang Hongyi","year":"2017","unstructured":"Hongyi Zhang, Moustapha Cisse, Yann N Dauphin, and David Lopez-Paz. 2017. mixup: Beyond empirical risk minimization. arXiv preprint arXiv:1710.09412 (2017)."},{"key":"e_1_3_2_2_86_1","volume-title":"Is chatgpt fair for recommendation? evaluating fairness in large language model recommendation. arXiv preprint arXiv:2305.07609","author":"Zhang Jizhi","year":"2023","unstructured":"Jizhi Zhang, Keqin Bao, Yang Zhang, Wenjie Wang, Fuli Feng, and Xiangnan He. 2023. Is chatgpt fair for recommendation? evaluating fairness in large language model recommendation. arXiv preprint arXiv:2305.07609 (2023)."},{"key":"e_1_3_2_2_87_1","volume-title":"Recommendation as instruction following: A large language model empowered recommendation approach. arXiv preprint arXiv:2305.07001","author":"Zhang Junjie","year":"2023","unstructured":"Junjie Zhang, Ruobing Xie, Yupeng Hou,Wayne Xin Zhao, Leyu Lin, and Ji-Rong Wen. 2023. Recommendation as instruction following: A large language model empowered recommendation approach. arXiv preprint arXiv:2305.07001 (2023)."},{"key":"e_1_3_2_2_88_1","volume-title":"Memoryaugmented llm personalization with short-and long-term memory coordination. arXiv preprint arXiv:2309.11696","author":"Zhang Kai","year":"2023","unstructured":"Kai Zhang, Fubang Zhao, Yangyang Kang, and Xiaozhong Liu. 2023. Memoryaugmented llm personalization with short-and long-term memory coordination. arXiv preprint arXiv:2309.11696 (2023)."},{"key":"e_1_3_2_2_89_1","volume-title":"Deep learning for click-through rate estimation. IJCAI","author":"Zhang Weinan","year":"2021","unstructured":"Weinan Zhang, Jiarui Qin,Wei Guo, Ruiming Tang, and Xiuqiang He. 2021. Deep learning for click-through rate estimation. IJCAI (2021)."},{"key":"e_1_3_2_2_90_1","volume-title":"TwHIN-BERT: a socially-enriched pretrained language model for multilingual Tweet representations. arXiv preprint arXiv:2209.07562","author":"Zhang Xinyang","year":"2022","unstructured":"Xinyang Zhang, Yury Malkov, Omar Florez, Serim Park, Brian McWilliams, Jiawei Han, and Ahmed El-Kishky. 2022. TwHIN-BERT: a socially-enriched pretrained language model for multilingual Tweet representations. arXiv preprint arXiv:2209.07562 (2022)."},{"key":"e_1_3_2_2_91_1","unstructured":"Yuhui Zhang Hao Ding Zeren Shui Yifei Ma James Zou Anoop Deoras and Hao Wang. 2021. Language models as recommender systems: Evaluations and limitations. (2021)."},{"key":"e_1_3_2_2_92_1","volume-title":"Prompt learning for news recommendation. arXiv preprint arXiv:2304.05263","author":"Zhang Zizhuo","year":"2023","unstructured":"Zizhuo Zhang and BangWang. 2023. Prompt learning for news recommendation. arXiv preprint arXiv:2304.05263 (2023)."},{"key":"e_1_3_2_2_93_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33015941"},{"key":"e_1_3_2_2_94_1","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3219823"}],"event":{"name":"WWW '24: The ACM Web Conference 2024","location":"Singapore Singapore","acronym":"WWW '24","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"]},"container-title":["Proceedings of the ACM Web Conference 2024"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3589334.3645467","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3589334.3645467","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T00:23:25Z","timestamp":1755822205000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3589334.3645467"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,13]]},"references-count":94,"alternative-id":["10.1145\/3589334.3645467","10.1145\/3589334"],"URL":"https:\/\/doi.org\/10.1145\/3589334.3645467","relation":{},"subject":[],"published":{"date-parts":[[2024,5,13]]},"assertion":[{"value":"2024-05-13","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}