{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,10]],"date-time":"2026-04-10T10:04:37Z","timestamp":1775815477179,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":59,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,9,22]]},"DOI":"10.1145\/3705328.3748068","type":"proceedings-article","created":{"date-parts":[[2025,9,6]],"date-time":"2025-09-06T10:46:13Z","timestamp":1757155573000},"page":"411-421","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["<i>R<\/i>\n            <sup>4<\/sup>\n            ec: A Reasoning, Reflection, and Refinement Framework for Recommendation Systems"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0006-6044-8239","authenticated-orcid":false,"given":"Hao","family":"Gu","sequence":"first","affiliation":[{"name":"Institute of Automation, Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-0704-015X","authenticated-orcid":false,"given":"Rui","family":"Zhong","sequence":"additional","affiliation":[{"name":"Kuaishou Technology, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-4128-6968","authenticated-orcid":false,"given":"Yu","family":"Xia","sequence":"additional","affiliation":[{"name":"University of Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-0151-2160","authenticated-orcid":false,"given":"Wei","family":"Yang","sequence":"additional","affiliation":[{"name":"Kuaishou Technology, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-9196-6241","authenticated-orcid":false,"given":"Chi","family":"Lu","sequence":"additional","affiliation":[{"name":"Kuaishou Technology, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9266-0780","authenticated-orcid":false,"given":"Peng","family":"Jiang","sequence":"additional","affiliation":[{"name":"Kuaishou Technology, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3636-3618","authenticated-orcid":false,"given":"Kun","family":"Gai","sequence":"additional","affiliation":[{"name":"Kuaishou Technology, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2025,9,7]]},"reference":[{"key":"e_1_3_3_1_2_2","unstructured":"Josh Achiam Steven Adler Sandhini Agarwal Lama Ahmad Ilge Akkaya Florencia\u00a0Leoni Aleman Diogo Almeida Janko Altenschmidt Sam Altman Shyamal Anadkat et\u00a0al. 2023. Gpt-4 technical report. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2303.08774 (2023)."},{"key":"e_1_3_3_1_3_2","unstructured":"Jinze Bai Shuai Bai Yunfei Chu Zeyu Cui Kai Dang Xiaodong Deng Yang Fan Wenbin Ge Yu Han Fei Huang et\u00a0al. 2023. Qwen technical report. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2309.16609 (2023)."},{"key":"e_1_3_3_1_4_2","unstructured":"Zhuoxi Bai Ning Wu Fengyu Cai Xinyi Zhu and Yun Xiong. 2024. Finetuning Large Language Model for Personalized Ranking. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2405.16127 (2024)."},{"key":"e_1_3_3_1_5_2","doi-asserted-by":"publisher","DOI":"10.1145\/3604915.3608857"},{"key":"e_1_3_3_1_6_2","unstructured":"Iz Beltagy Matthew\u00a0E Peters and Arman Cohan. 2020. Longformer: The long-document transformer. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2004.05150 (2020)."},{"key":"e_1_3_3_1_7_2","doi-asserted-by":"crossref","unstructured":"Jianlv Chen Shitao Xiao Peitian Zhang Kun Luo Defu Lian and Zheng Liu. 2024. Bge m3-embedding: Multi-lingual multi-functionality multi-granularity text embeddings through self-knowledge distillation. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2402.03216 (2024).","DOI":"10.18653\/v1\/2024.findings-acl.137"},{"key":"e_1_3_3_1_8_2","unstructured":"Xinyun Chen Maxwell Lin Nathanael Sch\u00e4rli and Denny Zhou. 2023. Teaching large language models to self-debug. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2304.05128 (2023)."},{"key":"e_1_3_3_1_9_2","doi-asserted-by":"publisher","DOI":"10.1145\/3604915.3610646"},{"key":"e_1_3_3_1_10_2","unstructured":"Jacob Devlin. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/1810.04805 (2018)."},{"key":"e_1_3_3_1_11_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i8.28678"},{"key":"e_1_3_3_1_12_2","unstructured":"Yunfan Gao Tao Sheng Youlin Xiang Yun Xiong Haofen Wang and Jiawei Zhang. 2023. Chat-rec: Towards interactive and explainable llms-augmented recommender system. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2303.14524 (2023)."},{"key":"e_1_3_3_1_13_2","doi-asserted-by":"publisher","DOI":"10.1145\/3523227.3546767"},{"key":"e_1_3_3_1_14_2","unstructured":"Zhibin Gou Zhihong Shao Yeyun Gong Yelong Shen Yujiu Yang Nan Duan and Weizhu Chen. 2023. Critic: Large language models can self-correct with tool-interactive critiquing. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2305.11738 (2023)."},{"key":"e_1_3_3_1_15_2","unstructured":"Hao Gu Jiangyan Yi Chenglong Wang Jianhua Tao Zheng Lian Jiayi He Yong Ren Yujie Chen and Zhengqi Wen. 2025. \\(\\mathcal {A}LLM4ADD\\): Unlocking the Capabilities of Audio Large Language Models for Audio Deepfake Detection. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2505.11079 (2025)."},{"key":"e_1_3_3_1_16_2","unstructured":"Huifeng Guo Ruiming Tang Yunming Ye Zhenguo Li and Xiuqiang He. 2017. DeepFM: a factorization-machine based neural network for CTR prediction. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/1703.04247 (2017)."},{"key":"e_1_3_3_1_17_2","doi-asserted-by":"publisher","DOI":"10.1145\/3397271.3401063"},{"key":"e_1_3_3_1_18_2","volume-title":"The Encyclopaedia Logic, with the Zus tze: Part I of the Encyclopaedia of Philosophical Sciences with the Zus\u00e4tze","author":"Hegel Georg Wilhelm\u00a0Friedrich","year":"1991","unstructured":"Georg Wilhelm\u00a0Friedrich Hegel. 1991. The Encyclopaedia Logic, with the Zus tze: Part I of the Encyclopaedia of Philosophical Sciences with the Zus\u00e4tze. Vol.\u00a01. Hackett Publishing."},{"key":"e_1_3_3_1_19_2","unstructured":"B Hidasi. 2015. Session-based Recommendations with Recurrent Neural Networks. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/1511.06939 (2015)."},{"key":"e_1_3_3_1_20_2","unstructured":"Edward\u00a0J Hu Yelong Shen Phillip Wallis Zeyuan Allen-Zhu Yuanzhi Li Shean Wang Lu Wang and Weizhu Chen. 2021. Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2106.09685 (2021)."},{"key":"e_1_3_3_1_21_2","unstructured":"Yixin Ji Juntao Li Hai Ye Kaixin Wu Jia Xu Linjian Mo and Min Zhang. 2025. Test-time Computing: from System-1 Thinking to System-2 Thinking. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2501.02497 (2025)."},{"key":"e_1_3_3_1_22_2","unstructured":"Nan Jiang Xiaopeng Li Shiqi Wang Qiang Zhou Soneya\u00a0Binta Hossain Baishakhi Ray Varun Kumar Xiaofei Ma and Anoop Deoras. 2024. Training LLMs to Better Self-Debug and Explain Code. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2405.18649 (2024)."},{"key":"e_1_3_3_1_23_2","unstructured":"Daniel Kahneman. 2011. Thinking fast and slow. Farrar Straus and Giroux (2011)."},{"key":"e_1_3_3_1_24_2","unstructured":"Pei Ke Bosi Wen Zhuoer Feng Xiao Liu Xuanyu Lei Jiale Cheng Shengyuan Wang Aohan Zeng Yuxiao Dong Hongning Wang et\u00a0al. 2023. Critiquellm: Scaling llm-as-critic for effective and explainable evaluation of large language model generation. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2311.18702 (2023)."},{"key":"e_1_3_3_1_25_2","unstructured":"Geunwoo Kim Pierre Baldi and Stephen McAleer. 2024. Language models can solve computer tasks. Advances in Neural Information Processing Systems 36 (2024)."},{"key":"e_1_3_3_1_26_2","doi-asserted-by":"crossref","unstructured":"Yehuda Koren Robert Bell and Chris Volinsky. 2009. Matrix factorization techniques for recommender systems. Computer 42 8 (2009) 30\u201337.","DOI":"10.1109\/MC.2009.263"},{"key":"e_1_3_3_1_27_2","unstructured":"Aviral Kumar Vincent Zhuang Rishabh Agarwal Yi Su John\u00a0D Co-Reyes Avi Singh Kate Baumli Shariq Iqbal Colton Bishop Rebecca Roelofs et\u00a0al. 2024. Training language models to self-correct via reinforcement learning. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2409.12917 (2024)."},{"key":"e_1_3_3_1_28_2","unstructured":"Junlong Li Shichao Sun Weizhe Yuan Run-Ze Fan Hai Zhao and Pengfei Liu. 2023. Generative judge for evaluating alignment. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2310.05470 (2023)."},{"key":"e_1_3_3_1_29_2","doi-asserted-by":"publisher","DOI":"10.1145\/3357384.3357951"},{"key":"e_1_3_3_1_30_2","unstructured":"Jiayi Liao Sihang Li Zhengyi Yang Jiancan Wu Yancheng Yuan Xiang Wang and Xiangnan He. 2023. Llara: Aligning large language models with sequential recommenders. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2312.02445 (2023)."},{"key":"e_1_3_3_1_31_2","doi-asserted-by":"publisher","DOI":"10.1145\/3637528.3671884"},{"key":"e_1_3_3_1_32_2","unstructured":"Junling Liu Chao Liu Peilin Zhou Renjie Lv Kang Zhou and Yan Zhang. 2023. Is chatgpt a good recommender? a preliminary study. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2304.10149 (2023)."},{"key":"e_1_3_3_1_33_2","unstructured":"Aman Madaan Niket Tandon Prakhar Gupta Skyler Hallinan Luyu Gao Sarah Wiegreffe Uri Alon Nouha Dziri Shrimai Prabhumoye Yiming Yang et\u00a0al. 2024. Self-refine: Iterative refinement with self-feedback. Advances in Neural Information Processing Systems 36 (2024)."},{"key":"e_1_3_3_1_34_2","volume-title":"The Twelfth International Conference on Learning Representations","author":"Olausson Theo\u00a0X","year":"2023","unstructured":"Theo\u00a0X Olausson, Jeevana\u00a0Priya Inala, Chenglong Wang, Jianfeng Gao, and Armando Solar-Lezama. 2023. Is Self-Repair a Silver Bullet for Code Generation?. In The Twelfth International Conference on Learning Representations."},{"key":"e_1_3_3_1_35_2","unstructured":"Long Ouyang Jeffrey Wu Xu Jiang Diogo Almeida Carroll Wainwright Pamela Mishkin Chong Zhang Sandhini Agarwal Katarina Slama Alex Ray et\u00a0al. 2022. Training language models to follow instructions with human feedback. Advances in neural information processing systems 35 (2022) 27730\u201327744."},{"key":"e_1_3_3_1_36_2","unstructured":"Liangming Pan Alon Albalak Xinyi Wang and William\u00a0Yang Wang. 2023. Logic-lm: Empowering large language models with symbolic solvers for faithful logical reasoning. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2305.12295 (2023)."},{"key":"e_1_3_3_1_37_2","unstructured":"Debjit Paul Mete Ismayilzada Maxime Peyrard Beatriz Borges Antoine Bosselut Robert West and Boi Faltings. 2023. Refiner: Reasoning feedback on intermediate representations. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2304.01904 (2023)."},{"key":"e_1_3_3_1_38_2","unstructured":"Yuxiao Qu Tianjun Zhang Naman Garg and Aviral Kumar. 2024. Recursive introspection: Teaching language model agents how to self-improve. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2407.18219 (2024)."},{"key":"e_1_3_3_1_39_2","unstructured":"Noah Shinn Federico Cassano Ashwin Gopinath Karthik Narasimhan and Shunyu Yao. 2024. Reflexion: Language agents with verbal reinforcement learning. Advances in Neural Information Processing Systems 36 (2024)."},{"key":"e_1_3_3_1_40_2","unstructured":"Charlie Snell Jaehoon Lee Kelvin Xu and Aviral Kumar. 2024. Scaling llm test-time compute optimally can be more effective than scaling model parameters. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2408.03314 (2024)."},{"key":"e_1_3_3_1_41_2","doi-asserted-by":"publisher","DOI":"10.1145\/3357384.3357925"},{"key":"e_1_3_3_1_42_2","unstructured":"Rohan Taori Ishaan Gulrajani Tianyi Zhang Yann Dubois Xuechen Li Carlos Guestrin Percy Liang and Tatsunori\u00a0B Hashimoto. 2023. Stanford alpaca: an instruction-following llama model (2023). URL https:\/\/github. com\/tatsu-lab\/stanford_alpaca 1 9 (2023)."},{"key":"e_1_3_3_1_43_2","unstructured":"Hugo Touvron Thibaut Lavril Gautier Izacard Xavier Martinet Marie-Anne Lachaux Timoth\u00e9e Lacroix Baptiste Rozi\u00e8re Naman Goyal Eric Hambro Faisal Azhar et\u00a0al. 2023. Llama: Open and efficient foundation language models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2302.13971 (2023)."},{"key":"e_1_3_3_1_44_2","unstructured":"Hanbing Wang Xiaorui Liu Wenqi Fan Xiangyu Zhao Venkataramana Kini Devendra Yadav Fei Wang Zhen Wen Jiliang Tang and Hui Liu. 2024. Rethinking large language model architectures for sequential recommendations. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2402.09543 (2024)."},{"key":"e_1_3_3_1_45_2","unstructured":"Lei Wang and Ee-Peng Lim. 2023. Zero-shot next-item recommendation using large pretrained language models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2304.03153 (2023)."},{"key":"e_1_3_3_1_46_2","doi-asserted-by":"publisher","DOI":"10.1145\/3124749.3124754"},{"key":"e_1_3_3_1_47_2","unstructured":"Wenjie Wang Honghui Bao Xinyu Lin Jizhi Zhang Yongqi Li Fuli Feng See-Kiong Ng and Tat-Seng Chua. 2024. Learnable Tokenizer for LLM-based Generative Recommendation. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2405.07314 (2024)."},{"key":"e_1_3_3_1_48_2","unstructured":"Xuezhi Wang Jason Wei Dale Schuurmans Quoc Le Ed Chi Sharan Narang Aakanksha Chowdhery and Denny Zhou. 2022. Self-consistency improves chain of thought reasoning in language models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2203.11171 (2022)."},{"key":"e_1_3_3_1_49_2","unstructured":"Sean Welleck Ximing Lu Peter West Faeze Brahman Tianxiao Shen Daniel Khashabi and Yejin Choi. 2022. Generating sequences by learning to self-correct. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2211.00053 (2022)."},{"key":"e_1_3_3_1_50_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.emnlp-main.714"},{"key":"e_1_3_3_1_51_2","doi-asserted-by":"publisher","DOI":"10.1145\/3640457.3688104"},{"key":"e_1_3_3_1_52_2","unstructured":"Zhiheng Xi Dingwen Yang Jixuan Huang Jiafu Tang Guanyu Li Yiwen Ding Wei He Boyang Hong Shihan Do Wenyu Zhan et\u00a0al. 2024. Enhancing LLM Reasoning via Critique Models with Test-Time and Training-Time Supervision. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2411.16579 (2024)."},{"key":"e_1_3_3_1_53_2","doi-asserted-by":"crossref","unstructured":"Yu Xia Rui Zhong Hao Gu Wei Yang Chi Lu Peng Jiang and Kun Gai. 2025. Hierarchical Tree Search-based User Lifelong Behavior Modeling on Large Language Model. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2505.19505 (2025).","DOI":"10.1145\/3726302.3729995"},{"key":"e_1_3_3_1_54_2","unstructured":"Lanling Xu Junjie Zhang Bingqian Li Jinpeng Wang Mingchen Cai Wayne\u00a0Xin Zhao and Ji-Rong Wen. 2024. Prompting large language models for recommender systems: A comprehensive framework and empirical analysis. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2401.04997 (2024)."},{"key":"e_1_3_3_1_55_2","unstructured":"An Yang Baosong Yang Beichen Zhang Binyuan Hui Bo Zheng Bowen Yu Chengyuan Li Dayiheng Liu Fei Huang Haoran Wei et\u00a0al. 2024. Qwen2. 5 Technical Report. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2412.15115 (2024)."},{"key":"e_1_3_3_1_56_2","unstructured":"Wenhao Yu Zhihan Zhang Zhenwen Liang Meng Jiang and Ashish Sabharwal. 2023. Improving language models via plug-and-play retrieval feedback. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2305.14002 (2023)."},{"key":"e_1_3_3_1_57_2","unstructured":"Junjie Zhang Ruobing Xie Yupeng Hou Xin Zhao Leyu Lin and Ji-Rong Wen. 2023. Recommendation as instruction following: A large language model empowered recommendation approach. ACM Transactions on Information Systems (2023)."},{"key":"e_1_3_3_1_58_2","unstructured":"Zihuai Zhao Wenqi Fan Jiatong Li Yunqing Liu Xiaowei Mei Yiqi Wang Zhen Wen Fei Wang Xiangyu Zhao Jiliang Tang et\u00a0al. 2023. Recommender systems in the era of large language models (llms). arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2307.02046 (2023)."},{"key":"e_1_3_3_1_59_2","unstructured":"Xin Zheng Jie Lou Boxi Cao Xueru Wen Yuqiu Ji Hongyu Lin Yaojie Lu Xianpei Han Debing Zhang and Le Sun. 2024. Critic-cot: Boosting the reasoning abilities of large language model via chain-of-thoughts critic. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2408.16326 (2024)."},{"key":"e_1_3_3_1_60_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33015941"}],"event":{"name":"RecSys '25: Nineteenth ACM Conference on Recommender Systems","location":"Prague Czech Republic","acronym":"RecSys '25","sponsor":["SIGCHI ACM Special Interest Group on Computer-Human Interaction","SIGAI ACM Special Interest Group on Artificial Intelligence","SIGIR ACM Special Interest Group on Information Retrieval","SIGKDD ACM Special Interest Group on Knowledge Discovery in Data","SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"]},"container-title":["Proceedings of the Nineteenth ACM Conference on Recommender Systems"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3705328.3748068","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,6]],"date-time":"2025-09-06T11:44:21Z","timestamp":1757159061000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3705328.3748068"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,7]]},"references-count":59,"alternative-id":["10.1145\/3705328.3748068","10.1145\/3705328"],"URL":"https:\/\/doi.org\/10.1145\/3705328.3748068","relation":{},"subject":[],"published":{"date-parts":[[2025,9,7]]},"assertion":[{"value":"2025-09-07","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}