{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,24]],"date-time":"2026-03-24T11:48:01Z","timestamp":1774352881177,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":38,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,10,21]],"date-time":"2024-10-21T00:00:00Z","timestamp":1729468800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"Innovation Found from the Engineering Research Center of Integration and Application of Digital Learning Technology, Ministry of Education","award":["1321004"],"award-info":[{"award-number":["1321004"]}]},{"DOI":"10.13039\/https:\/\/doi.org\/10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","award":["2023JBZY038"],"award-info":[{"award-number":["2023JBZY038"]}],"id":[{"id":"10.13039\/https:\/\/doi.org\/10.13039\/501100012226","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,10,21]]},"DOI":"10.1145\/3627673.3679945","type":"proceedings-article","created":{"date-parts":[[2024,10,20]],"date-time":"2024-10-20T19:34:11Z","timestamp":1729452851000},"page":"3887-3891","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["Improving Prompt-based News Recommendation with Individual Template and Customized Answer"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0005-1733-5645","authenticated-orcid":false,"given":"Yijiang","family":"Li","sequence":"first","affiliation":[{"name":"MoE Key Lab of Big Data &amp; Artificial Intelligence in Transportation, Beijing Jiaotong University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5733-3621","authenticated-orcid":false,"given":"Jun","family":"Wu","sequence":"additional","affiliation":[{"name":"MoE Key Lab of Big Data &amp; Artificial Intelligence in Transportation, Beijing Jiaotong University &amp; MoE Engineering Research Center of Integration and Application of Digital Learning Technology, The Open University of China, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2024,10,21]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"Mingxiao An Fangzhao Wu Chuhan Wu Kun Zhang Zheng Liu and Xing Xie. 2019. Neural News Recommendation with Long- and Short-term User Representations. In ACL. 336--345."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"crossref","unstructured":"Qiwei Bi Jian Li Lifeng Shang Xin Jiang Qun Liu and Hanfang Yang. 2022. MTRec: Multi-Task Learning over BERT for News Recommendation. In ACL. 2663--2669.","DOI":"10.18653\/v1\/2022.findings-acl.209"},{"key":"e_1_3_2_1_3_1","unstructured":"Tom B. Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell et al. 2020. Language Models are Few-Shot Learners. In NeurIPS."},{"key":"e_1_3_2_1_4_1","volume-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL. 4171--4186.","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL. 4171--4186."},{"key":"e_1_3_2_1_5_1","unstructured":"Tianyu Gao Adam Fisch and Danqi Chen. 2021. Making Pre-trained Language Models Better Few-shot Learners. In ACL-IJCNLP. 3816--3830."},{"key":"e_1_3_2_1_6_1","unstructured":"Pengcheng He Xiaodong Liu Jianfeng Gao and Weizhu Chen. 2021. Deberta: decoding-Enhanced Bert with Disentangled Attention. In ICLR."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"crossref","unstructured":"Linmei Hu Chen Li Chuan Shi Cheng Yang and Chao Shao. 2020. Graph neural news recommendation with long-term and short-term interest modeling. Inf. Process. Manag. (2020) 102142.","DOI":"10.1016\/j.ipm.2019.102142"},{"key":"e_1_3_2_1_8_1","unstructured":"Eric Jang Shixiang Gu and Ben Poole. 2017. Categorical Reparameterization with Gumbel-Softmax. In ICLR."},{"key":"e_1_3_2_1_9_1","volume-title":"RMBERT: News Recommendation via Recurrent Reasoning Memory Network over BERT. In SIGIR. 1773--1777.","author":"Jia Qinglin","year":"2021","unstructured":"Qinglin Jia, Jingjie Li, Qi Zhang, Xiuqiang He, and Jieming Zhu. 2021. RMBERT: News Recommendation via Recurrent Reasoning Memory Network over BERT. In SIGIR. 1773--1777."},{"key":"e_1_3_2_1_10_1","volume-title":"Malthouse","author":"Li Xinyi","year":"2023","unstructured":"Xinyi Li, Yongfeng Zhang, and Edward C. Malthouse. 2023. PBNR: Prompt-based News Recommender System. CoRR abs\/2304.07862 (2023)."},{"key":"e_1_3_2_1_11_1","unstructured":"Xiang Lisa Li and Percy Liang. 2021. Prefix-Tuning: Optimizing Continuous Prompts for Generation. In ACL-IJCNLP. 4582--4597."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"crossref","unstructured":"Pengfei Liu Weizhe Yuan Jinlan Fu Zhengbao Jiang Hiroaki Hayashi and Graham Neubig. 2023. Pre-train Prompt and Predict: A Systematic Survey of Prompting Methods in Natural Language Processing. ACM Comput. Surv. (2023) 195:1--195:35.","DOI":"10.1145\/3560815"},{"key":"e_1_3_2_1_13_1","volume-title":"P-Tuning v2: Prompt Tuning Can Be Comparable to Fine-tuning Universally Across Scales and Tasks. CoRR abs\/2110.07602","author":"Liu Xiao","year":"2021","unstructured":"Xiao Liu, Kaixuan Ji, Yicheng Fu, Zhengxiao Du, Zhilin Yang, and Jie Tang. 2021. P-Tuning v2: Prompt Tuning Can Be Comparable to Fine-tuning Universally Across Scales and Tasks. CoRR abs\/2110.07602 (2021)."},{"key":"e_1_3_2_1_14_1","volume-title":"CoRR abs\/2103.10385","author":"Liu Xiao","year":"2021","unstructured":"Xiao Liu, Yanan Zheng, Zhengxiao Du, Ming Ding, Yujie Qian, Zhilin Yang, and Jie Tang. 2021. GPT Understands, Too. CoRR abs\/2103.10385 (2021)."},{"key":"e_1_3_2_1_15_1","volume-title":"RoBERTa: A Robustly Optimized BERT Pretraining Approach. CoRR abs\/1907.11692","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. RoBERTa: A Robustly Optimized BERT Pretraining Approach. CoRR abs\/1907.11692 (2019)."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"crossref","unstructured":"Xiangkui Lu Jun Wu and Jianbo Yuan. 2023. Optimizing Reciprocal Rank with Bayesian Average for improved Next Item Recommendation. In SIGIR. 2236--2240.","DOI":"10.1145\/3539618.3592033"},{"key":"e_1_3_2_1_17_1","volume-title":"Discrete Listwise Content-aware Recommendation. ACM Trans. Knowl. Discov. Data","author":"Luo Fangyuan","year":"2024","unstructured":"Fangyuan Luo, Jun Wu, and Tao Wang. 2024. Discrete Listwise Content-aware Recommendation. ACM Trans. Knowl. Discov. Data (2024), 7:1--7:20."},{"key":"e_1_3_2_1_18_1","volume-title":"KDD. 1933--","author":"Okura Shumpei","year":"1942","unstructured":"Shumpei Okura, Yukihiro Tagami, Shingo Ono, and Akira Tajima. 2017. Embedding-based News Recommendation for Millions of Users. In KDD. 1933-- 1942."},{"key":"e_1_3_2_1_19_1","volume-title":"Miller","author":"Petroni Fabio","year":"2019","unstructured":"Fabio Petroni, Tim Rockt\u00e4schel, Sebastian Riedel, Patrick S. H. Lewis, Anton Bakhtin, Yuxiang Wu, and Alexander H. Miller. 2019. Language Models as Knowledge Bases?. In EMNLP-IJCNLP. 2463--2473."},{"key":"e_1_3_2_1_20_1","volume":"202","author":"Raffel Colin","unstructured":"Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. 2020. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. J. Mach. Learn. Res. (2020), 140:1--140:67.","journal-title":"Peter J. Liu."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"crossref","unstructured":"Timo Schick and Hinrich Sch\u00fctze. 2021. Exploiting Cloze-Questions for Few-Shot Text Classification and Natural Language Inference. In EACL. 255--269.","DOI":"10.18653\/v1\/2021.eacl-main.20"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"crossref","unstructured":"Taylor Shin Yasaman Razeghi Robert L. Logan IV Eric Wallace and Sameer Singh. 2020. AutoPrompt: Eliciting Knowledge from Language Models with Automatically Generated Prompts. In EMNLP. 4222--4235.","DOI":"10.18653\/v1\/2020.emnlp-main.346"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"crossref","unstructured":"Shuang Tang Fangyuan Luo and Jun Wu. 2022. Smooth-AUC: Smoothing the Path Towards Rank-based CTR Prediction. In SIGIR. 2400--2404.","DOI":"10.1145\/3477495.3531865"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"crossref","unstructured":"Heyuan Wang Fangzhao Wu Zheng Liu and Xing Xie. 2020. Fine-grained Interest Matching for Neural News Recommendation. In ACL. 836--845.","DOI":"10.18653\/v1\/2020.acl-main.77"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3178876.3186175"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"crossref","unstructured":"YuqiWang andWenqian Shang. 2015. Personalized news recommendation based on consumers? click behavior. In FSKD. 634--638.","DOI":"10.1109\/FSKD.2015.7382016"},{"key":"e_1_3_2_1_27_1","volume-title":"Sylvain Gugger, Mariama Drame, Quentin Lhoest, and Alexander M. Rush.","author":"Debut Lysandre","year":"2020","unstructured":"ThomasWolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, R\u00e9mi Louf, Morgan Funtowicz, Joe Davison, Sam Shleifer, Patrick von Platen, Clara Ma, Yacine Jernite, Julien Plu, Canwen Xu, Teven Le Scao, Sylvain Gugger, Mariama Drame, Quentin Lhoest, and Alexander M. Rush. 2020. Transformers: State-of-the-Art Natural Language Processing. In EMNLP. 38--45."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"crossref","unstructured":"Chuhan Wu Fangzhao Wu Mingxiao An Jianqiang Huang Yongfeng Huang and Xing Xie. 2019. Neural News Recommendation with Attentive Multi-View Learning. In IJCAI. 3863--3869.","DOI":"10.24963\/ijcai.2019\/536"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/3292500.3330665"},{"key":"e_1_3_2_1_30_1","unstructured":"Chuhan Wu Fangzhao Wu Suyu Ge Tao Qi Yongfeng Huang and Xing Xie. 2019. Neural News Recommendation with Multi-Head Self-Attention. In EMNLPIJCNLP. 6388--6393."},{"key":"e_1_3_2_1_31_1","unstructured":"Chuhan Wu Fangzhao Wu Yongfeng Huang and Xing Xie. 2023. Personalized News Recommendation: Methods and Challenges. ACM Trans. Inf. Syst. (2023) 24:1--24:50."},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"crossref","unstructured":"Chuhan Wu Fangzhao Wu Tao Qi and Yongfeng Huang. 2021. Empowering News Recommendation with Pre-trained Language Models. In SIGIR. 1652--1656.","DOI":"10.1145\/3404835.3463069"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"crossref","unstructured":"Chuhan Wu Fangzhao Wu Yang Yu Tao Qi Yongfeng Huang and Qi Liu. 2021. NewsBERT: Distilling Pre-trained Language Model for Intelligent News Application. In EMNLP. 3285--3295.","DOI":"10.18653\/v1\/2021.findings-emnlp.280"},{"key":"e_1_3_2_1_34_1","volume-title":"MIND: A Large-scale Dataset for News Recommendation. In ACL. 3597--3606.","author":"Wu Fangzhao","year":"2020","unstructured":"Fangzhao Wu, Ying Qiao, Jiun-Hung Chen, Chuhan Wu, Tao Qi, Jianxun Lian, Danyang Liu, Xing Xie, Jianfeng Gao, Winnie Wu, and Ming Zhou. 2020. MIND: A Large-scale Dataset for News Recommendation. In ACL. 3597--3606."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"crossref","unstructured":"Shitao Xiao Zheng Liu Yingxia Shao Tao Di Bhuvan Middha Fangzhao Wu and Xing Xie. 2022. Training Large-Scale News Recommenders with Pretrained Language Models in the Loop. In KDD. 4215--4225.","DOI":"10.1145\/3534678.3539120"},{"key":"e_1_3_2_1_36_1","volume-title":"UNBERT: User-News Matching BERT for News Recommendation. In IJCAI. 3356--3362.","author":"Zhang Qi","year":"2021","unstructured":"Qi Zhang, Jingjie Li, Qinglin Jia, Chuyuan Wang, Jieming Zhu, Zhaowei Wang, and Xiuqiang He. 2021. UNBERT: User-News Matching BERT for News Recommendation. In IJCAI. 3356--3362."},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"crossref","unstructured":"Zizhuo Zhang and BangWang. 2023. Prompt Learning for News Recommendation. In SIGIR. 227--237.","DOI":"10.1145\/3539618.3591752"},{"key":"e_1_3_2_1_38_1","volume-title":"DAN: Deep Attention Neural Network for News Recommendation. In AAAI. 5973--5980.","author":"Zhu Qiannan","year":"2019","unstructured":"Qiannan Zhu, Xiaofei Zhou, Zeliang Song, Jianlong Tan, and Li Guo. 2019. DAN: Deep Attention Neural Network for News Recommendation. In AAAI. 5973--5980."}],"event":{"name":"CIKM '24: The 33rd ACM International Conference on Information and Knowledge Management","location":"Boise ID USA","acronym":"CIKM '24","sponsor":["SIGIR ACM Special Interest Group on Information Retrieval"]},"container-title":["Proceedings of the 33rd ACM International Conference on Information and Knowledge Management"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3627673.3679945","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3627673.3679945","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T00:58:09Z","timestamp":1750294689000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3627673.3679945"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,21]]},"references-count":38,"alternative-id":["10.1145\/3627673.3679945","10.1145\/3627673"],"URL":"https:\/\/doi.org\/10.1145\/3627673.3679945","relation":{},"subject":[],"published":{"date-parts":[[2024,10,21]]},"assertion":[{"value":"2024-10-21","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}