{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,10]],"date-time":"2026-04-10T10:05:35Z","timestamp":1775815535808,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":76,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,10,21]],"date-time":"2024-10-21T00:00:00Z","timestamp":1729468800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/https:\/\/doi.org\/10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62203425"],"award-info":[{"award-number":["62203425"]}],"id":[{"id":"10.13039\/https:\/\/doi.org\/10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Postdoctoral Fellowship Program of CPSF","award":["GZC20232738"],"award-info":[{"award-number":["GZC20232738"]}]},{"name":"International Postdoctoral Exchange Fellowship Program by Office of China Postdoc Council","award":["YJ20220198"],"award-info":[{"award-number":["YJ20220198"]}]},{"DOI":"10.13039\/https:\/\/doi.org\/10.13039\/501100002858","name":"China Postdoctoral Science Foundation","doi-asserted-by":"publisher","award":["2022TQ0344"],"award-info":[{"award-number":["2022TQ0344"]}],"id":[{"id":"10.13039\/https:\/\/doi.org\/10.13039\/501100002858","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Open Research Project of the State Key Laboratory of Media Convergence and Communication","award":["SKLMCC2022KF001"],"award-info":[{"award-number":["SKLMCC2022KF001"]}]},{"name":"CIPSC-SMP-Zhipu.AI Large Model Cross-Disciplinary Fund"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,10,21]]},"DOI":"10.1145\/3627673.3679519","type":"proceedings-article","created":{"date-parts":[[2024,10,20]],"date-time":"2024-10-20T19:34:21Z","timestamp":1729452861000},"page":"1732-1742","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":40,"title":["Let Silence Speak: Enhancing Fake News Detection with Generated Comments from Large Language Models"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0008-3558-7629","authenticated-orcid":false,"given":"Qiong","family":"Nan","sequence":"first","affiliation":[{"name":"Institute of Computing Technology, Chinese Academy of Sciences &amp; University of Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2481-5023","authenticated-orcid":false,"given":"Qiang","family":"Sheng","sequence":"additional","affiliation":[{"name":"Institute of Computing Technology, Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7857-1546","authenticated-orcid":false,"given":"Juan","family":"Cao","sequence":"additional","affiliation":[{"name":"Institute of Computing Technology, Chinese Academy of Sciences &amp; University of Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-1678-5772","authenticated-orcid":false,"given":"Beizhe","family":"Hu","sequence":"additional","affiliation":[{"name":"Institute of Computing Technology, Chinese Academy of Sciences &amp; University of Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3603-9743","authenticated-orcid":false,"given":"Danding","family":"Wang","sequence":"additional","affiliation":[{"name":"Institute of Computing Technology, Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4597-8534","authenticated-orcid":false,"given":"Jintao","family":"Li","sequence":"additional","affiliation":[{"name":"Institute of Computing Technology, Chinese Academy of Sciences, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2024,10,21]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2019.8683170"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.17645\/mac.v9i1.3523"},{"key":"e_1_3_2_1_3_1","volume-title":"The Dark Side of Language Models: Exploring the Potential of LLMs in Multimedia Disinformation Generation and Dissemination. Machine Learning with Applications","author":"Barman Dipto","year":"2024","unstructured":"Dipto Barman, Ziyi Guo, and Owen Conlan. 2024. The Dark Side of Language Models: Exploring the Potential of LLMs in Multimedia Disinformation Generation and Dissemination. Machine Learning with Applications (2024), 100545."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/1963405.1963500"},{"key":"e_1_3_2_1_5_1","volume-title":"The Twelfth International Conference on Learning Representations.","author":"Chen Canyu","year":"2023","unstructured":"Canyu Chen and Kai Shu. 2023. Can LLM-Generated Misinformation Be Detected?. In The Twelfth International Conference on Learning Representations."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.37"},{"key":"e_1_3_2_1_7_1","volume-title":"Gemma: Open Models Based on Gemini Research and Technology. arXiv preprint arXiv:2403.08295","author":"DeepMind Google","year":"2024","unstructured":"Google DeepMind. 2024. Gemma: Open Models Based on Gemini Research and Technology. arXiv preprint arXiv:2403.08295 (2024)."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_9_1","volume-title":"False Rumor of Explosion at White House Causes Stocks to Briefly Plunge","author":"Domm Patti","year":"2024","unstructured":"Patti Domm. 2013. False Rumor of Explosion at White House Causes Stocks to Briefly Plunge; AP Confirms Its Twitter Feed Was Hacked. https:\/\/www.cnbc.com\/id\/100646197. Accessed: 2024-05--20."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.3389\/fcomm.2021.661801"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.22329\/il.v38i1.5068"},{"key":"e_1_3_2_1_12_1","volume-title":"Integrating Large Language Models with Graphical Session-Based Recommendation. arXiv preprint arXiv:2402.16539","author":"Guo Naicheng","year":"2024","unstructured":"Naicheng Guo, Hongwei Cheng, Qianqiao Liang, Linxun Chen, and Bing Han. 2024. Integrating Large Language Models with Graphical Session-Based Recommendation. arXiv preprint arXiv:2402.16539 (2024)."},{"key":"e_1_3_2_1_13_1","volume-title":"Good Advisor: Exploring the Role of Large Language Models in Fake News Detection. Proceedings of the AAAI Conference on Artificial Intelligence","volume":"38","author":"Hu Beizhe","year":"2024","unstructured":"Beizhe Hu, Qiang Sheng, Juan Cao, Yuhui Shi, Yang Li, Danding Wang, and Peng Qi. 2024. Bad Actor, Good Advisor: Exploring the Role of Large Language Models in Fake News Detection. Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 38, 20 (2024), 22105--22113."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-industry.13"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.aiopen.2022.09.001"},{"key":"e_1_3_2_1_16_1","volume-title":"Large Language Model Interaction Simulator for Cold-Start Item Recommendation. arXiv preprint arXiv:2402.09176","author":"Huang Feiran","year":"2024","unstructured":"Feiran Huang, Zhenghang Yang, Junyi Jiang, Yuanchen Bei, Yijie Zhang, and Hao Chen. 2024. Large Language Model Interaction Simulator for Cold-Start Item Recommendation. arXiv preprint arXiv:2402.09176 (2024)."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1137\/1.9781611978032.50"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v30i1.10382"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-020-10183-2"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3492855"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM50108.2020.00037"},{"key":"e_1_3_2_1_22_1","volume-title":"Your Large Language Model is Secretly a Fairness Proponent and You Should Prompt it Like One. arXiv preprint arXiv:2402.12150","author":"Li Tianlin","year":"2024","unstructured":"Tianlin Li, Xiaoyu Zhang, Chao Du, Tianyu Pang, Qian Liu, Qing Guo, Chao Shen, and Yang Liu. 2024. Your Large Language Model is Secretly a Fairness Proponent and You Should Prompt it Like One. arXiv preprint arXiv:2402.12150 (2024)."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/BigData59044.2023.10386777"},{"key":"e_1_3_2_1_24_1","volume-title":"TELLER: A Trustworthy Framework for Explainable, Generalizable and Controllable Fake News Detection. arXiv preprint arXiv:2402.07776","author":"Liu Hui","year":"2024","unstructured":"Hui Liu, Wenya Wang, Haoru Li, and Haoliang Li. 2024. TELLER: A Trustworthy Framework for Explainable, Generalizable and Controllable Fake News Detection. arXiv preprint arXiv:2402.07776 (2024)."},{"key":"e_1_3_2_1_25_1","volume-title":"Can Large Language Models Detect Rumors on Social Media? arXiv preprint arXiv:2402.03916","author":"Liu Qiang","year":"2024","unstructured":"Qiang Liu, Xiang Tao, Junfei Wu, Shu Wu, and Liang Wang. 2024. Can Large Language Models Detect Rumors on Social Media? arXiv preprint arXiv:2402.03916 (2024)."},{"key":"e_1_3_2_1_26_1","volume-title":"From Skepticism to Acceptance: Simulating the Attitude Dynamics Toward Fake News. arXiv preprint arXiv:2403.09498","author":"Liu Yuhan","year":"2024","unstructured":"Yuhan Liu, Xiuying Chen, Xiaoqing Zhang, Xing Gao, Ji Zhang, and Rui Yan. 2024. From Skepticism to Acceptance: Simulating the Attitude Dynamics Toward Fake News. arXiv preprint arXiv:2403.09498 (2024)."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3386253"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.883"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/3184558.3188729"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1145\/3459637.3482139"},{"key":"e_1_3_2_1_31_1","volume-title":"Exploiting User Comments for Early Detection of Fake News Prior to Users' Commenting. arXiv preprint arXiv:2310.10429","author":"Nan Qiong","year":"2023","unstructured":"Qiong Nan, Qiang Sheng, Juan Cao, Yongchun Zhu, Danding Wang, Guang Yang, Jintao Li, and Kai Shu. 2023. Exploiting User Comments for Early Detection of Fake News Prior to Users' Commenting. arXiv preprint arXiv:2310.10429 (2023)."},{"key":"e_1_3_2_1_32_1","unstructured":"OpenAI. 2022. ChatGPT: Optimizing Language Models for Dialogue. https:\/\/openai.com\/blog\/chatgpt\/. Accessed: 2024-03-05."},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i01.5386"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i12.26689"},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/3474085.3481548"},{"key":"e_1_3_2_1_36_1","volume-title":"Nghia Khau, Bruna Guedes, Modestas Filipavicius, Andr\u00e9 Ferreira Manso, and Roland Mathis.","author":"Sekuli\u0107 Ivan","year":"2024","unstructured":"Ivan Sekuli\u0107, Silvia Terragni, Victor Guimar aes, Nghia Khau, Bruna Guedes, Modestas Filipavicius, Andr\u00e9 Ferreira Manso, and Roland Mathis. 2024. Reliable LLM-based User Simulator for Task-Oriented Dialogue Systems. arXiv preprint arXiv:2402.13374 (2024)."},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ipm.2022.102959"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-long.311"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3292500.3330935"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1089\/big.2020.0062"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1145\/3137597.3137600"},{"key":"e_1_3_2_1_42_1","volume-title":"Scott Ruston, and Huan Liu.","author":"Shu Kai","year":"2020","unstructured":"Kai Shu, Guoqing Zheng, Yichuan Li, Subhabrata Mukherjee, Ahmed Hassan Awadallah, Scott Ruston, and Huan Liu. 2020. Leveraging multi-source weak social supervision for early detection of fake news. arXiv preprint arXiv:2004.01732 (2020)."},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1109\/BigMM.2019.00--44"},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.4"},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.3961\/jpmph.20.094"},{"key":"e_1_3_2_1_46_1","unstructured":"Hugo Touvron Louis Martin Kevin Stone Peter Albert Amjad Almahairi Yasmine Babaei Nikolay Bashlykov Soumya Batra Prajjwal Bhargava Shruti Bhosale Dan Bikel Lukas Blecher Cristian Canton Ferrer Moya Chen Guillem Cucurull David Esiobu Jude Fernandes Jeremy Fu Wenyin Fu Brian Fuller Cynthia Gao Vedanuj Goswami Naman Goyal Anthony Hartshorn Saghar Hosseini Rui Hou Hakan Inan Marcin Kardas Viktor Kerkez Madian Khabsa Isabel Kloumann Artem Korenev Punit Singh Koura Marie-Anne Lachaux Thibaut Lavril Jenya Lee Diana Liskovich Yinghai Lu Yuning Mao Xavier Martinet Todor Mihaylov Pushkar Mishra Igor Molybog Yixin Nie Andrew Poulton Jeremy Reizenstein Rashi Rungta Kalyan Saladi Alan Schelten Ruan Silva Eric Michael Smith Ranjan Subramanian Xiaoqing Ellen Tan Binh Tang Ross Taylor Adina Williams Jian Xiang Kuan Puxin Xu Zheng Yan Iliyan Zarov Yuchen Zhang Angela Fan Melanie Kambadur Sharan Narang Aurelien Rodriguez Robert Stojnic Sergey Edunov and Thomas Scialom. 2023. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288 (2023)."},{"key":"e_1_3_2_1_47_1","volume-title":"Exploring the Impact of Large Language Models on Recommender Systems: An Extensive Review. arXiv preprint arXiv:2402.18590","author":"Vats Arpita","year":"2024","unstructured":"Arpita Vats, Vinija Jain, Rahul Raja, and Aman Chadha. 2024. Exploring the Impact of Large Language Models on Recommender Systems: An Extensive Review. arXiv preprint arXiv:2402.18590 (2024)."},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1126\/science.aap9559"},{"key":"e_1_3_2_1_49_1","volume-title":"Fact-checking: A meta-analysis of what works and for whom. Political communication","author":"Walter Nathan","year":"2020","unstructured":"Nathan Walter, Jonathan Cohen, R Lance Holbert, and Yasmin Morag. 2020. Fact-checking: A meta-analysis of what works and for whom. Political communication, Vol. 37, 3 (2020), 350--375."},{"key":"e_1_3_2_1_50_1","volume-title":"DELL: Generating Reactions and Explanations for LLM-Based Misinformation Detection. arXiv preprint arXiv:2402.10426","author":"Wan Herun","year":"2024","unstructured":"Herun Wan, Shangbin Feng, Zhaoxuan Tan, Heng Wang, Yulia Tsvetkov, and Minnan Luo. 2024. DELL: Generating Reactions and Explanations for LLM-Based Misinformation Detection. arXiv preprint arXiv:2402.10426 (2024)."},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1145\/3589334.3645471"},{"key":"e_1_3_2_1_52_1","volume-title":"MMIDR: Teaching Large Language Model to Interpret Multimodal Misinformation via Knowledge Distillation. arXiv preprint arXiv:2403.14171","author":"Wang Longzheng","year":"2024","unstructured":"Longzheng Wang, Xiaohan Xu, Lei Zhang, Jiarui Lu, Yongxiu Xu, Hongbo Xu, and Chuang Zhang. 2024. MMIDR: Teaching Large Language Model to Interpret Multimodal Misinformation via Knowledge Distillation. arXiv preprint arXiv:2403.14171 (2024)."},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i16.17647"},{"key":"e_1_3_2_1_54_1","volume-title":"DEEM: Dynamic Experienced Expert Modeling for Stance Detection. arXiv preprint arXiv:2402.15264","author":"Wang Xiaolong","year":"2024","unstructured":"Xiaolong Wang, Yile Wang, Sijie Cheng, Peng Li, and Yang Liu. 2024. DEEM: Dynamic Experienced Expert Modeling for Stance Detection. arXiv preprint arXiv:2402.15264 (2024)."},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3219903"},{"key":"e_1_3_2_1_56_1","volume-title":"Understanding News Creation Intents: Frame, Dataset, and Method. arXiv preprint arXiv:2312.16490","author":"Wang Zhengjia","year":"2023","unstructured":"Zhengjia Wang, Danding Wang, Qiang Sheng, Juan Cao, Silong Su, Yifan Sun, Beizhe Hu, and Siyuan Ma. 2023. Understanding News Creation Intents: Frame, Dataset, and Method. arXiv preprint arXiv:2312.16490 (2023)."},{"key":"e_1_3_2_1_57_1","unstructured":"Jason Wei Yi Tay Rishi Bommasani Colin Raffel Barret Zoph Sebastian Borgeaud Dani Yogatama Maarten Bosma Denny Zhou Donald Metzler et al. 2022. Emergent abilities of large language models. arXiv preprint arXiv:2206.07682 (2022)."},{"key":"e_1_3_2_1_58_1","volume-title":"Weibo Community Convention (Trial). https:\/\/service.account.weibo.com\/roles\/gongyue. Accessed: 2024-05--13","year":"2021","unstructured":"Weibo. 2021. Weibo Community Convention (Trial). https:\/\/service.account.weibo.com\/roles\/gongyue. Accessed: 2024-05--13."},{"key":"e_1_3_2_1_59_1","volume-title":"Risks Report","author":"Forum World Economic","year":"2024","unstructured":"World Economic Forum. 2024. Global Risks Report 2024. https:\/\/www3.weforum.org\/docs\/WEF_The_Global_Risks_Report_2024.pdf. Accessed: 2024-05--13."},{"key":"e_1_3_2_1_60_1","volume-title":"Fake News in Sheep's Clothing: Robust Fake News Detection Against LLM-Empowered Style Attacks. arXiv preprint arXiv:2310.10830","author":"Wu Jiaying","year":"2023","unstructured":"Jiaying Wu and Bryan Hooi. 2023. Fake News in Sheep's Clothing: Robust Fake News Detection Against LLM-Empowered Style Attacks. arXiv preprint arXiv:2310.10830 (2023)."},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.3233\/FAIA200348"},{"key":"e_1_3_2_1_62_1","volume-title":"Can Large Language Model Agents Simulate Human Trust Behaviors? arXiv preprint arXiv:2402.04559","author":"Xie Chengxing","year":"2024","unstructured":"Chengxing Xie, Canyu Chen, Feiran Jia, Ziyu Ye, Kai Shu, Adel Bibi, Ziniu Hu, Philip Torr, Bernard Ghanem, and Guohao Li. 2024. Can Large Language Model Agents Simulate Human Trust Behaviors? arXiv preprint arXiv:2402.04559 (2024)."},{"key":"e_1_3_2_1_63_1","volume-title":"The Earth is Flat because...: Investigating LLMs' Belief towards Misinformation via Persuasive Conversation. arXiv preprint arXiv:2312.09085","author":"Xu Rongwu","year":"2023","unstructured":"Rongwu Xu, Brian S Lin, Shujian Yang, Tianqi Zhang, Weiyan Shi, Tianwei Zhang, Zhixuan Fang, Wei Xu, and Han Qiu. 2023. The Earth is Flat because...: Investigating LLMs' Belief towards Misinformation via Persuasive Conversation. arXiv preprint arXiv:2312.09085 (2023)."},{"key":"e_1_3_2_1_64_1","volume-title":"LEMMA: Towards LVLM-Enhanced Multimodal Misinformation Detection with External Knowledge Augmentation. arXiv preprint arXiv:2402.11943","author":"Xuan Keyang","year":"2024","unstructured":"Keyang Xuan, Li Yi, Fan Yang, Ruochen Wu, Yi R Fung, and Heng Ji. 2024. LEMMA: Towards LVLM-Enhanced Multimodal Misinformation Detection with External Knowledge Augmentation. arXiv preprint arXiv:2402.11943 (2024)."},{"key":"e_1_3_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.1109\/INES49302.2020.9147195"},{"key":"e_1_3_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33015644"},{"key":"e_1_3_2_1_67_1","volume-title":"Proceedings of the 29th International Conference on Computational Linguistics. International Committee on Computational Linguistics, Gyeongju, Republic of Korea, 2608--2621","author":"Yang Zhiwei","year":"2022","unstructured":"Zhiwei Yang, Jing Ma, Hechang Chen, Hongzhan Lin, Ziyang Luo, and Yi Chang. 2022. A Coarse-to-fine Cascaded Evidence-Distillation Neural Network for Explainable Fake News Detection. In Proceedings of the 29th International Conference on Computational Linguistics. International Committee on Computational Linguistics, Gyeongju, Republic of Korea, 2608--2621. https:\/\/aclanthology.org\/2022.coling-1.230"},{"key":"e_1_3_2_1_68_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_69_1","volume-title":"PopALM: Popularity-Aligned Language Models for Social Media Trendy Response Prediction. arXiv preprint arXiv:2402.18950","author":"Yu Erxin","year":"2024","unstructured":"Erxin Yu, Jing Li, and Chunpu Xu. 2024. PopALM: Popularity-Aligned Language Models for Social Media Trendy Response Prediction. arXiv preprint arXiv:2402.18950 (2024)."},{"key":"e_1_3_2_1_70_1","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2017\/545"},{"key":"e_1_3_2_1_71_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442381.3450004"},{"key":"e_1_3_2_1_72_1","doi-asserted-by":"publisher","DOI":"10.1145\/3589334.3645680"},{"key":"e_1_3_2_1_73_1","unstructured":"Zhipu AI. 2024. GLM-4. https:\/\/open.bigmodel.cn\/dev\/howuse\/glm4. Accessed: 2024-03-05."},{"key":"e_1_3_2_1_74_1","doi-asserted-by":"publisher","DOI":"10.1145\/3477495.3531816"},{"key":"e_1_3_2_1_75_1","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2022.3185151"},{"key":"e_1_3_2_1_76_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISI58743.2023.10297146"}],"event":{"name":"CIKM '24: The 33rd ACM International Conference on Information and Knowledge Management","location":"Boise ID USA","acronym":"CIKM '24","sponsor":["SIGIR ACM Special Interest Group on Information Retrieval"]},"container-title":["Proceedings of the 33rd ACM International Conference on Information and Knowledge Management"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3627673.3679519","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3627673.3679519","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T00:03:29Z","timestamp":1750291409000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3627673.3679519"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,21]]},"references-count":76,"alternative-id":["10.1145\/3627673.3679519","10.1145\/3627673"],"URL":"https:\/\/doi.org\/10.1145\/3627673.3679519","relation":{},"subject":[],"published":{"date-parts":[[2024,10,21]]},"assertion":[{"value":"2024-10-21","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}