{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,10]],"date-time":"2026-04-10T10:04:57Z","timestamp":1775815497664,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":61,"publisher":"ACM","funder":[{"name":"QSDL","award":["RUC24QSDL015"],"award-info":[{"award-number":["RUC24QSDL015"]}]},{"name":"NSFC","award":["62122089"],"award-info":[{"award-number":["62122089"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,7,13]]},"DOI":"10.1145\/3726302.3730092","type":"proceedings-article","created":{"date-parts":[[2025,7,14]],"date-time":"2025-07-14T01:18:36Z","timestamp":1752455916000},"page":"504-514","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":10,"title":["<i>The Truth Becomes Clearer Through Debate!<\/i>\n            Multi-Agent Systems with Large Language Models Unmask Fake News"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2912-750X","authenticated-orcid":false,"given":"Yuhan","family":"Liu","sequence":"first","affiliation":[{"name":"Gaoling School of Artificial Intelligence, Renmin University of China, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-2154-7215","authenticated-orcid":false,"given":"Yuxuan","family":"Liu","sequence":"additional","affiliation":[{"name":"Gaoling School of Artificial Intelligence, Renmin University of China, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4532-2156","authenticated-orcid":false,"given":"Xiaoqing","family":"Zhang","sequence":"additional","affiliation":[{"name":"Gaoling School of Artificial Intelligence, Renmin University of China, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6633-0796","authenticated-orcid":false,"given":"Xiuying","family":"Chen","sequence":"additional","affiliation":[{"name":"Mohamed bin Zayed University of Artificial Intelligence, Jeddah, United Arab Emirates"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3356-6823","authenticated-orcid":false,"given":"Rui","family":"Yan","sequence":"additional","affiliation":[{"name":"Gaoling School of Artificial Intelligence, Renmin University of China, Beijing, China and Wuhan University, Wuhan, China"}]}],"member":"320","published-online":{"date-parts":[[2025,7,13]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jocs.2010.12.007"},{"key":"e_1_3_2_1_2_1","volume-title":"Debate and critical analysis: The harmony of conflict","author":"Branham Robert James","unstructured":"Robert James Branham. 2013. Debate and critical analysis: The harmony of conflict. Routledge."},{"key":"e_1_3_2_1_3_1","volume-title":"The Twelfth International Conference on Learning Representations.","author":"Chan Chi-Min","year":"2024","unstructured":"Chi-Min Chan, Weize Chen, Yusheng Su, Jianxuan Yu, Wei Xue, Shanghang Zhang, Jie Fu, and Zhiyuan Liu. 2024. ChatEval: Towards Better LLM-based Evaluators through Multi-Agent Debate. In The Twelfth International Conference on Learning Representations."},{"key":"e_1_3_2_1_4_1","volume-title":"Reconcile: Round-table conference improves reasoning via consensus among diverse llms. arXiv preprint arXiv:2309","author":"Chih-Yao Chen Justin","year":"2023","unstructured":"Justin Chih-Yao Chen, Swarnadeep Saha, and Mohit Bansal. 2023c. Reconcile: Round-table conference improves reasoning via consensus among diverse llms. arXiv preprint arXiv:2309.13007 (2023)."},{"key":"e_1_3_2_1_5_1","volume-title":"A Topic-aware Summarization Framework with Different Modal Side Information. SIGIR","author":"Chen Xiuying","year":"2023","unstructured":"Xiuying Chen, Mingzhe Li, Shen Gao, Xin Cheng, Qiang Yang, Qishen Zhang, Xin Gao, and Xiangliang Zhang. 2023a. A Topic-aware Summarization Framework with Different Modal Side Information. SIGIR (2023)."},{"key":"e_1_3_2_1_6_1","volume-title":"Improving the Robustness of Summarization Systems with Dual Augmentation. ACL","author":"Chen Xiuying","year":"2023","unstructured":"Xiuying Chen, Guodong Long, Chongyang Tao, Mingzhe Li, Xin Gao, Chengqi Zhang, and Xiangliang Zhang. 2023b. Improving the Robustness of Summarization Systems with Dual Augmentation. ACL (2023)."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N19-1423"},{"key":"e_1_3_2_1_8_1","first-page":"2008","article-title":"The Ultimate Lincoln-Douglas Debate Handbook","volume":"16","author":"Djuranovic Marko","year":"2003","unstructured":"Marko Djuranovic. 2003. The Ultimate Lincoln-Douglas Debate Handbook. Retrieved Sep, Vol. 16 (2003), 2008.","journal-title":"Retrieved Sep"},{"key":"e_1_3_2_1_9_1","volume-title":"Improving factuality and reasoning in language models through multiagent debate. arXiv preprint arXiv:2305.14325","author":"Du Yilun","year":"2023","unstructured":"Yilun Du, Shuang Li, Antonio Torralba, Joshua B Tenenbaum, and Igor Mordatch. 2023. Improving factuality and reasoning in language models through multiagent debate. arXiv preprint arXiv:2305.14325 (2023)."},{"key":"e_1_3_2_1_10_1","unstructured":"Abhimanyu Dubey Abhinav Jauhri Abhinav Pandey Abhishek Kadian Ahmad Al-Dahle Aiesha Letman Akhil Mathur Alan Schelten Amy Yang Angela Fan et al. 2024. The llama 3 herd of models. arXiv preprint arXiv:2407.21783 (2024)."},{"key":"e_1_3_2_1_11_1","volume-title":"Science","volume":"363","author":"Grinberg Nir","year":"2019","unstructured":"Nir Grinberg, Kenneth Joseph, Lisa Friedland, Briony Swire-Thompson, and David Lazer. 2019. Fake news on Twitter during the 2016 US presidential election. Science, Vol. 363, 6425 (2019), 374--378."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442442.3452328"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/2487788.2488033"},{"key":"e_1_3_2_1_14_1","volume-title":"Zijuan Lin, Liyang Zhou, et al.","author":"Hong Sirui","year":"2023","unstructured":"Sirui Hong, Xiawu Zheng, Jonathan Chen, Yuheng Cheng, Jinlin Wang, Ceyao Zhang, Zili Wang, Steven Ka Shing Yau, Zijuan Lin, Liyang Zhou, et al. 2023. Metagpt: Meta programming for multi-agent collaborative framework. arXiv preprint arXiv:2308.00352 (2023)."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i20.30214"},{"key":"e_1_3_2_1_16_1","volume-title":"Learn over past, evolve for future: Forecasting temporal trends for fake news detection. arXiv preprint arXiv:2306.14728","author":"Hu Beizhe","year":"2023","unstructured":"Beizhe Hu, Qiang Sheng, Juan Cao, Yongchun Zhu, Danding Wang, Zhengjia Wang, and Zhiwei Jin. 2023. Learn over past, evolve for future: Forecasting temporal trends for fake news detection. arXiv preprint arXiv:2306.14728 (2023)."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v36i5.20517"},{"key":"e_1_3_2_1_18_1","volume-title":"Guangyu Robert Yang, and Andrew Ahn","author":"Kaiya Zhao","year":"2023","unstructured":"Zhao Kaiya, Michelangelo Naim, Jovana Kondic, Manuel Cortes, Jiaxin Ge, Shuying Luo, Guangyu Robert Yang, and Andrew Ahn. 2023. Lyfe Agents: Generative agents for low-cost real-time social interactions. arXiv preprint arXiv:2310.02172 (2023)."},{"key":"e_1_3_2_1_19_1","volume-title":"FakeBERT: Fake news detection in social media with a BERT-based deep learning approach. Multimedia tools and applications","author":"Kaliyar Rohit Kumar","year":"2021","unstructured":"Rohit Kumar Kaliyar, Anurag Goswami, and Pratik Narang. 2021. FakeBERT: Fake news detection in social media with a BERT-based deep learning approach. Multimedia tools and applications, Vol. 80, 8 (2021), 11765--11788."},{"key":"e_1_3_2_1_20_1","volume-title":"Proceedings 13","author":"Kazil Jackie","year":"2020","unstructured":"Jackie Kazil, David Masad, and Andrew Crooks. 2020. Utilizing python for agent-based modeling: The mesa framework. In Social, Cultural, and Behavioral Modeling: 13th International Conference, SBP-BRiMS 2020, Washington, DC, USA, October 18--21, 2020, Proceedings 13. Springer, 308--317."},{"key":"e_1_3_2_1_21_1","volume-title":"Forty-first International Conference on Machine Learning.","author":"Khan Akbir","year":"2024","unstructured":"Akbir Khan, John Hughes, Dan Valentine, Laura Ruis, Kshitij Sachan, Ansh Radhakrishnan, Edward Grefenstette, Samuel R Bowman, Tim Rockt\u00e4schel, and Ethan Perez. 2024. Debating with More Persuasive LLMs Leads to More Truthful Answers. In Forty-first International Conference on Machine Learning."},{"key":"e_1_3_2_1_22_1","volume-title":"Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980","author":"Kingma Diederik P","year":"2014","unstructured":"Diederik P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)."},{"key":"e_1_3_2_1_23_1","volume-title":"Quantifying the impact of large language models on collective opinion dynamics. arXiv preprint arXiv:2308.03313","author":"Li Chao","year":"2023","unstructured":"Chao Li, Xing Su, Chao Fan, Haoying Han, Cong Xue, and Chunmo Zheng. 2023b. Quantifying the impact of large language models on collective opinion dynamics. arXiv preprint arXiv:2308.03313 (2023)."},{"key":"e_1_3_2_1_24_1","first-page":"51991","article-title":"Camel: Communicative agents for'' mind'' exploration of large language model society","volume":"36","author":"Li Guohao","year":"2023","unstructured":"Guohao Li, Hasan Hammoud, Hani Itani, Dmitrii Khizbullin, and Bernard Ghanem. 2023a. Camel: Communicative agents for'' mind'' exploration of large language model society. Advances in Neural Information Processing Systems, Vol. 36 (2023), 51991--52008.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_25_1","volume-title":"Large Language Model Agent for Fake News Detection. arXiv preprint arXiv:2405.01593","author":"Li Xinyi","year":"2024","unstructured":"Xinyi Li, Yongfeng Zhang, and Edward C Malthouse. 2024. Large Language Model Agent for Fake News Detection. arXiv preprint arXiv:2405.01593 (2024)."},{"key":"e_1_3_2_1_26_1","volume-title":"Encouraging divergent thinking in large language models through multi-agent debate. arXiv preprint arXiv:2305.19118","author":"Liang Tian","year":"2023","unstructured":"Tian Liang, Zhiwei He, Wenxiang Jiao, Xing Wang, Yan Wang, Rui Wang, Yujiu Yang, Shuming Shi, and Zhaopeng Tu. 2023. Encouraging divergent thinking in large language models through multi-agent debate. arXiv preprint arXiv:2305.19118 (2023)."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2024\/873"},{"key":"e_1_3_2_1_28_1","volume-title":"From a tiny slip to a giant leap: An llm-based simulation for fake news evolution. arXiv preprint arXiv:2410.19064","author":"Liu Yuhan","year":"2024","unstructured":"Yuhan Liu, Zirui Song, Xiaoqing Zhang, Xiuying Chen, and Rui Yan. 2024b. From a tiny slip to a giant leap: An llm-based simulation for fake news evolution. arXiv preprint arXiv:2410.19064 (2024)."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v39i1.32034"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1145\/3690624.3709171"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/3485447.3512258"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/3459637.3482139"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/3586183.3606763"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1145\/3526113.3545616"},{"key":"e_1_3_2_1_35_1","volume-title":"Communicative agents for software development. arXiv preprint arXiv:2307.07924","author":"Qian Chen","year":"2023","unstructured":"Chen Qian, Xin Cong, Cheng Yang, Weize Chen, Yusheng Su, Juyuan Xu, Zhiyuan Liu, and Maosong Sun. 2023. Communicative agents for software development. arXiv preprint arXiv:2307.07924, Vol. 6 (2023)."},{"key":"e_1_3_2_1_36_1","first-page":"3834","article-title":"Neural User Response Generator: Fake News Detection with Collective User Intelligence","volume":"18","author":"Qian Feng","year":"2018","unstructured":"Feng Qian, Chengyue Gong, Karishma Sharma, and Yan Liu. 2018. Neural User Response Generator: Fake News Detection with Collective User Intelligence.. In IJCAI, Vol. 18. 3834--3840.","journal-title":"IJCAI"},{"key":"e_1_3_2_1_37_1","unstructured":"Yoel Roth. 2022. The vast majority of content we take action on for misinformation is identified proactively."},{"key":"e_1_3_2_1_38_1","volume-title":"Fakenewsnet: A data repository with news content, social context, and spatiotemporal information for studying fake news on social media. Big data","author":"Shu Kai","year":"2020","unstructured":"Kai Shu, Deepak Mahudeswaran, Suhang Wang, Dongwon Lee, and Huan Liu. 2020. Fakenewsnet: A data repository with news content, social context, and spatiotemporal information for studying fake news on social media. Big data, Vol. 8, 3 (2020), 171--188."},{"key":"e_1_3_2_1_39_1","volume-title":"Mmac-copilot: Multi-modal agent collaboration operating system copilot. arXiv e-prints","author":"Song Zirui","year":"2024","unstructured":"Zirui Song, Yaohang Li, Meng Fang, Zhenhao Chen, Zecheng Shi, and Yuan Huang. 2024. Mmac-copilot: Multi-modal agent collaboration operating system copilot. arXiv e-prints (2024), arXiv--2404."},{"key":"e_1_3_2_1_40_1","volume-title":"IConference 2014 proceedings","author":"Starbird Kate","year":"2014","unstructured":"Kate Starbird, Jim Maddock, Mania Orand, Peg Achterman, and Robert M Mason. 2014. Rumors, false flags, and digital vigilantes: Misinformation on twitter after the 2013 boston marathon bombing. IConference 2014 proceedings (2014)."},{"key":"e_1_3_2_1_41_1","volume-title":"Multi-agent collaboration: Harnessing the power of intelligent llm agents. arXiv preprint arXiv:2306.03314","author":"Talebirad Yashar","year":"2023","unstructured":"Yashar Talebirad and Amirhossein Nadiri. 2023. Multi-agent collaboration: Harnessing the power of intelligent llm agents. arXiv preprint arXiv:2306.03314 (2023)."},{"key":"e_1_3_2_1_42_1","volume-title":"Simulating social media using large language models to evaluate alternative news feed algorithms. arXiv preprint arXiv:2310.05984","author":"T\u00f6rnberg Petter","year":"2023","unstructured":"Petter T\u00f6rnberg, Diliara Valeeva, Justus Uitermark, and Christopher Bail. 2023. Simulating social media using large language models to evaluate alternative news feed algorithms. arXiv preprint arXiv:2310.05984 (2023)."},{"key":"e_1_3_2_1_43_1","volume-title":"Graph Attention Networks. In International Conference on Learning Representations.","author":"Veli\u010dkovi\u0107 Petar","year":"2018","unstructured":"Petar Veli\u010dkovi\u0107, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Li\u00f2, and Yoshua Bengio. 2018. Graph Attention Networks. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3219903"},{"key":"e_1_3_2_1_45_1","volume-title":"Foundations and recent trends in multimodal mobile agents: A survey. arXiv preprint arXiv:2411.02006","author":"Wu Biao","year":"2024","unstructured":"Biao Wu, Yanda Li, Meng Fang, Zirui Song, Zhiwei Zhang, Yunchao Wei, and Ling Chen. 2024. Foundations and recent trends in multimodal mobile agents: A survey. arXiv preprint arXiv:2411.02006 (2024)."},{"key":"e_1_3_2_1_46_1","volume-title":"Autogen: Enabling next-gen llm applications via multi-agent conversation framework. arXiv preprint arXiv:2308.08155","author":"Wu Qingyun","year":"2023","unstructured":"Qingyun Wu, Gagan Bansal, Jieyu Zhang, Yiran Wu, Shaokun Zhang, Erkang Zhu, Beibin Li, Li Jiang, Xiaoyun Zhang, and Chi Wang. 2023. Autogen: Enabling next-gen llm applications via multi-agent conversation framework. arXiv preprint arXiv:2308.08155 (2023)."},{"key":"e_1_3_2_1_47_1","volume-title":"Small models are valuable plug-ins for large language models. arXiv preprint arXiv:2305.08848","author":"Xu Canwen","year":"2023","unstructured":"Canwen Xu, Yichong Xu, Shuohang Wang, Yang Liu, Chenguang Zhu, and Julian McAuley. 2023. Small models are valuable plug-ins for large language models. arXiv preprint arXiv:2305.08848 (2023)."},{"key":"e_1_3_2_1_48_1","unstructured":"An Yang Baosong Yang Beichen Zhang Binyuan Hui Bo Zheng Bowen Yu Chengyuan Li Dayiheng Liu Fei Huang Haoran Wei et al. 2024. Qwen2. 5 Technical Report. arXiv preprint arXiv:2412.15115 (2024)."},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"crossref","unstructured":"Feng Yu Qiang Liu Shu Wu Liang Wang Tieniu Tan et al. 2017. A Convolutional Approach for Misinformation Identification.. In IJCAI. 3901--3907.","DOI":"10.24963\/ijcai.2017\/545"},{"key":"e_1_3_2_1_50_1","volume-title":"Proceedings of the 22nd Chinese National Conference on Computational Linguistics (Volume 2: Frontier Forum). 16--29","author":"Yuhan Liu","year":"2023","unstructured":"Liu Yuhan, Chen Xiuying, and Yan Rui. 2023. Unleashing the power of large models: Exploring human-machine conversations. In Proceedings of the 22nd Chinese National Conference on Computational Linguistics (Volume 2: Frontier Forum). 16--29."},{"key":"e_1_3_2_1_51_1","volume-title":"Weaving Context Across Images: Improving Vision-Language Models through Focus-Centric Visual Chains. arXiv preprint arXiv:2504.20199","author":"Zhang Juntian","year":"2025","unstructured":"Juntian Zhang, Chuanqi Cheng, Yuhan Liu, Wei Liu, Jian Luan, and Rui Yan. 2025. Weaving Context Across Images: Improving Vision-Language Models through Focus-Centric Visual Chains. arXiv preprint arXiv:2504.20199 (2025)."},{"key":"e_1_3_2_1_52_1","volume-title":"Exploring Collaboration Mechanisms for LLM Agents: A Social Psychology View. In ICLR 2024 Workshop on Large Language Model (LLM) Agents.","author":"Zhang Jintian","year":"2024","unstructured":"Jintian Zhang, Xin Xu, Ningyu Zhang, Ruibo Liu, Bryan Hooi, and Shumin Deng. 2024 e. Exploring Collaboration Mechanisms for LLM Agents: A Social Psychology View. In ICLR 2024 Workshop on Large Language Model (LLM) Agents."},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442381.3450004"},{"key":"e_1_3_2_1_54_1","volume-title":"A large-scale time-aware agents simulation for influencer selection in digital advertising campaigns. arXiv preprint arXiv:2411.01143","author":"Zhang Xiaoqing","year":"2024","unstructured":"Xiaoqing Zhang, Xiuying Chen, Yuhan Liu, Jianzhou Wang, Zhenxing Hu, and Rui Yan. 2024a. A large-scale time-aware agents simulation for influencer selection in digital advertising campaigns. arXiv preprint arXiv:2411.01143 (2024)."},{"key":"e_1_3_2_1_55_1","volume-title":"SAGraph: A Large-scale Text-Rich Social Graph Dataset for Advertising Campaigns. arXiv preprint arXiv:2403.15105","author":"Zhang Xiaoqing","year":"2024","unstructured":"Xiaoqing Zhang, Xiuying Chen, Yuhan Liu, Jianzhou Wang, Zhenxing Hu, and Rui Yan. 2024b. SAGraph: A Large-scale Text-Rich Social Graph Dataset for Advertising Campaigns. arXiv preprint arXiv:2403.15105 (2024)."},{"key":"e_1_3_2_1_56_1","volume-title":"2024 d. Thinking Before Running! Efficient Code Generation with Thorough Exploration and Optimal Refinement. arXiv preprint arXiv:2502.17442","author":"Zhang Xiaoqing","year":"2024","unstructured":"Xiaoqing Zhang, Yuhan Liu, Flood Sung, Xiuying Chen, and Rui Yan. 2024 d. Thinking Before Running! Efficient Code Generation with Thorough Exploration and Optimal Refinement. arXiv preprint arXiv:2502.17442 (2024)."},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1145\/3589334.3645369"},{"key":"e_1_3_2_1_58_1","volume-title":"Proceedings of the 37th International Conference on Neural Information Processing Systems. 41432--41450","author":"Zhang Zheng","year":"2023","unstructured":"Zheng Zhang, Qi Liu, Hao Jiang, Fei Wang, Yan Zhuang, Le Wu, Weibo Gao, and Enhong Chen. 2023. FairLISA: fair user modeling with limited sensitive attributes information. In Proceedings of the 37th International Conference on Neural Information Processing Systems. 41432--41450."},{"key":"e_1_3_2_1_59_1","volume-title":"The Thirty-eighth Annual Conference on Neural Information Processing Systems.","author":"Zhang Zheng","unstructured":"Zheng Zhang, Wei Song, Qi Liu, Qingyang Mao, Yiyan Wang, Weibo Gao, Zhenya Huang, Shijin Wang, and Enhong Chen. [n.,d.]. Towards Accurate and Fair Cognitive Diagnosis via Monotonic Data Augmentation. In The Thirty-eighth Annual Conference on Neural Information Processing Systems."},{"key":"e_1_3_2_1_60_1","volume-title":"MFAN: Multi-modal Feature-enhanced Attention Networks for Rumor Detection. IJCAI.","author":"Zheng Jiaqi","year":"2022","unstructured":"Jiaqi Zheng, Xi Zhang, Sanchuan Guo, Quan Wang, Wenyu Zang, and Yongdong Zhang. 2022. MFAN: Multi-modal Feature-enhanced Attention Networks for Rumor Detection. IJCAI."},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1145\/3477495.3531816"}],"event":{"name":"SIGIR '25: The 48th International ACM SIGIR Conference on Research and Development in Information Retrieval","location":"Padua Italy","acronym":"SIGIR '25","sponsor":["SIGIR ACM Special Interest Group on Information Retrieval"]},"container-title":["Proceedings of the 48th International ACM SIGIR Conference on Research and Development in Information Retrieval"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3726302.3730092","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T18:30:52Z","timestamp":1755887452000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3726302.3730092"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,13]]},"references-count":61,"alternative-id":["10.1145\/3726302.3730092","10.1145\/3726302"],"URL":"https:\/\/doi.org\/10.1145\/3726302.3730092","relation":{},"subject":[],"published":{"date-parts":[[2025,7,13]]},"assertion":[{"value":"2025-07-13","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}