{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T15:45:35Z","timestamp":1776095135535,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":44,"publisher":"ACM","license":[{"start":{"date-parts":[[2025,4,22]],"date-time":"2025-04-22T00:00:00Z","timestamp":1745280000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,4,28]]},"DOI":"10.1145\/3696410.3714934","type":"proceedings-article","created":{"date-parts":[[2025,4,22]],"date-time":"2025-04-22T22:52:18Z","timestamp":1745362338000},"page":"3751-3761","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":8,"title":["Supernotes: Driving Consensus in Crowd-Sourced Fact-Checking"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7907-1335","authenticated-orcid":false,"given":"Soham","family":"De","sequence":"first","affiliation":[{"name":"University of Washington, Seattle, Washington, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4474-7109","authenticated-orcid":false,"given":"Michiel A.","family":"Bakker","sequence":"additional","affiliation":[{"name":"Massachusetts Institute of Technology, Cambridge, Massachusetts, USA and Google DeepMind, Cambridge, Massachusetts, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-7342-420X","authenticated-orcid":false,"given":"Jay","family":"Baxter","sequence":"additional","affiliation":[{"name":"X Community Notes, Seattle, Washington, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9908-7454","authenticated-orcid":false,"given":"Martin","family":"Saveski","sequence":"additional","affiliation":[{"name":"University of Washington, Seattle, Washington, USA"}]}],"member":"320","published-online":{"date-parts":[[2025,4,22]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al.","author":"Achiam Josh","year":"2023","unstructured":"Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, Florencia Leoni Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al. 2023. Gpt-4 technical report. In arXiv:2303.08774."},{"key":"e_1_3_2_1_2_1","volume-title":"Scaling up fact-checking using the wisdom of crowds. Science Advances","author":"Allen Jennifer","year":"2021","unstructured":"Jennifer Allen, Antonio A Arechar, Gordon Pennycook, and David G Rand. 2021. Scaling up fact-checking using the wisdom of crowds. Science Advances (2021)."},{"key":"e_1_3_2_1_3_1","unstructured":"Michiel Bakker Martin Chadwick Hannah Sheahan Michael Tessler Lucy Campbell-Gillingham Jan Balaguer Nat McAleese Amelia Glaese John Aslanides Matt Botvinick et al. 2022. Fine-tuning language models to find agreement among humans with diverse preferences. In Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/3415164"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"crossref","unstructured":"Jason W Burton Ezequiel Lopez-Lopez Shahar Hechtlinger Zoe Rahwan Samuel Aeschbach Michiel A Bakker Joshua A Becker Aleks Berditchevskaia Julian Berger Levin Brinkmann et al. 2024. How large language models can reshape collective intelligence. Nature Human Behaviour (2024).","DOI":"10.1038\/s41562-024-01959-9"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"crossref","unstructured":"Yuwei Chuai Moritz Pilarski Thomas Renault David Restrepo-Amariles Aurore Troussel-Cl\u00e9ment Gabriele Lenzini and Nicolas Pr\u00f6llochs. 2024. Community-based fact-checking reduces the spread of misleading posts on social media. In arXiv:2409.08781.","DOI":"10.31219\/osf.io\/3a4fe"},{"key":"e_1_3_2_1_7_1","volume-title":"Kai-Cheng Yang, and Filippo Menczer.","author":"DeVerna Matthew R.","year":"2024","unstructured":"Matthew R. DeVerna, Harry Yaojun Yan, Kai-Cheng Yang, and Filippo Menczer. 2024. Fact-checking information from large language models can decrease headline discernment. Proceedings of the National Academy of Sciences (2024)."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3610058"},{"key":"e_1_3_2_1_9_1","volume-title":"Community notes increase trust in fact-checking on social media. PNAS Nexus","author":"Drolsbach Chiara Patricia","year":"2024","unstructured":"Chiara Patricia Drolsbach, Kirill Solovev, and Nicolas Pr\u00f6llochs. 2024. Community notes increase trust in fact-checking on social media. PNAS Nexus (2024)."},{"key":"e_1_3_2_1_10_1","volume-title":"Proceedings of the Conference on Human Factors in Computing Systems.","author":"Epstein Ziv","year":"2020","unstructured":"Ziv Epstein, Gordon Pennycook, and David Rand. 2020. Will the crowd game the algorithm? Using layperson judgments to combat misinformation on social media by downranking distrusted sources. In Proceedings of the Conference on Human Factors in Computing Systems."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"crossref","unstructured":"Sara Fish Paul G\u00f6lz David C Parkes Ariel D Procaccia Gili Rusak Itai Shapira and Manuel W\u00fcthrich. 2023. Generative social choice. In arXiv:2309.01291.","DOI":"10.1145\/3670865.3673547"},{"key":"e_1_3_2_1_12_1","volume-title":"Most Republicans don't trust fact-checkers, and most Americans don't trust the media","author":"Flamini Daniela","year":"2019","unstructured":"Daniela Flamini. 2019. Most Republicans don't trust fact-checkers, and most Americans don't trust the media. Poynter Institute. (2019)."},{"key":"e_1_3_2_1_13_1","volume-title":"Crowdsourced fact-checking? What we learned from Truthsquad. Mediashift","author":"Florin Fabrice","year":"2010","unstructured":"Fabrice Florin. 2010. Crowdsourced fact-checking? What we learned from Truthsquad. Mediashift (2010)."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/3491102.3502004"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/3097983.3098131"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3543507.3583388"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"crossref","unstructured":"Emma Hoes Sacha Altay and Juan Bermeo. 2023. Leveraging ChatGPT for efficient fact-checking. In osf.io\/preprints\/psyarxiv\/qnjkf.","DOI":"10.31234\/osf.io\/qnjkf"},{"key":"e_1_3_2_1_18_1","volume-title":"Foundation Models for Decision Making Workshop, NeurIPS.","author":"Jarrett Daniel","year":"2023","unstructured":"Daniel Jarrett, Miruna Pislar, Michiel A Bakker, Michael Henry Tessler, Raphael Koster, Jan Balaguer, Romuald Elie, Christopher Summerfield, and Andrea Tacchetti. 2023. Language agents as digital representatives in collective decision-making. In Foundation Models for Decision Making Workshop, NeurIPS."},{"key":"e_1_3_2_1_19_1","volume-title":"Group decision making using cardinal social welfare functions. Management Science","author":"Keeney Ralph L","year":"1975","unstructured":"Ralph L Keeney and Craig W Kirkwood. 1975. Group decision making using cardinal social welfare functions. Management Science (1975)."},{"key":"e_1_3_2_1_20_1","volume-title":"Leveraging volunteer fact checking to identify misinformation about COVID-19 in social media. HKS Misinformation Review","author":"Kim Hyunuk","year":"2020","unstructured":"Hyunuk Kim and Dylan Walker. 2020. Leveraging volunteer fact checking to identify misinformation about COVID-19 in social media. HKS Misinformation Review (2020)."},{"key":"e_1_3_2_1_21_1","volume-title":"Manling Li, and Heng Ji.","author":"Kim Kyungha","year":"2024","unstructured":"Kyungha Kim, Sangyun Lee, Kung-Hsiang Huang, Hou Pong Chan, Manling Li, and Heng Ji. 2024. Can LLMs Produce Faithful Explanations For Fact-checking? Towards Faithful Explainable Fact-Checking via Multi-Agent Debate. In arXiv:2402.07401."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1007\/s42001-024-00338-8"},{"key":"e_1_3_2_1_23_1","unstructured":"Helene Landemore. 2022. Can AI bring deliberative democracy to the masses. In HAI Weekly Seminar Working Paper."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"crossref","unstructured":"David MJ Lazer Matthew A Baum Yochai Benkler Adam J Berinsky Kelly M Greenhill Filippo Menczer Miriam J Metzger Brendan Nyhan Gordon Pennycook David Rothschild et al. 2018. The science of fake news. Science (2018).","DOI":"10.1126\/science.aao2998"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1109\/BigData50022.2020.9377956"},{"key":"e_1_3_2_1_26_1","volume-title":"Fair division and collective welfare","author":"Moulin Herv\u00e9","unstructured":"Herv\u00e9 Moulin. 2004. Fair division and collective welfare. MIT press."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/1102351.1102430"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3306446.3340818"},{"key":"e_1_3_2_1_29_1","volume-title":"Shifting attention to accuracy can reduce misinformation online. Nature","author":"Pennycook Gordon","year":"2021","unstructured":"Gordon Pennycook, Ziv Epstein, Mohsen Mosleh, Antonio A Arechar, Dean Eckles, and David G Rand. 2021. Shifting attention to accuracy can reduce misinformation online. Nature (2021)."},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.1806781116"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1609\/icwsm.v16i1.19335"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.01240"},{"key":"e_1_3_2_1_33_1","volume-title":"The perils and promises of fact-checking with large language models. Frontiers in Artificial Intelligence","author":"Quelle Dorian","year":"2024","unstructured":"Dorian Quelle and Alexandre Bovet. 2024. The perils and promises of fact-checking with large language models. Frontiers in Artificial Intelligence (2024)."},{"key":"e_1_3_2_1_34_1","volume-title":"Searching for or reviewing evidence improves crowdworkers' misinformation judgments and reduces partisan bias. Collective Intelligence","author":"Resnick Paul","year":"2023","unstructured":"Paul Resnick, Aljohara Alfayez, Jane Im, and Eric Gilbert. 2023. Searching for or reviewing evidence improves crowdworkers' misinformation judgments and reduces partisan bias. Collective Intelligence (2023)."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/2872518.2890098"},{"key":"e_1_3_2_1_36_1","volume-title":"Community notes moderate engagement with and diffusion of false information online. Preprint","author":"Slaughter Isaac","year":"2025","unstructured":"Isaac Slaughter, Axel Peytavin, Johan Ugander, and Martin Saveski. 2025. Community notes moderate engagement with and diffusion of false information online. Preprint (2025)."},{"key":"e_1_3_2_1_37_1","unstructured":"Christopher T Small Ivan Vendrov Esin Durmus Hadjar Homaei Elizabeth Barry Julien Cornebise Ted Suzman Deep Ganguli and Colin Megill. 2023. Opportunities and Risks of LLMs for Scalable Deliberation with Polis. In arXiv:2306.11932."},{"key":"e_1_3_2_1_38_1","volume-title":"Americans' perspectives on online media warning labels. Behavioral Sciences","author":"Straub Jeremy","year":"2022","unstructured":"Jeremy Straub and Matthew Spradling. 2022. Americans' perspectives on online media warning labels. Behavioral Sciences (2022)."},{"key":"e_1_3_2_1_39_1","unstructured":"Christopher Summerfield Lisa Argyle Michiel Bakker Teddy Collins Esin Durmus Tyna Eloundou Iason Gabriel Deep Ganguli Kobi Hackenburg Gillian Hadfield et al. 2024. How will advanced AI systems impact democracy?. In arXiv:2409.06729."},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-emnlp.416"},{"key":"e_1_3_2_1_41_1","volume-title":"Birdwatch: Crowd wisdom and bridging algorithms can inform understanding and reduce the spread of misinformation. In arXiv:2210.15723.","author":"Wojcik Stefan","year":"2022","unstructured":"Stefan Wojcik, Sophie Hilgard, Nick Judd, Delia Mocanu, Stephen Ragain, MB Hunzaker, Keith Coleman, and Jay Baxter. 2022. Birdwatch: Crowd wisdom and bridging algorithms can inform understanding and reduce the spread of misinformation. In arXiv:2210.15723."},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.naacl-long.313"},{"key":"e_1_3_2_1_43_1","volume-title":"JustiLM: Few-shot Justification Generation for Explainable Fact-Checking of Real-world Claims. Transactions of the Association for Computational Linguistics","author":"Zeng Fengzhu","year":"2024","unstructured":"Fengzhu Zeng and Wei Gao. 2024. JustiLM: Few-shot Justification Generation for Explainable Fact-Checking of Real-world Claims. Transactions of the Association for Computational Linguistics (2024)."},{"key":"e_1_3_2_1_44_1","unstructured":"Xinyi Zhou Ashish Sharma Amy X Zhang and Tim Althoff. 2024. Correcting misinformation on social media with a large language model. In arXiv:2403.11169."}],"event":{"name":"WWW '25: The ACM Web Conference 2025","location":"Sydney NSW Australia","acronym":"WWW '25","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"]},"container-title":["Proceedings of the ACM on Web Conference 2025"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3696410.3714934","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3696410.3714934","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:18:54Z","timestamp":1750295934000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3696410.3714934"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4,22]]},"references-count":44,"alternative-id":["10.1145\/3696410.3714934","10.1145\/3696410"],"URL":"https:\/\/doi.org\/10.1145\/3696410.3714934","relation":{},"subject":[],"published":{"date-parts":[[2025,4,22]]},"assertion":[{"value":"2025-04-22","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}