{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T17:17:13Z","timestamp":1776100633606,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":60,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,6,12]],"date-time":"2023-06-12T00:00:00Z","timestamp":1686528000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/501100000780","name":"European Commission","doi-asserted-by":"publisher","award":["951911"],"award-info":[{"award-number":["951911"]}],"id":[{"id":"10.13039\/501100000780","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,6,12]]},"DOI":"10.1145\/3591106.3592278","type":"proceedings-article","created":{"date-parts":[[2023,6,8]],"date-time":"2023-06-08T22:33:38Z","timestamp":1686263618000},"page":"627-635","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":11,"title":["Framing the News: From Human Perception to Large Language Model Inferences"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-6728-6774","authenticated-orcid":false,"given":"David","family":"Alonso del Barrio","sequence":"first","affiliation":[{"name":"Idiap Research Institute, Switzerland"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5488-2182","authenticated-orcid":false,"given":"Daniel","family":"Gatica-Perez","sequence":"additional","affiliation":[{"name":"Idiap Research Institute, Switzerland and EPFL, Switzerland"}]}],"member":"320","published-online":{"date-parts":[[2023,6,12]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.22146\/jsp.56457"},{"key":"e_1_3_2_1_2_1","volume-title":"RAFT: A real-world few-shot text classification benchmark. arXiv preprint arXiv:2109.14076","author":"Alex Neel","year":"2021","unstructured":"Neel Alex, Eli Lifland, Lewis Tunstall, Abhishek Thakur, Pegah Maham, C\u00a0Jess Riedel, Emmie Hine, Carolyn Ashurst, Paul Sedille, Alexis Carlier, 2021. RAFT: A real-world few-shot text classification benchmark. arXiv preprint arXiv:2109.14076 (2021)."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","unstructured":"David Alonso\u00a0del Barrio and Daniel Gatica-Perez. 2022. How Did Europe\u2019s Press Cover Covid-19 Vaccination News? A Five-Country Analysis. (2022) 35\u201343. https:\/\/doi.org\/10.1145\/3512732.3533588","DOI":"10.1145\/3512732.3533588"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445922"},{"key":"e_1_3_2_1_5_1","volume-title":"Artificial intelligence in journalism: A boon or bane? In Optimization in machine learning and applications","author":"Biswal Santosh\u00a0Kumar","unstructured":"Santosh\u00a0Kumar Biswal and Nikhil\u00a0Kumar Gouda. 2020. Artificial intelligence in journalism: A boon or bane? In Optimization in machine learning and applications. Springer, 155\u2013167."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1080\/1369183X.2014.1002200"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","unstructured":"Michael Bommarito and Daniel\u00a0Martin Katz. 2022. GPT Takes the Bar Exam. https:\/\/doi.org\/10.48550\/ARXIV.2212.14402","DOI":"10.48550\/ARXIV.2212.14402"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1177\/1077699019859901"},{"key":"e_1_3_2_1_9_1","volume-title":"Language models are few-shot learners. Advances in neural information processing systems 33","author":"Brown Tom","year":"2020","unstructured":"Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared\u00a0D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, 2020. Language models are few-shot learners. Advances in neural information processing systems 33 (2020), 1877\u20131901."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1080\/19312458.2014.937527"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1177\/0894439315596385"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.3115\/v1"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.3390\/ijerph17218136"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.4185\/RLCS-2019-1357en"},{"key":"e_1_3_2_1_15_1","volume-title":"Clarifying journalism\u2019s quantitative turn: A typology for evaluating data journalism, computational journalism, and computer-assisted reporting. Digital journalism 3, 3","author":"Coddington Mark","year":"2015","unstructured":"Mark Coddington. 2015. Clarifying journalism\u2019s quantitative turn: A typology for evaluating data journalism, computational journalism, and computer-assisted reporting. Digital journalism 3, 3 (2015), 331\u2013348."},{"key":"e_1_3_2_1_16_1","volume-title":"Doing News Framing Analysis","author":"Cooper D","unstructured":"Stephen\u00a0D Cooper. 2010. The oppositional framing of bloggers. In Doing News Framing Analysis. Routledge, 151\u2013172."},{"key":"e_1_3_2_1_17_1","volume-title":"GPT-3: What\u2019s it good for?Natural Language Engineering 27, 1","author":"Dale Robert","year":"2021","unstructured":"Robert Dale. 2021. GPT-3: What\u2019s it good for?Natural Language Engineering 27, 1 (2021), 113\u2013118."},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1177\/0963662509352044"},{"key":"e_1_3_2_1_19_1","volume-title":"To frame is to explain: A deductive frame-analysis of Dutch and French climate change coverage during the annual UN Conferences of the Parties. Public understanding of science 19, 6","author":"Dirikx Astrid","year":"2010","unstructured":"Astrid Dirikx and Dave Gelders. 2010. To frame is to explain: A deductive frame-analysis of Dutch and French climate change coverage during the annual UN Conferences of the Parties. Public understanding of science 19, 6 (2010), 732\u2013742."},{"key":"e_1_3_2_1_20_1","volume-title":"Gsum: A general framework for guided neural abstractive summarization. arXiv preprint arXiv:2010.08014","author":"Dou Zi-Yi","year":"2020","unstructured":"Zi-Yi Dou, Pengfei Liu, Hiroaki Hayashi, Zhengbao Jiang, and Graham Neubig. 2020. Gsum: A general framework for guided neural abstractive summarization. arXiv preprint arXiv:2010.08014 (2020)."},{"key":"e_1_3_2_1_21_1","first-page":"1","article-title":"The corona chronicles: Framing analysis of online news headlines of the COVID-19 pandemic in Italy, USA and South Africa","volume":"27","author":"Ebrahim Sumayya","year":"2022","unstructured":"Sumayya Ebrahim. 2022. The corona chronicles: Framing analysis of online news headlines of the COVID-19 pandemic in Italy, USA and South Africa. Health SA Gesondheid (Online) 27 (2022), 1\u20138.","journal-title":"Health SA Gesondheid (Online)"},{"key":"e_1_3_2_1_22_1","unstructured":"Hend Abdelgaber\u00a0Ahmed El-Behary. 2021. A Feverish Spring: A Comparative Analysis of COVID-19 News Framing in Sweden the UK and Egypt. (2021)."},{"key":"e_1_3_2_1_23_1","volume-title":"Framing: Towards clarification of a fractured paradigm. McQuail\u2019s reader in mass communication theory 390","author":"Entman M","year":"1993","unstructured":"Robert\u00a0M Entman. 1993. Framing: Towards clarification of a fractured paradigm. McQuail\u2019s reader in mass communication theory 390 (1993), 397."},{"key":"e_1_3_2_1_24_1","volume-title":"Making pre-trained language models better few-shot learners. arXiv preprint arXiv:2012.15723","author":"Gao Tianyu","year":"2020","unstructured":"Tianyu Gao, Adam Fisch, and Danqi Chen. 2020. Making pre-trained language models better few-shot learners. arXiv preprint arXiv:2012.15723 (2020)."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2021.3062875"},{"key":"e_1_3_2_1_26_1","volume-title":"A Lens-Mapping Framework for Understanding the Encoding and Decoding of Interpersonal Dispositions in Nonverbal Behavior. Journal of Personality and Social Psychology 66 (02","author":"Gifford Robert","year":"1994","unstructured":"Robert Gifford. 1994. A Lens-Mapping Framework for Understanding the Encoding and Decoding of Interpersonal Dispositions in Nonverbal Behavior. Journal of Personality and Social Psychology 66 (02 1994), 398\u2013412. https:\/\/doi.org\/10.1037\/\/0022-3514.66.2.398"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.eacl-main.154"},{"key":"e_1_3_2_1_28_1","volume-title":"Sustainable Advanced Computing","author":"Gupta Anushka","unstructured":"Anushka Gupta, Diksha Chugh, Rahul Katarya, 2022. Automated news summarization using transformers. In Sustainable Advanced Computing. Springer, 249\u2013259."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1080\/21670811.2016.1162663"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1111\/psj.12343"},{"key":"e_1_3_2_1_31_1","volume-title":"Quantitative analysis of large amounts of journalistic texts using topic modelling. Digital journalism 4, 1","author":"Jacobi Carina","year":"2016","unstructured":"Carina Jacobi, Wouter Van\u00a0Atteveldt, and Kasper Welbers. 2016. Quantitative analysis of large amounts of journalistic texts using topic modelling. Digital journalism 4, 1 (2016), 89\u2013106."},{"key":"e_1_3_2_1_32_1","volume-title":"How can we know what language models know?Transactions of the Association for Computational Linguistics 8","author":"Jiang Zhengbao","year":"2020","unstructured":"Zhengbao Jiang, Frank\u00a0F Xu, Jun Araki, and Graham Neubig. 2020. How can we know what language models know?Transactions of the Association for Computational Linguistics 8 (2020), 423\u2013438."},{"key":"e_1_3_2_1_33_1","unstructured":"Shima Khanehzar Andrew Turpin and Gosia Miko\u0142ajczak. 2019. Modeling Political Framing Across Policy Issues and Contexts. In ALTA."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1080\/10714421.2018.1479616"},{"key":"e_1_3_2_1_35_1","volume-title":"Holistic evaluation of language models. arXiv preprint arXiv:2211.09110","author":"Liang Percy","year":"2022","unstructured":"Percy Liang, Rishi Bommasani, Tony Lee, Dimitris Tsipras, Dilara Soylu, Michihiro Yasunaga, Yian Zhang, Deepak Narayanan, Yuhuai Wu, Ananya Kumar, 2022. Holistic evaluation of language models. arXiv preprint arXiv:2211.09110 (2022)."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","unstructured":"Pengfei Liu Weizhe Yuan Jinlan Fu Zhengbao Jiang Hiroaki Hayashi and Graham Neubig. 2021. Pre-train prompt and predict: A systematic survey of prompting methods in natural language processing. (2021). https:\/\/doi.org\/10.48550\/ARXIV.2107.13586","DOI":"10.48550\/ARXIV.2107.13586"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/K19-1047"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1111\/j.1460-2466.2008.00384.x"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3543829.3544529"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1109\/MIC.2018.112102235"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1177\/1464884919861598"},{"key":"e_1_3_2_1_42_1","unstructured":"R. Monarch. 2021. Human-in-the-Loop Machine Learning: Active Learning and Annotation for Human-centered AI. Manning. https:\/\/books.google.ch\/books?id=LCh0zQEACAAJ"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1080\/10584609.2020.1812777"},{"key":"e_1_3_2_1_44_1","volume-title":"Framing analysis: An approach to news discourse. Political communication 10, 1","author":"Pan Zhongdang","year":"1993","unstructured":"Zhongdang Pan and Gerald\u00a0M Kosicki. 1993. Framing analysis: An approach to news discourse. Political communication 10, 1 (1993), 55\u201375."},{"key":"e_1_3_2_1_45_1","volume-title":"Zero-shot text classification with generative language models. arXiv preprint arXiv:1912.10165","author":"Puri Raul","year":"2019","unstructured":"Raul Puri and Bryan Catanzaro. 2019. Zero-shot text classification with generative language models. arXiv preprint arXiv:1912.10165 (2019)."},{"key":"e_1_3_2_1_46_1","volume-title":"Learning how to ask: Querying lms with mixtures of soft prompts. arXiv preprint arXiv:2104.06599","author":"Qin Guanghui","year":"2021","unstructured":"Guanghui Qin and Jason Eisner. 2021. Learning how to ask: Querying lms with mixtures of soft prompts. arXiv preprint arXiv:2104.06599 (2021)."},{"key":"e_1_3_2_1_47_1","unstructured":"Rabindra Lamsal. 2021. Sentiment Analysis of English Tweets with BERTsent. https:\/\/huggingface.co\/rabindralamsal\/finetuned-bertweet-sentiment-analysis."},{"key":"e_1_3_2_1_48_1","volume-title":"Language models are unsupervised multitask learners. OpenAI blog 1, 8","author":"Radford Alec","year":"2019","unstructured":"Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever, 2019. Language models are unsupervised multitask learners. OpenAI blog 1, 8 (2019), 9."},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijcce.2022.03.003"},{"key":"e_1_3_2_1_50_1","volume-title":"Framing of the Covid-19 pandemic and its organizational predictors. Cuadernos. info50","author":"Rodelo V","year":"2021","unstructured":"Frida\u00a0V Rodelo. 2021. Framing of the Covid-19 pandemic and its organizational predictors. Cuadernos. info50 (2021), 91\u2013112."},{"key":"e_1_3_2_1_51_1","volume-title":"Bloom: A 176b-parameter open-access multilingual language model. arXiv preprint arXiv:2211.05100","author":"Scao Teven\u00a0Le","year":"2022","unstructured":"Teven\u00a0Le Scao, Angela Fan, Christopher Akiki, Ellie Pavlick, Suzana Ili\u0107, Daniel Hesslow, Roman Castagn\u00e9, Alexandra\u00a0Sasha Luccioni, Fran\u00e7ois Yvon, Matthias Gall\u00e9, 2022. Bloom: A 176b-parameter open-access multilingual language model. arXiv preprint arXiv:2211.05100 (2022)."},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1111\/j.1460-2466.2000.tb02843.x"},{"key":"e_1_3_2_1_53_1","volume-title":"Constrained language models yield few-shot semantic parsers. arXiv preprint arXiv:2104.08768","author":"Shin Richard","year":"2021","unstructured":"Richard Shin, Christopher\u00a0H Lin, Sam Thomson, Charles Chen, Subhro Roy, Emmanouil\u00a0Antonios Platanios, Adam Pauls, Dan Klein, Jason Eisner, and Benjamin Van\u00a0Durme. 2021. Constrained language models yield few-shot semantic parsers. arXiv preprint arXiv:2104.08768 (2021)."},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.11114\/smc.v5i1.2279"},{"key":"e_1_3_2_1_55_1","volume-title":"Energy and policy considerations for deep learning in NLP. arXiv preprint arXiv:1906.02243","author":"Strubell Emma","year":"2019","unstructured":"Emma Strubell, Ananya Ganesh, and Andrew McCallum. 2019. Energy and policy considerations for deep learning in NLP. arXiv preprint arXiv:1906.02243 (2019)."},{"key":"e_1_3_2_1_56_1","volume-title":"Understanding the capabilities, limitations, and societal impact of large language models. arXiv preprint arXiv:2102.02503","author":"Tamkin Alex","year":"2021","unstructured":"Alex Tamkin, Miles Brundage, Jack Clark, and Deep Ganguli. 2021. Understanding the capabilities, limitations, and societal impact of large language models. arXiv preprint arXiv:2102.02503 (2021)."},{"key":"e_1_3_2_1_57_1","volume-title":"A simple method for commonsense reasoning. arXiv preprint arXiv:1806.02847","author":"Trinh H","year":"2018","unstructured":"Trieu\u00a0H Trinh and Quoc\u00a0V Le. 2018. A simple method for commonsense reasoning. arXiv preprint arXiv:1806.02847 (2018)."},{"key":"e_1_3_2_1_58_1","first-page":"200","article-title":"Multimodal few-shot learning with frozen language models","volume":"34","author":"Tsimpoukelli Maria","year":"2021","unstructured":"Maria Tsimpoukelli, Jacob\u00a0L Menick, Serkan Cabi, SM Eslami, Oriol Vinyals, and Felix Hill. 2021. Multimodal few-shot learning with frozen language models. Advances in Neural Information Processing Systems 34 (2021), 200\u2013212.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1145\/1743546.1743585"},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1177\/17427665211023984"}],"event":{"name":"ICMR '23: International Conference on Multimedia Retrieval","location":"Thessaloniki Greece","acronym":"ICMR '23","sponsor":["SIGMM ACM Special Interest Group on Multimedia"]},"container-title":["Proceedings of the 2023 ACM International Conference on Multimedia Retrieval"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3591106.3592278","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3591106.3592278","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:37:30Z","timestamp":1750178250000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3591106.3592278"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,12]]},"references-count":60,"alternative-id":["10.1145\/3591106.3592278","10.1145\/3591106"],"URL":"https:\/\/doi.org\/10.1145\/3591106.3592278","relation":{},"subject":[],"published":{"date-parts":[[2023,6,12]]},"assertion":[{"value":"2023-06-12","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}