{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,9]],"date-time":"2026-04-09T17:06:28Z","timestamp":1775754388769,"version":"3.50.1"},"reference-count":63,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"2","license":[{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Artif. Intell."],"published-print":{"date-parts":[[2026,2]]},"DOI":"10.1109\/tai.2025.3579452","type":"journal-article","created":{"date-parts":[[2025,6,13]],"date-time":"2025-06-13T13:54:12Z","timestamp":1749822852000},"page":"740-754","source":"Crossref","is-referenced-by-count":5,"title":["Irony Detection, Reasoning, and Understanding in Zero-Shot Learning"],"prefix":"10.1109","volume":"7","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6680-6320","authenticated-orcid":false,"given":"Peiling","family":"Yi","sequence":"first","affiliation":[{"name":"University of Essex, Colchester, U.K."}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-2640-6745","authenticated-orcid":false,"given":"Yuhan","family":"Xia","sequence":"additional","affiliation":[{"name":"University of Essex, Colchester, U.K."}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4407-578X","authenticated-orcid":false,"given":"Yunfei","family":"Long","sequence":"additional","affiliation":[{"name":"Queen Mary University of London, Mile End, U.K."}]}],"member":"263","reference":[{"key":"ref1","volume-title":"A Rhetoric of Irony","author":"Booth","year":"1974"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.4159\/harvard.9780674063143"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1145\/3124420"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1037\/0096-3445.123.2.129"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1016\/j.datak.2012.02.005"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1007\/s00521-020-05102-3"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.naacl-long.238"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1162\/coli_a_00337"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.semeval-1.111"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/S18-1005"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1145\/3649506"},{"key":"ref12","article-title":"Language models are few-shot learners","author":"Brown","year":"2020"},{"key":"ref13","article-title":"ChainLM: Empowering large language models with improved chain-of-thought prompting","author":"Cheng","year":"2024"},{"key":"ref14","first-page":"24824","article-title":"Chain-of-thought prompting elicits reasoning in large language models","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"35","author":"Wei","year":"2022"},{"key":"ref15","article-title":"Unlocking the potential of ChatGPT: A comprehensive exploration of its applications, advantages, limitations, and future directions in natural language processing","author":"Hariri","year":"2023"},{"key":"ref16","first-page":"27730","article-title":"Training language models to follow instructions with human feedback","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"35","author":"Ouyang","year":"2022"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1016\/j.iotcps.2023.04.003"},{"key":"ref18","article-title":"A prompt pattern catalog to enhance prompt engineering with ChatGPT","author":"White","year":"2023"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1162\/dint_a_00235"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1145\/2808797.2808910"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D13-1066"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1145\/3136273.3136298"},{"key":"ref23","first-page":"581","article-title":"Identifying sarcasm in Twitter: A closer look","volume-title":"Proc. 49th Annu. Meeting Assoc. Comput. Linguistics: Hum. Lang. Technol.","author":"Gonz\u00e1lez-Ib\u00e1nez","year":"2011"},{"key":"ref24","first-page":"4238","article-title":"Who cares about sarcastic tweets? Investigating the impact of sarcasm on sentiment analysis","volume-title":"Proc. 9th Int. Conf. Lang. Resources and Evaluation (LREC)","author":"Maynard","year":"2014"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-00825-3_11"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/DSAA.2015.7344888"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-19390-8_38"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/P15-2106"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1007\/s12652-019-01419-7"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-56608-5_45"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1007\/s13369-021-06193-3"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.3390\/computers12110231"},{"key":"ref33","article-title":"Leveraging ChatGPT as text annotation tool for sentiment analysis","author":"Belal","year":"2023"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/cascon62161.2024.10837875"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/IISEC59749.2023.10391005"},{"key":"ref36","article-title":"Improving language understanding by generative pre-training","author":"Radford","year":"2018"},{"issue":"8","key":"ref37","first-page":"9","article-title":"Language models are unsupervised multitask learners","volume":"1","author":"Radford","year":"2019","journal-title":"OpenAI Blog"},{"key":"ref38","first-page":"22199","article-title":"Large language models are zero-shot reasoners","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"35","author":"Kojima","year":"2022"},{"key":"ref39","article-title":"Self-consistency improves chain of thought reasoning in language models","author":"Wang","year":"2022"},{"key":"ref40","article-title":"Automatic chain of thought prompting in large language models","author":"Zhang","year":"2022"},{"key":"ref41","first-page":"11809","article-title":"Tree of thoughts: Deliberate problem solving with large language models","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"36","author":"Yao","year":"2024"},{"key":"ref42","article-title":"Large language model guided tree-of-thought","author":"Long","year":"2023"},{"key":"ref43","article-title":"Using tree-of-thought prompting to boost ChatGPT\u2019s reasoning","author":"Hulbert","year":"2023","journal-title":"Zenodo"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-long.225"},{"key":"ref45","article-title":"Large language models as optimizers","author":"Yang","year":"2023"},{"key":"ref46","article-title":"Promptbreeder: Self-referential self-improvement via prompt evolution","author":"Fernando","year":"2023"},{"key":"ref47","article-title":"Large language models are human-level prompt engineers","author":"Zhou","year":"2022"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.acl-short.54"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.147"},{"key":"ref50","article-title":"Finetuned language models are zero-shot learners","author":"Wei","year":"2021"},{"key":"ref51","article-title":"Multitask prompted training enables zero-shot task generalization","author":"Sanh","year":"2021"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/w16-3604"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/P14-2084"},{"key":"ref54","article-title":"Bert: Pre-training of deep bidirectional transformers for language understanding","author":"Devlin","year":"2018"},{"key":"ref55","article-title":"Roberta: A robustly optimized BERT pretraining approach","author":"Liu","year":"2019"},{"key":"ref56","first-page":"16857","article-title":"MPNet: Masked and permuted pre-training for language understanding","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"33","author":"Song","year":"2020"},{"key":"ref57","article-title":"A survey in automatic irony processing: Linguistic, cognitive, and multi-x perspectives","author":"Zeng","year":"2022"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.1017\/CBO9781107298019"},{"key":"ref59","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-tutorials.1"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.3389\/fpsyg.2017.00606"},{"key":"ref61","article-title":"Meta prompting for AGI systems","author":"Zhang","year":"2023"},{"key":"ref62","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.147"},{"key":"ref63","doi-asserted-by":"publisher","DOI":"10.1037\/h0062427"}],"container-title":["IEEE Transactions on Artificial Intelligence"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/9078688\/11370309\/11036129.pdf?arnumber=11036129","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,3]],"date-time":"2026-02-03T05:52:40Z","timestamp":1770097960000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11036129\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2]]},"references-count":63,"journal-issue":{"issue":"2"},"URL":"https:\/\/doi.org\/10.1109\/tai.2025.3579452","relation":{},"ISSN":["2691-4581"],"issn-type":[{"value":"2691-4581","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,2]]}}}