{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,7]],"date-time":"2025-11-07T11:29:44Z","timestamp":1762514984831,"version":"build-2065373602"},"reference-count":59,"publisher":"Springer Science and Business Media LLC","issue":"11","license":[{"start":{"date-parts":[[2025,7,19]],"date-time":"2025-07-19T00:00:00Z","timestamp":1752883200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,7,19]],"date-time":"2025-07-19T00:00:00Z","timestamp":1752883200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","award":["No. 2022YFF0902100","No. 2022YFF0902100","No. 2022YFF0902100","No. 2022YFF0902100"],"award-info":[{"award-number":["No. 2022YFF0902100","No. 2022YFF0902100","No. 2022YFF0902100","No. 2022YFF0902100"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Nature Scientific Foundation of Heilongjiang Province","award":["YQ2021F006","YQ2021F006","YQ2021F006","YQ2021F006"],"award-info":[{"award-number":["YQ2021F006","YQ2021F006","YQ2021F006","YQ2021F006"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int. J. Mach. Learn. &amp; Cyber."],"published-print":{"date-parts":[[2025,11]]},"DOI":"10.1007\/s13042-025-02741-7","type":"journal-article","created":{"date-parts":[[2025,7,19]],"date-time":"2025-07-19T09:49:05Z","timestamp":1752918545000},"page":"9049-9066","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Pragmatic inference and mapping for conversational implicature"],"prefix":"10.1007","volume":"16","author":[{"given":"Jingwen","family":"Hu","sequence":"first","affiliation":[]},{"given":"Yuanxing","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Longxuan","family":"Ma","sequence":"additional","affiliation":[]},{"given":"Wei-Nan","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Ting","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,7,19]]},"reference":[{"key":"2741_CR1","unstructured":"Achiam J, Adler S, Agarwal S, Ahmad L, Akkaya I, Aleman FL, Almeida D, Altenschmidt J, Altman S, Anadkat S, et al (2023) Gpt-4 technical report. arXiv preprint arXiv:2303.08774"},{"issue":"1","key":"2741_CR2","doi-asserted-by":"publisher","first-page":"123","DOI":"10.1016\/j.jml.2011.09.005","volume":"66","author":"L Bott","year":"2012","unstructured":"Bott L, Bailey TM, Grodner D (2012) Distinguishing speed from accuracy in scalar implicatures. J Mem Lang 66(1):123\u2013142","journal-title":"J Mem Lang"},{"key":"2741_CR3","doi-asserted-by":"crossref","unstructured":"Bellman R (1957) A markovian decision process. J Math Mech 679\u2013684","DOI":"10.1512\/iumj.1957.6.56038"},{"key":"2741_CR4","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown T, Mann B, Ryder N, Subbiah M, Kaplan JD, Dhariwal P, Neelakantan A, Shyam P, Sastry G, Askell A et al (2020) Language models are few-shot learners. Adv Neural Inf Process Syst 33:1877\u20131901","journal-title":"Adv Neural Inf Process Syst"},{"key":"2741_CR5","doi-asserted-by":"crossref","unstructured":"Chen M, Du J, Pasunuru R, Mihaylov T, Iyer S, Stoyanov V, Kozareva Z (2022) Improving in-context few-shot learning via self-supervised training. arXiv preprint arXiv:2205.01703","DOI":"10.18653\/v1\/2022.naacl-main.260"},{"issue":"12","key":"2741_CR6","doi-asserted-by":"publisher","first-page":"5725","DOI":"10.1007\/s13042-024-02274-5","volume":"15","author":"J Chen","year":"2024","unstructured":"Chen J, Ma L, Li X, Xu J, Cho JH, Nag K, Korpeoglu E, Kumar S, Achan K (2024) Relation labeling in product knowledge graphs with large language models for e-commerce. Int J Mach Learn Cybern 15(12):5725\u20135743","journal-title":"Int J Mach Learn Cybern"},{"key":"2741_CR7","doi-asserted-by":"crossref","unstructured":"Chen D, Song S, Yu Q, Li Z, Wang W, Xiong F, Tang B (2024) Grimoire is all you need for enhancing large language models. arXiv preprint arXiv:2401.03385","DOI":"10.21203\/rs.3.rs-3845612\/v1"},{"key":"2741_CR8","unstructured":"Dubey A, Jauhri A, Pandey A, Kadian A, Al-Dahle A, Letman A, Mathur A, Schelten A, Yang A, Fan A, et al (2024) The llama 3 herd of models. arXiv preprint arXiv:2407.21783"},{"key":"2741_CR9","doi-asserted-by":"publisher","unstructured":"Dong Q, Li L, Dai D, Zheng C, Ma J, Li R, Xia H, Xu J, Wu Z, Chang B, Sun X, Li L, Sui Z (2024) A survey on in-context learning. In: Al-Onaizan Y, Bansal M, Chen Y-N (eds) Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, pp. 1107\u20131128. Association for Computational Linguistics, Miami, Florida, USA. https:\/\/doi.org\/10.18653\/v1\/2024.emnlp-main.64","DOI":"10.18653\/v1\/2024.emnlp-main.64"},{"key":"2741_CR10","doi-asserted-by":"publisher","unstructured":"Duan H, Wei J, Wang C, Liu H, Fang Y, Zhang S, Lin D, Chen K (2024) BotChat: Evaluating LLMs\u2019 capabilities of having multi-turn dialogues. In: Duh K, Gomez H, Bethard S (eds) Findings of the Association for Computational Linguistics: NAACL 2024, pp 3184\u20133200. Association for Computational Linguistics, Mexico City, Mexico. https:\/\/doi.org\/10.18653\/v1\/2024.findings-naacl.201","DOI":"10.18653\/v1\/2024.findings-naacl.201"},{"issue":"5","key":"2741_CR11","doi-asserted-by":"publisher","first-page":"378","DOI":"10.1037\/h0031619","volume":"76","author":"JL Fleiss","year":"1971","unstructured":"Fleiss JL (1971) Measuring nominal scale agreement among many raters. Psychol Bull 76(5):378","journal-title":"Psychol Bull"},{"key":"2741_CR12","doi-asserted-by":"crossref","unstructured":"Gu Y, Dong L, Wei F, Huang M (2023) Pre-training to learn in context. arXiv preprint arXiv:2305.09137","DOI":"10.18653\/v1\/2023.acl-long.267"},{"key":"2741_CR13","doi-asserted-by":"publisher","unstructured":"Gonen H, Iyer S, Blevins T, Smith N, Zettlemoyer L (2023) Demystifying prompts in language models via perplexity estimation. In: Bouamor H, Pino J, Bali K (eds) Findings of the Association for Computational Linguistics: EMNLP 2023, pp 10136\u201310148. Association for Computational Linguistics, Singapore. https:\/\/doi.org\/10.18653\/v1\/2023.findings-emnlp.679","DOI":"10.18653\/v1\/2023.findings-emnlp.679"},{"key":"2741_CR14","doi-asserted-by":"publisher","first-page":"2316","DOI":"10.1016\/j.procs.2020.04.251","volume":"171","author":"EJ George","year":"2020","unstructured":"George EJ, Mamidi R (2020) Conversational implicatures in English dialogue: annotated dataset. Procedia Comput Sci 171:2316\u20132323","journal-title":"Procedia Comput Sci"},{"key":"2741_CR15","doi-asserted-by":"crossref","unstructured":"Grice H (1975) Logic and conversation. Syntax and semantics 3","DOI":"10.1163\/9789004368811_003"},{"key":"2741_CR16","unstructured":"GLM T, Zeng A, Xu B, Wang B, Zhang C, Yin D, Rojas D, Feng G, Zhao H, Lai H et al (2024) Chatglm: a family of large language models from glm-130b to glm-4 all tools. arXiv preprint arXiv:2406.12793"},{"key":"2741_CR17","doi-asserted-by":"publisher","unstructured":"Hu J, Floyd S, Jouravlev O, Fedorenko E, Gibson E (2023) A fine-grained comparison of pragmatic language understanding in humans and language models. In: Rogers A, Boyd-Graber J, Okazaki N (eds) Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp 4194\u20134213. Association for Computational Linguistics, Toronto, Canada. https:\/\/doi.org\/10.18653\/v1\/2023.acl-long.230","DOI":"10.18653\/v1\/2023.acl-long.230"},{"key":"2741_CR18","doi-asserted-by":"publisher","unstructured":"Jeretic P, Warstadt A, Bhooshan S, Williams A (2020) Are natural language inference models IMPPRESsive? Learning IMPlicature and PRESupposition. In: Jurafsky D, Chai J, Schluter N, Tetreault J (eds) Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp 8690\u20138705. Association for Computational Linguistics, Online. https:\/\/doi.org\/10.18653\/v1\/2020.acl-main.768","DOI":"10.18653\/v1\/2020.acl-main.768"},{"key":"2741_CR19","first-page":"22199","volume":"35","author":"T Kojima","year":"2022","unstructured":"Kojima T, Gu SS, Reid M, Matsuo Y, Iwasawa Y (2022) Large language models are zero-shot reasoners. Adv Neural Inf Process Syst 35:22199\u201322213","journal-title":"Adv Neural Inf Process Syst"},{"issue":"10","key":"2741_CR20","doi-asserted-by":"publisher","first-page":"260","DOI":"10.1007\/s10462-024-10888-y","volume":"57","author":"P Kumar","year":"2024","unstructured":"Kumar P (2024) Large language models (llms): Survey, technical frameworks, and future challenges. Artif Intell Rev 57(10):260","journal-title":"Artif Intell Rev"},{"key":"2741_CR21","doi-asserted-by":"crossref","unstructured":"Lu Y, Bartolo M, Moore A, Riedel S, Stenetorp P (2021) Fantastically ordered prompts and where to find them: Overcoming few-shot prompt order sensitivity. arXiv preprint arXiv:2104.08786","DOI":"10.18653\/v1\/2022.acl-long.556"},{"key":"2741_CR22","doi-asserted-by":"publisher","unstructured":"Li X, Lv K, Yan H, Lin T, Zhu W, Ni Y, Xie G, Wang X, Qiu X (2023) Unified demonstration retriever for in-context learning. In: Rogers, A, Boyd-Graber J, Okazaki N (eds) Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp 4644\u20134668. Association for Computational Linguistics, Toronto, Canada. https:\/\/doi.org\/10.18653\/v1\/2023.acl-long.256","DOI":"10.18653\/v1\/2023.acl-long.256"},{"key":"2741_CR23","unstructured":"Li Y, Ma X, Lu S, Lee K, Liu X, Guo C (2024) Mend: Meta demonstration distillation for efficient and effective in-context learning. arXiv preprint arXiv:2403.06914"},{"key":"2741_CR24","doi-asserted-by":"publisher","unstructured":"Li X, Qiu X (2023) Finding support examples for in-context learning. In: Bouamor, H., Pino, J., Bali, K. (eds.) Findings of the Association for Computational Linguistics: EMNLP 2023, pp. 6219\u20136235. Association for Computational Linguistics, Singapore . https:\/\/doi.org\/10.18653\/v1\/2023.findings-emnlp.411","DOI":"10.18653\/v1\/2023.findings-emnlp.411"},{"key":"2741_CR25","doi-asserted-by":"publisher","unstructured":"Louis A, Roth D, Radlinski F (2020) \u201cI\u2019d rather just go to bed\u201d: Understanding indirect answers. In: Webber, B., Cohn, T., He, Y., Liu, Y. (eds.) Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 7411\u20137425. Association for Computational Linguistics, Online. https:\/\/doi.org\/10.18653\/v1\/2020.emnlp-main.601","DOI":"10.18653\/v1\/2020.emnlp-main.601"},{"key":"2741_CR26","unstructured":"Li E, Schuster S, Degen J (2021) Predicting scalar inferences from \u201cor\u201d to \u201cnot both\u201d using neural sentence encoders. In: Proceedings of the Society for Computation in Linguistics 2021, pp. 446\u2013450"},{"key":"2741_CR27","doi-asserted-by":"publisher","unstructured":"Liu J, Shen D, Zhang Y, Dolan B, Carin L, Chen W (2022) What makes good in-context examples for GPT-3? In: Agirre, E., Apidianaki, M., Vuli\u0107, I. (eds.) Proceedings of Deep Learning Inside Out (DeeLIO 2022): The 3rd Workshop on Knowledge Extraction and Integration for Deep Learning Architectures, pp. 100\u2013114. Association for Computational Linguistics, Dublin, Ireland and Online. https:\/\/doi.org\/10.18653\/v1\/2022.deelio-1.10","DOI":"10.18653\/v1\/2022.deelio-1.10"},{"key":"2741_CR28","doi-asserted-by":"publisher","unstructured":"Min S, Lewis M, Zettlemoyer L, Hajishirzi H (2022) MetaICL: Learning to learn in context. In: Carpuat, M., Marneffe, M.-C., Meza\u00a0Ruiz, I.V. (eds.) Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 2791\u20132809. Association for Computational Linguistics, Seattle, United States. https:\/\/doi.org\/10.18653\/v1\/2022.naacl-main.201","DOI":"10.18653\/v1\/2022.naacl-main.201"},{"issue":"5","key":"2741_CR29","doi-asserted-by":"publisher","first-page":"886","DOI":"10.17507\/tpls.1205.08","volume":"12","author":"R Musa","year":"2022","unstructured":"Musa R, Mohammed BK (2022) The role of conversational implicature in daily conversations-what matters, content or context? Theory and Practice in Language Studies 12(5):886\u2013893","journal-title":"Theory and Practice in Language Studies"},{"key":"2741_CR30","unstructured":"Nguyen T, Wong E (2023) In-context example selection with influences. arXiv preprint arXiv:2302.11042"},{"key":"2741_CR31","doi-asserted-by":"publisher","unstructured":"Ou J, Lu J, Liu C, Tang Y, Zhang F, Zhang D, Gai K (2024) DialogBench: Evaluating LLMs as human-like dialogue systems. In: Duh, K., Gomez, H., Bethard, S. (eds.) Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), pp. 6137\u20136170. Association for Computational Linguistics, Mexico City, Mexico . https:\/\/doi.org\/10.18653\/v1\/2024.naacl-long.341","DOI":"10.18653\/v1\/2024.naacl-long.341"},{"key":"2741_CR32","doi-asserted-by":"crossref","unstructured":"Qin C, Zhang A, Dagar A, Ye W (2023) In-context learning with iterative demonstration selection. arXiv preprint arXiv:2310.09881","DOI":"10.18653\/v1\/2024.findings-emnlp.438"},{"key":"2741_CR33","doi-asserted-by":"publisher","unstructured":"Rubin O, Herzig J, Berant J (2022) Learning to retrieve prompts for in-context learning. In: Carpuat, M., Marneffe, M.-C., Meza\u00a0Ruiz, I.V. (eds.) Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 2655\u20132671. Association for Computational Linguistics, Seattle, United States. https:\/\/doi.org\/10.18653\/v1\/2022.naacl-main.191","DOI":"10.18653\/v1\/2022.naacl-main.191"},{"key":"2741_CR34","unstructured":"Ruis L, Khan A, Biderman S, Hooker S, Rockt\u00e4schel T, Grefenstette E (2024) The goldilocks of pragmatic understanding: Fine-tuning strategy matters for implicature resolution by llms. Advances in Neural Information Processing Systems 36"},{"key":"2741_CR35","doi-asserted-by":"publisher","unstructured":"Schuster S, Chen Y, Degen J (2020) Harnessing the linguistic signal to predict scalar inferences. In: Jurafsky, D., Chai, J., Schluter, N., Tetreault, J. (eds.) Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 5387\u20135403. Association for Computational Linguistics, Online. https:\/\/doi.org\/10.18653\/v1\/2020.acl-main.479","DOI":"10.18653\/v1\/2020.acl-main.479"},{"key":"2741_CR36","doi-asserted-by":"publisher","unstructured":"Sravanthi S, Doshi M, Tankala P, Murthy R, Dabre R, Bhattacharyya P (2024) PUB: A pragmatics understanding benchmark for assessing LLMs\u2019 pragmatics capabilities. In: Ku, L.-W., Martins, A., Srikumar, V. (eds.) Findings of the Association for Computational Linguistics: ACL 2024, pp. 12075\u201312097. Association for Computational Linguistics, Bangkok, Thailand . https:\/\/doi.org\/10.18653\/v1\/2024.findings-acl.719","DOI":"10.18653\/v1\/2024.findings-acl.719"},{"key":"2741_CR37","unstructured":"Su H, Kasai J, Wu CH, Shi W, Wang T, Xin J, Zhang R, Ostendorf M, Zettlemoyer L, Smith NA, et al (2022) Selective annotation makes language models better few-shot learners. arXiv preprint arXiv:2209.01975"},{"key":"2741_CR38","unstructured":"Shi W, Min S, Lomeli M, Zhou C, Li M, Lin V, Smith NA, Zettlemoyer L, Yih S, Lewis M (2023) In-context pretraining: Language modeling beyond document boundaries. arXiv preprint arXiv:2310.10638"},{"key":"2741_CR39","doi-asserted-by":"publisher","unstructured":"Sorensen T, Robinson J, Rytting C, Shaw A, Rogers K, Delorey A, Khalil M, Fulda N, Wingate D (2022) An information-theoretic approach to prompt engineering without ground truth labels. In: Muresan, S., Nakov, P., Villavicencio, A. (eds.) Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 819\u2013862. Association for Computational Linguistics, Dublin, Ireland . https:\/\/doi.org\/10.18653\/v1\/2022.acl-long.60","DOI":"10.18653\/v1\/2022.acl-long.60"},{"key":"2741_CR40","unstructured":"Srivastava A, Rastogi A, Rao A, Shoeb AAM, Abid A, Fisch A, Brown AR, Santoro A, Gupta A, Garriga-Alonso A, et al (2022) Beyond the imitation game: Quantifying and extrapolating the capabilities of language models. arXiv preprint arXiv:2206.04615"},{"key":"2741_CR41","doi-asserted-by":"publisher","unstructured":"Sanagavarapu K, Singaraju J, Kakileti A, Kaza A, Mathews A, Li H, Brito N, Blanco E (2022) Disentangling indirect answers to yes-no questions in real conversations. In: Carpuat, M., Marneffe, M.-C., Meza\u00a0Ruiz, I.V. (eds.) Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 4677\u20134695. Association for Computational Linguistics, Seattle, United States . https:\/\/doi.org\/10.18653\/v1\/2022.naacl-main.345","DOI":"10.18653\/v1\/2022.naacl-main.345"},{"issue":"3","key":"2741_CR42","first-page":"93","volume":"2","author":"A Triyuono","year":"2022","unstructured":"Triyuono A, Dirham UR (2022) Pragmatic competence or pragmatic knowledge and its role in linguistic communication. Journal of Learning and Instructional Studies 2(3):93\u2013106","journal-title":"Journal of Learning and Instructional Studies"},{"key":"2741_CR43","doi-asserted-by":"crossref","unstructured":"Tanwar E, Dutta S, Borthakur M, Chakraborty T (2023) Multilingual llms are better cross-lingual in-context learners with alignment. arXiv preprint arXiv:2305.05940","DOI":"10.18653\/v1\/2023.acl-long.346"},{"key":"2741_CR44","doi-asserted-by":"publisher","unstructured":"Takayama J, Kajiwara T, Arase Y (2021) DIRECT: Direct and indirect responses in conversational text corpus. In: Moens, M.-F., Huang, X., Specia, L., Yih, S.W.-t. (eds.) Findings of the Association for Computational Linguistics: EMNLP 2021, pp. 1980\u20131989. Association for Computational Linguistics, Punta Cana, Dominican Republic. https:\/\/doi.org\/10.18653\/v1\/2021.findings-emnlp.170","DOI":"10.18653\/v1\/2021.findings-emnlp.170"},{"key":"2741_CR45","unstructured":"Touvron H, Martin L, Stone K, Albert P, Almahairi A, Babaei Y, Bashlykov N, Batra S, Bhargava P, Bhosale S, et al (2023) Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288"},{"key":"2741_CR46","doi-asserted-by":"crossref","unstructured":"Wei J, Hou L, Lampinen A, Chen X, Huang D, Tay Y, Chen X, Lu Y, Zhou D, Ma T, et al (2023) Symbol tuning improves in-context learning in language models. arXiv preprint arXiv:2305.08298","DOI":"10.18653\/v1\/2023.emnlp-main.61"},{"key":"2741_CR47","doi-asserted-by":"publisher","unstructured":"Wang Z, Hossain M, Mathur S, Melo T, Ozler, K, Park K, Quintero J, Rezaei M, Shakya S, Uddin M, Blanco E (2023) Interpreting indirect answers to yes-no questions in multiple languages. In: Bouamor, H., Pino, J., Bali, K. (eds.) Findings of the Association for Computational Linguistics: EMNLP 2023, pp. 2210\u20132227. Association for Computational Linguistics, Singapore. https:\/\/doi.org\/10.18653\/v1\/2023.findings-emnlp.146","DOI":"10.18653\/v1\/2023.findings-emnlp.146"},{"key":"2741_CR48","doi-asserted-by":"crossref","unstructured":"Wang Y, Kordi Y, Mishra S, Liu A, Smith NA, Khashabi D, Hajishirzi H (2022) Self-instruct: Aligning language models with self-generated instructions. arXiv preprint arXiv:2212.10560","DOI":"10.18653\/v1\/2023.acl-long.754"},{"key":"2741_CR49","doi-asserted-by":"publisher","unstructured":"Wang Y, Mishra S, Alipoormolabashi P, Kordi Y, Mirzaei A, Naik A, Ashok A, Dhanasekaran AS, Arunkumar A, Stap D, Pathak E, Karamanolakis G, Lai H, Purohit I, Mondal I, Anderson J, Kuznia K, Doshi K, Pal KK, Patel M, Moradshahi M, Parmar M, Purohit M, Varshney N, Kaza PR, Verma P, Puri RS, Karia R, Doshi S, Sampat SK, Mishra S, Reddy AS, Patro S, Dixit T, Shen X (2022) Super-NaturalInstructions: Generalization via declarative instructions on 1600+ NLP tasks. In: Goldberg, Y., Kozareva, Z., Zhang, Y. (eds.) Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp. 5085\u20135109. Association for Computational Linguistics, Abu Dhabi, United Arab Emirates. https:\/\/doi.org\/10.18653\/v1\/2022.emnlp-main.340","DOI":"10.18653\/v1\/2022.emnlp-main.340"},{"key":"2741_CR50","doi-asserted-by":"publisher","unstructured":"Wang Z, Rashid F, Blanco E (2024) Interpreting answers to yes-no questions in dialogues from multiple domains. In: Duh, K., Gomez, H., Bethard, S. (eds.) Findings of the Association for Computational Linguistics: NAACL 2024, pp. 2111\u20132128. Association for Computational Linguistics, Mexico City, Mexico. https:\/\/doi.org\/10.18653\/v1\/2024.findings-naacl.136","DOI":"10.18653\/v1\/2024.findings-naacl.136"},{"key":"2741_CR51","first-page":"24824","volume":"35","author":"J Wei","year":"2022","unstructured":"Wei J, Wang X, Schuurmans D, Bosma M, Xia F, Chi E, Le QV, Zhou D et al (2022) Chain-of-thought prompting elicits reasoning in large language models. Adv Neural Inf Process Syst 35:24824\u201324837","journal-title":"Adv Neural Inf Process Syst"},{"key":"2741_CR52","doi-asserted-by":"crossref","unstructured":"Wu Z, Wang Y, Ye J, Kong L (2022) Self-adaptive in-context learning: An information compression perspective for in-context example selection and ordering. arXiv preprint arXiv:2212.10375","DOI":"10.18653\/v1\/2023.acl-long.79"},{"key":"2741_CR53","doi-asserted-by":"publisher","unstructured":"Wu S, Yang S, Chen Z, Su Q (2024) Rethinking pragmatics in large language models: Towards open-ended evaluation and preference tuning. In: Al-Onaizan, Y., Bansal, M., Chen, Y.-N. (eds.) Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, pp. 22583\u201322599. Association for Computational Linguistics, Miami, Florida, USA . https:\/\/doi.org\/10.18653\/v1\/2024.emnlp-main.1258","DOI":"10.18653\/v1\/2024.emnlp-main.1258"},{"key":"2741_CR54","doi-asserted-by":"crossref","unstructured":"Yue S, Song S, Cheng X, Hu H (2024) Do large language models understand conversational implicature\u2013a case study with a chinese sitcom. In: China National Conference on Chinese Computational Linguistics, pp. 402\u2013418. Springer","DOI":"10.1007\/978-981-97-8367-0_24"},{"key":"2741_CR55","doi-asserted-by":"publisher","unstructured":"Yerukola A, Vaduguru S, Fried D, Sap M (2024) Is the pope catholic? yes, the pope is catholic. generative evaluation of non-literal intent resolution in LLMs. In: Ku, L.-W., Martins, A., Srikumar, V. (eds.) Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pp. 265\u2013275. Association for Computational Linguistics, Bangkok, Thailand. https:\/\/doi.org\/10.18653\/v1\/2024.acl-short.26","DOI":"10.18653\/v1\/2024.acl-short.26"},{"key":"2741_CR56","doi-asserted-by":"crossref","unstructured":"Zhang Y, Feng S, Tan C (2022) Active example selection for in-context learning. arXiv preprint arXiv:2211.04486","DOI":"10.18653\/v1\/2022.emnlp-main.622"},{"key":"2741_CR57","unstructured":"Zhou Y, Muresanu AI, Han Z, Paster K, Pitis S, Chan H, Ba J (2022) Large language models are human-level prompt engineers. arXiv preprint arXiv:2211.01910"},{"key":"2741_CR58","doi-asserted-by":"crossref","unstructured":"Zheng Z, Qiu S, Fan L, Zhu Y, Zhu S-C (2021) Grice: A grammar-based dataset for recovering implicature and conversational reasoning. In: Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pp. 2074\u20132085","DOI":"10.18653\/v1\/2021.findings-acl.182"},{"key":"2741_CR59","unstructured":"Zhao Z, Wallace E, Feng S, Klein D, Singh S (2021) Calibrate before use: Improving few-shot performance of language models. In: International Conference on Machine Learning, pp. 12697\u201312706. PMLR"}],"container-title":["International Journal of Machine Learning and Cybernetics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13042-025-02741-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s13042-025-02741-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13042-025-02741-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,7]],"date-time":"2025-11-07T11:23:54Z","timestamp":1762514634000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s13042-025-02741-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,19]]},"references-count":59,"journal-issue":{"issue":"11","published-print":{"date-parts":[[2025,11]]}},"alternative-id":["2741"],"URL":"https:\/\/doi.org\/10.1007\/s13042-025-02741-7","relation":{},"ISSN":["1868-8071","1868-808X"],"issn-type":[{"type":"print","value":"1868-8071"},{"type":"electronic","value":"1868-808X"}],"subject":[],"published":{"date-parts":[[2025,7,19]]},"assertion":[{"value":"4 January 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 July 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 July 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no Conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}