{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,20]],"date-time":"2026-01-20T14:23:53Z","timestamp":1768919033696,"version":"3.49.0"},"reference-count":70,"publisher":"Springer Science and Business Media LLC","issue":"22","license":[{"start":{"date-parts":[[2024,4,22]],"date-time":"2024-04-22T00:00:00Z","timestamp":1713744000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,4,22]],"date-time":"2024-04-22T00:00:00Z","timestamp":1713744000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Science and Technology Program project of Shanghai Municipal Committee of Science and Technology","award":["22511104800"],"award-info":[{"award-number":["22511104800"]}]},{"name":"Science and Technology Program project of Shanghai Municipal Committee of Science and Technology","award":["22DZ204903"],"award-info":[{"award-number":["22DZ204903"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Comput &amp; Applic"],"published-print":{"date-parts":[[2024,8]]},"DOI":"10.1007\/s00521-024-09728-5","type":"journal-article","created":{"date-parts":[[2024,4,22]],"date-time":"2024-04-22T09:02:25Z","timestamp":1713776545000},"page":"13491-13512","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["Diluie: constructing diverse demonstrations of in-context learning with large language model for unified information extraction"],"prefix":"10.1007","volume":"36","author":[{"given":"Qian","family":"Guo","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6088-7198","authenticated-orcid":false,"given":"Yi","family":"Guo","sequence":"additional","affiliation":[]},{"given":"Jin","family":"Zhao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,4,22]]},"reference":[{"key":"9728_CR1","doi-asserted-by":"crossref","unstructured":"Al-Rfou R, Kulkarni V, Perozzi B, Skiena S (2015) Polyglot-ner: massive multilingual named entity recognition. In: In: Proceedings of the 2015 SIAM international conference on data mining. SIAM, pp 586\u2013594","DOI":"10.1137\/1.9781611974010.66"},{"key":"9728_CR2","unstructured":"Chen W, Ma X, Wang X, Cohen WW (2022) Program of thoughts prompting: disentangling computation from reasoning for numerical reasoning tasks. arXiv preprint arXiv:2211.12588"},{"key":"9728_CR3","doi-asserted-by":"crossref","unstructured":"Chen P, Xu H, Zhang C, Huang R (2022) Crossroads, buildings and neighborhoods: a dataset for fine-grained location recognition. In: Proceedings of the 2022 conference of the North American chapter of the association for computational linguistics: human language technologies, pp 3329\u20133339","DOI":"10.18653\/v1\/2022.naacl-main.243"},{"key":"9728_CR4","unstructured":"Chen X, Ye J, Zu C, Xu N, Zheng R, Peng M, Zhou J, Gui T, Zhang Q, Huang X (2023) How robust is gpt-3.5 to predecessors? A comprehensive study on language understanding tasks. arXiv preprint arXiv:2303.00293"},{"key":"9728_CR5","doi-asserted-by":"crossref","unstructured":"Chia YK, Bing L, Poria S, Si L (2022) Relationprompt: leveraging prompts to generate synthetic data for zero-shot relation triplet extraction. arXiv preprint arXiv:2203.09101","DOI":"10.18653\/v1\/2022.findings-acl.5"},{"key":"9728_CR6","unstructured":"Chowdhery A, Narang S, Devlin J, Bosma M, Mishra G, Roberts A, Barham P, Chung HW, Sutton C, Gehrmann S, Schuh P (2022) Palm: scaling language modeling with pathways. arXiv preprint arXiv:2204.02311"},{"key":"9728_CR7","unstructured":"Chung HW, Hou L, Longpre S, Zoph B, Tay Y, Fedus W, Li Y, Wang X, Dehghani M, Brahma S, Webson A (2022) Scaling instruction-finetuned language models. arXiv preprint arXiv:2210.11416"},{"key":"9728_CR8","doi-asserted-by":"publisher","unstructured":"Wang X, Zhu W, Wang WY (2023) Large language models are im\u00adplicitly topic models: Explaining and finding good demonstrations for in\u00adcontext learning. corr, abs\/2301.11916. https:\/\/doi.org\/10.48550\/arXiv.2301.11916","DOI":"10.48550\/arXiv.2301.11916"},{"key":"9728_CR9","unstructured":"Derczynski L, Bontcheva K, Roberts I (2016) Broad twitter corpus: a diverse named entity recognition resource. In: Proceedings of COLING 2016, the 26th international conference on computational linguistics: technical papers, pp 1169\u20131179"},{"key":"9728_CR10","unstructured":"Devlin J, Chang MW, Lee K, Toutanova K (2018) Bert: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805"},{"key":"9728_CR11","unstructured":"Dong Q, Li L, Dai D, Zheng C, Wu Z, Chang B, Sun X, Xu J, Sui Z (2022) A survey for in-context learning. arXiv preprint arXiv:2301.00234"},{"key":"9728_CR12","first-page":"15460","volume":"35","author":"H Fei","year":"2022","unstructured":"Fei H, Wu S, Li J, Li B, Li F, Qin L, Zhang M, Zhang M, Chua TS (2022) Lasuie: unifying information extraction with latent adaptive structure-aware generative language model. Adv Neural Inf Process Syst 35:15460\u201315475","journal-title":"Adv Neural Inf Process Syst"},{"key":"9728_CR13","doi-asserted-by":"crossref","unstructured":"Guan R, Man KL, Chen F, Yao S, Hu R, Zhu X, Smith J, Lim EG, Yue Y (2023) Findvehicle and vehiclefinder: a ner dataset for natural language-based vehicle retrieval and a keyword-based cross-modal vehicle retrieval system. arXiv preprint arXiv:2304.10893","DOI":"10.1007\/s11042-023-16373-y"},{"issue":"5","key":"9728_CR14","doi-asserted-by":"publisher","first-page":"885","DOI":"10.1016\/j.jbi.2012.04.008","volume":"45","author":"H Gurulingappa","year":"2012","unstructured":"Gurulingappa H, Rajput AM, Roberts A, Fluck J, Hofmann-Apitius M, Toldo L (2012) Development of a benchmark corpus to support the automatic extraction of drug-related adverse effects from medical case reports. J Biomed Informat 45(5):885\u2013892","journal-title":"J Biomed Informat"},{"key":"9728_CR15","unstructured":"Hendrickx I, Kim SN, Kozareva Z, Nakov P, S\u00e9aghdha DO, Pad\u00f3 S, Pennacchiotti M, Romano L, Szpakowicz S (2019) Semeval-2010 task 8: multi-way classification of semantic relations between pairs of nominals. arXiv preprint arXiv:1911.10422"},{"key":"9728_CR16","doi-asserted-by":"crossref","unstructured":"He J, Wang L, Hu Y, Liu N, Liu H, Xu X, Shen HT (2023) Icl-d3ie: in-context learning with diverse demonstrations updating for document information extraction. arXiv preprint arXiv:2303.05063","DOI":"10.1109\/ICCV51070.2023.01785"},{"key":"9728_CR17","doi-asserted-by":"crossref","unstructured":"Hovy E, Marcus M, Palmer M, Ramshaw L, Weischedel R (2006) Ontonotes: the 90% solution. In: Proceedings of the human language technology conference of the NAACL, companion volume: short papers, pp 57\u201360","DOI":"10.3115\/1614049.1614064"},{"key":"9728_CR18","unstructured":"Jat S, Khandelwal S, Talukdar P (2018) Improving distantly supervised relation extraction using word and entity based attention. arXiv preprint arXiv:1804.06987"},{"key":"9728_CR19","unstructured":"Jordan Michael\u00a0I, Kearns Michael\u00a0J, Solla Sara\u00a0A (1998) Advances in Neural Information Processing Systems 10: Proceedings of the 1997 Conference, vol\u00a010. MIT Press"},{"key":"9728_CR20","doi-asserted-by":"crossref","unstructured":"Kim J-D, Ohta T, Tateisi Y, Tsujii J (2003) Genia corpus-a semantically annotated corpus for bio-textmining. Bioinformatics, 19(suppl_1):i180\u2013i182","DOI":"10.1093\/bioinformatics\/btg1023"},{"key":"9728_CR21","unstructured":"Kingma Diederik\u00a0P, Ba Jimmy (2014) Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980"},{"key":"9728_CR22","doi-asserted-by":"publisher","DOI":"10.1016\/j.simpa.2022.100373","volume":"13","author":"V Kocaman","year":"2022","unstructured":"Kocaman V, Talby D (2022) Accurate clinical and biomedical named entity recognition at scale. Softw Impacts 13:100373","journal-title":"Softw Impacts"},{"key":"9728_CR23","doi-asserted-by":"crossref","unstructured":"Kocaman V, Talby D (2021) Biomedical named entity recognition at scale. In: Pattern recognition. ICPR international workshops and challenges: virtual event, January 10\u201315, 2021, proceedings, part I. Springer, pp 635\u2013646","DOI":"10.1007\/978-3-030-68763-2_48"},{"issue":"1","key":"9728_CR24","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/1758-2946-7-S1-S1","volume":"7","author":"M Krallinger","year":"2015","unstructured":"Krallinger M, Rabal O, Leitner F, Vazquez M, Salgado D, Lu Z, Leaman R, Lu Y, Ji D, Lowe DM, Sayle RA (2015) The chemdner corpus of chemicals and drugs and its annotation principles. J Cheminformat 7(1):1\u201317","journal-title":"J Cheminformat"},{"key":"9728_CR25","doi-asserted-by":"crossref","unstructured":"Abramski K, Citraro S, Lombardi L, Rossetti G, Stella M (2023) Cognitive network science reveals bias in gpt\u00ad3, gpt\u00ad3.5 turbo, and gpt\u00ad4 mirroring math anxiety in high\u00adschool students. Big Data and Cognitive Computing 7(3):124","DOI":"10.3390\/bdcc7030124"},{"key":"9728_CR26","unstructured":"Li M, Gong S, Feng J, Xu Y, Zhang J, Wu Z, Kong L (2023) In-context learning with many demonstration examples. arXiv preprint arXiv:2302.04931"},{"key":"9728_CR27","doi-asserted-by":"crossref","unstructured":"Li J, Sun Y, Johnson RJ, Sciaky D, Wei CH, Leaman R, Davis AP, Mattingly CJ, Wiegers TC, Lu Z (2016) Biocreative v cdr task corpus: a resource for chemical disease relation extraction. In: Database","DOI":"10.1093\/database\/baw068"},{"key":"9728_CR28","first-page":"13452","volume":"35","author":"Z Liu","year":"2021","unstructured":"Liu Z, Xu Y, Yu T, Dai W, Ji Z, Cahyawijaya S, Madotto A, Fung P (2021) Crossner: evaluating cross-domain named entity recognition. Proc AAAI Confer Artif Intell 35:13452\u201313460","journal-title":"Proc AAAI Confer Artif Intell"},{"key":"9728_CR29","doi-asserted-by":"crossref","unstructured":"Liu Y, Meng F, Zhang J, Xu J, Chen Y, Zhou J (2019) Gcdt: a global context enhanced deep transition architecture for sequence labeling. arXiv preprint arXiv:1906.02437","DOI":"10.18653\/v1\/P19-1233"},{"key":"9728_CR30","doi-asserted-by":"crossref","unstructured":"Liu J, Shen D, Zhang Y, Dolan B, Carin L, Chen W (2021) What makes good in-context examples for gpt-3? arXiv preprint arXiv:2101.06804","DOI":"10.18653\/v1\/2022.deelio-1.10"},{"key":"9728_CR31","doi-asserted-by":"crossref","unstructured":"Lou J, Lu Y, Dai D, Jia W, Lin H, Han X, Sun L, Wu H (2023) Universal information extraction as unified semantic matching. arXiv preprint arXiv:2301.03282","DOI":"10.1609\/aaai.v37i11.26563"},{"key":"9728_CR32","doi-asserted-by":"crossref","unstructured":"Luan Y, He L, Ostendorf M, Hajishirzi H (2018) Multi-task identification of entities, relations, and coreference for scientific knowledge graph construction. arXiv preprint arXiv:1808.09602","DOI":"10.18653\/v1\/D18-1360"},{"key":"9728_CR33","doi-asserted-by":"crossref","unstructured":"Lu Y, Bartolo M, Moore A, Riedel S, Stenetorp P (2021) Fantastically ordered prompts and where to find them: overcoming few-shot prompt order sensitivity. arXiv preprint arXiv:2104.08786","DOI":"10.18653\/v1\/2022.acl-long.556"},{"key":"9728_CR34","doi-asserted-by":"crossref","unstructured":"Lu Y, Lin H, Xu J, Han X, Tang J, Li A, Sun L, Liao M, Chen S (2021) Text2event: controllable sequence-to-structure generation for end-to-end event extraction. arXiv preprint arXiv:2106.09232","DOI":"10.18653\/v1\/2021.acl-long.217"},{"key":"9728_CR35","doi-asserted-by":"crossref","unstructured":"Lu Y, Liu Q, Dai D, Xiao X, Lin H, Han X, Sun L, Wu H (2022) Unified structure generation for universal information extraction. arXiv preprint arXiv:2203.12277","DOI":"10.18653\/v1\/2022.acl-long.395"},{"key":"9728_CR36","doi-asserted-by":"crossref","unstructured":"Min S, Lewis M, Zettlemoyer L, Hajishirzi H (2021) Metaicl: learning to learn in context. arXiv preprint arXiv:2110.15943","DOI":"10.18653\/v1\/2022.naacl-main.201"},{"key":"9728_CR37","doi-asserted-by":"crossref","unstructured":"Mirowski P, Steck H, Whiting P, Palaniappan R, MacDonald M, Ho TK (2011) Kl-divergence kernel regression for non-gaussian fingerprint based localization. In: 2011 international conference on Indoor positioning and Indoor navigation. IEEE, pp 1\u201310","DOI":"10.1109\/IPIN.2011.6071928"},{"key":"9728_CR38","first-page":"1","volume":"1","author":"A Mitchell","year":"2005","unstructured":"Mitchell A, Strassel S, Huang S, Zakhary R (2005) Ace 2004 multilingual training corpus. Linguist Data Consort, Phila 1:1","journal-title":"Linguist Data Consort, Phila"},{"key":"9728_CR39","doi-asserted-by":"crossref","unstructured":"Ott M, Edunov S, Baevski A, Fan A, Gross S, Ng N, Grangier D, Auli M (2019) fairseq: a fast, extensible toolkit for sequence modeling. arXiv preprint arXiv:1904.01038","DOI":"10.18653\/v1\/N19-4009"},{"key":"9728_CR40","first-page":"27730","volume":"35","author":"L Ouyang","year":"2022","unstructured":"Ouyang L, Wu J, Jiang X, Almeida D, Wainwright C, Mishkin P, Zhang C, Agarwal S, Slama K, Ray A, Schulman J (2022) Training language models to follow instructions with human feedback. Adv Neural Inf Process Syst 35:27730\u201327744","journal-title":"Adv Neural Inf Process Syst"},{"key":"9728_CR41","doi-asserted-by":"crossref","unstructured":"Pan X, Zhang B, May J, Nothman J, Knight K, Ji H (2017) Cross-lingual name tagging and linking for 282 languages. In: Proceedings of the 55th annual meeting of the association for computational linguistics (volume 1: long papers), pp 1946\u20131958","DOI":"10.18653\/v1\/P17-1178"},{"key":"9728_CR42","first-page":"1","volume":"1","author":"N Poolsawad","year":"2014","unstructured":"Poolsawad N, Kambhampati C, Cleland JGF (2014) Balancing class for performance of classification with a clinical dataset. Proc World Congr Eng 1:1\u20136","journal-title":"Proc World Congr Eng"},{"key":"9728_CR43","doi-asserted-by":"crossref","unstructured":"Riedel S, Yao L, McCallum A (2010) Modeling relations and their mentions without labeled text. In: Machine learning and knowledge discovery in databases: European conference, ECML PKDD 2010, Barcelona, Spain, September 20\u201324, 2010, proceedings, part III, vol 21. Springer, pp 148\u2013163","DOI":"10.1007\/978-3-642-15939-8_10"},{"key":"9728_CR44","unstructured":"Roth D, Yih W-T (2004) A linear programming formulation for global inference in natural language tasks. In: Proceedings of the 8th conference on computational natural language learning (CoNLL-2004) at HLT-NAACL 2004, pp 1\u20138"},{"key":"9728_CR45","unstructured":"Sang EF, De Meulder F (2003) Introduction to the conll-2003 shared task: Language-independent named entity recognition. arXiv preprint arXiv:cs\/0306050"},{"key":"9728_CR46","doi-asserted-by":"crossref","unstructured":"Shaikh O, Zhang H, Held W, Bernstein M, Yang D (2022) On second thought, let\u2019s not think step by step! bias and toxicity in zero-shot reasoning. arXiv preprint arXiv:2212.08061","DOI":"10.18653\/v1\/2023.acl-long.244"},{"key":"9728_CR47","doi-asserted-by":"crossref","unstructured":"Sun Z, Li J, Pergola G, Wallace BC, John B, Greene N, Kim J, He Y (2022) Phee: a dataset for pharmacovigilance event extraction from text. arXiv preprint arXiv:2210.12560","DOI":"10.18653\/v1\/2022.emnlp-main.376"},{"key":"9728_CR48","first-page":"7072","volume":"33","author":"R Takanobu","year":"2019","unstructured":"Takanobu R, Zhang T, Liu J, Huang M (2019) A hierarchical framework for relation extraction with reinforcement learning. Proc AAAI Confer Artif Intell 33:7072\u20137079","journal-title":"Proc AAAI Confer Artif Intell"},{"key":"9728_CR49","doi-asserted-by":"crossref","unstructured":"Tedeschi S, Navigli R (2022) Multinerd: a multilingual, multi-genre and fine-grained dataset for named entity recognition (and disambiguation). In: Findings of the association for computational linguistics: NAACL 2022, pp 801\u2013812","DOI":"10.18653\/v1\/2022.findings-naacl.60"},{"key":"9728_CR50","doi-asserted-by":"crossref","unstructured":"Tedeschi S, Maiorca V, Campolungo N, Cecconi F, Navigli R (2021) Wikineural: combined neural and knowledge-based silver data creation for multilingual NER. In: Findings of the association for computational linguistics: EMNLP 2021, pp 2521\u20132533","DOI":"10.18653\/v1\/2021.findings-emnlp.215"},{"key":"9728_CR51","unstructured":"Ushio A, Neves L, Silva V, Barbieri F, Camacho-Collados J (2022) Named entity recognition in twitter: a dataset and analysis on short-term temporal shifts. arXiv preprint arXiv:2210.03797"},{"key":"9728_CR52","first-page":"45","volume":"57","author":"C Walker","year":"2006","unstructured":"Walker C, Strassel S, Medero J, Maeda K (2006) Ace 2005 multilingual training corpus. Linguist Data Consort, Phila 57:45","journal-title":"Linguist Data Consort, Phila"},{"key":"9728_CR53","doi-asserted-by":"crossref","unstructured":"Wan Z, Cheng F, Mao Z, Liu Q, Song H, Li J, Kurohashi S (2023) Gpt-re: in-context learning for relation extraction using large language models. arXiv preprint arXiv:2305.02105","DOI":"10.18653\/v1\/2023.emnlp-main.214"},{"key":"9728_CR54","doi-asserted-by":"crossref","unstructured":"Wang X, Dou S, Xiong L, Zou Y, Zhang Q, Gui T, Qiao L, Cheng Z, Huang X (2022) Miner: Improving out-of-vocabulary named entity recognition from an information theoretic perspective. arXiv preprint arXiv:2204.04391","DOI":"10.18653\/v1\/2022.acl-long.383"},{"key":"9728_CR55","doi-asserted-by":"crossref","unstructured":"Wang B, Min S, Deng X, Shen J, Wu Y, Zettlemoyer L, Sun H (2022) Toward understanding chain-of-thought prompting: an empirical study of what matters. arXiv preprint arXiv:2212.10001","DOI":"10.18653\/v1\/2023.acl-long.153"},{"key":"9728_CR56","unstructured":"Wang X, Zhou W, Zu C, Xia H, Chen T, Zhang Y, Zheng R, Ye J, Zhang Q, Gui T, Kang J (2023) Instructuie: Multi-task instruction tuning for unified information extraction. arXiv preprint arXiv:2304.08085"},{"key":"9728_CR57","unstructured":"Wang X, Zhu W, Saxon M, Steyvers M, Wang WY (2023) Large language models are implicitly topic models: explaining and finding good demonstrations for in-context learning. arXiv preprint arXiv:2301.11916"},{"key":"9728_CR58","first-page":"24824","volume":"35","author":"J Wei","year":"2022","unstructured":"Wei J, Wang X, Schuurmans D, Bosma M, Xia F, Chi E, Le QV, Zhou D (2022) Chain-of-thought prompting elicits reasoning in large language models. Adv Neural Inf Process Syst 35:24824\u201324837","journal-title":"Adv Neural Inf Process Syst"},{"key":"9728_CR59","unstructured":"Wu Z, Wang Y, Ye J, Kong L (2022) Self-adaptive in-context learning. arXiv preprint arXiv:2212.10375"},{"key":"9728_CR60","unstructured":"Xie SM, Raghunathan A, Liang P, Ma T (2021) An explanation of in-context learning as implicit bayesian inference. arXiv preprint arXiv:2111.02080"},{"key":"9728_CR61","unstructured":"Xie SM, Raghunathan A, Liang P, Ma T (2023) Efficient attention via control variates. arXiv preprint arXiv:2302.04542"},{"key":"9728_CR62","doi-asserted-by":"crossref","unstructured":"Yan H, Dai J, Qiu X, Zhang Z (2021) A unified generative framework for aspect-based sentiment analysis. arXiv preprint arXiv:2106.04300","DOI":"10.18653\/v1\/2021.acl-long.188"},{"key":"9728_CR63","unstructured":"Ye J, Chen X, Xu N, Zu C, Shao Z, Liu S, Cui Y, Zhou Z, Gong C, Shen Y, Zhou J (2023) A comprehensive capability analysis of gpt-3 and gpt-3.5 series models. arXiv preprint arXiv:2303.10420"},{"key":"9728_CR64","unstructured":"Zeng A, Attarian M, Ichter B, Choromanski K, Wong A, Welker S, Tombari F, Purohit A, Ryoo M, Sindhwani V, Lee J (2022) Socratic models: Composing zero-shot multimodal reasoning with language. arXiv preprint arXiv:2204.00598"},{"key":"9728_CR65","unstructured":"Zhang S, Roller S, Goyal N, Artetxe M, Chen M, Chen S, Dewan C, Diab M, Li X, Lin XV, Mihaylov T (2022) Opt: open pre-trained transformer language models. arXiv preprint arXiv:2205.01068"},{"key":"9728_CR66","unstructured":"Zhang D, Wang D (2015) Relation classification via recurrent neural network. arXiv preprint arXiv:1508.01006"},{"key":"9728_CR67","doi-asserted-by":"crossref","unstructured":"Zhang Y, Zhong V, Chen D, Angeli G, Manning CD (2017) Position-aware attention and supervised data improve slot filling. In: Conference on empirical methods in natural language processing","DOI":"10.18653\/v1\/D17-1004"},{"key":"9728_CR68","unstructured":"Zheng L, Wang C, Kong L (2022) Linear complexity randomized self-attention mechanism. In: International conference on machine learning. PMLR, pp 27011\u201327041"},{"key":"9728_CR69","first-page":"3081","volume":"36","author":"Yang Zhengyuan","year":"2022","unstructured":"Zhengyuan Yang, Zhe Gan, Jianfeng Wang, Xiaowei Hu, Yumao Lu, Zicheng Liu, Lijuan Wang (2022) An empirical study of gpt-3 for few-shot knowledge-based VQA. Proc AAAI Confer Artif Intell 36:3081\u20133089","journal-title":"Proc AAAI Confer Artif Intell"},{"key":"9728_CR70","unstructured":"Zhou D, Scharli N, Hou L, Wei J, Scales N, Wang X, Schuurmans D, Cui C, Bousquet O, Le Q, Chi E (2022) Least-to-most prompting enables complex reasoning in large language models. arXiv preprint arXiv:2205.10625"}],"container-title":["Neural Computing and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-024-09728-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00521-024-09728-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-024-09728-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,8,9]],"date-time":"2024-08-09T18:05:25Z","timestamp":1723226725000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00521-024-09728-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,4,22]]},"references-count":70,"journal-issue":{"issue":"22","published-print":{"date-parts":[[2024,8]]}},"alternative-id":["9728"],"URL":"https:\/\/doi.org\/10.1007\/s00521-024-09728-5","relation":{},"ISSN":["0941-0643","1433-3058"],"issn-type":[{"value":"0941-0643","type":"print"},{"value":"1433-3058","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,4,22]]},"assertion":[{"value":"15 September 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 March 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 April 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"No potential Conflict of interest was reported by the authors.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}