{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T12:37:29Z","timestamp":1776083849507,"version":"3.50.1"},"reference-count":69,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2025,6,16]],"date-time":"2025-06-16T00:00:00Z","timestamp":1750032000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2025,6,16]],"date-time":"2025-06-16T00:00:00Z","timestamp":1750032000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"name":"The Alan Turning Institute\/DSO grant"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int. J. Mach. Learn. &amp; Cyber."],"published-print":{"date-parts":[[2025,10]]},"abstract":"<jats:title>Abstract<\/jats:title>\n          <jats:p>Language models (LMs) utilize chain-of-thought (CoT) to imitate human reasoning and inference processes, achieving notable success in multi-hop question answering (QA). Despite this, a disparity remains between the reasoning capabilities of LMs and humans when addressing complex challenges. Psychological research highlights the crucial interplay between explicit content in texts and prior human knowledge during reading. However, current studies have inadequately addressed the relationship between input texts and the pre-training-derived knowledge of LMs from the standpoint of human cognition. In this paper, we propose a <jats:bold>P<\/jats:bold>rompting <jats:bold>E<\/jats:bold>xplicit and <jats:bold>I<\/jats:bold>mplicit knowledge (PEI) framework, which employs CoT prompt-based learning to bridge explicit and implicit knowledge, aligning with human reading comprehension for multi-hop QA. PEI leverages CoT prompts to elicit implicit knowledge from LMs within the input context, while integrating question type information to boost model performance. Moreover, we propose two training paradigms for PEI, and extend our framework on biomedical domain QA to further explore the fusion and relation of explicit and implicit biomedical knowledge via employing biomedical LMs in the Knowledge Prompter to invoke biomedical implicit knowledge and analyze the consistency of the domain knowledge fusion. The experimental results indicate that our proposed PEI performs comparably to the state-of-the-art on HotpotQA, and surpasses baselines on 2WikiMultihopQA and MuSiQue. Additionally, our method achieves a\u00a0significant improvement compared to baselines on MEDHOP. Ablation studies further validate the efficacy of PEI framework in bridging and integrating explicit and implicit knowledge.<\/jats:p>","DOI":"10.1007\/s13042-025-02712-y","type":"journal-article","created":{"date-parts":[[2025,6,16]],"date-time":"2025-06-16T09:30:08Z","timestamp":1750066208000},"page":"8103-8118","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Improving multi-hop question answering with prompting explicit and implicit knowledge aligned human reading comprehension"],"prefix":"10.1007","volume":"16","author":[{"given":"Guangming","family":"Huang","sequence":"first","affiliation":[]},{"given":"Yunfei","family":"Long","sequence":"additional","affiliation":[]},{"given":"Cunjin","family":"Luo","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,6,16]]},"reference":[{"issue":"6","key":"2712_CR1","first-page":"125","volume":"5","author":"NM Abdelaal","year":"2014","unstructured":"Abdelaal NM, Sase AS (2014) Relationship between prior knowledge and reading comprehension. Adv Lang Literary Stud 5(6):125\u2013131","journal-title":"Adv Lang Literary Stud"},{"key":"2712_CR2","doi-asserted-by":"crossref","unstructured":"Baldwin RS, Peleg-Bruckner Z, McClintock AH (1985) Effects of topic interest and prior knowledge on reading comprehension. Read Res Quart pp 497\u2013504","DOI":"10.2307\/747856"},{"key":"2712_CR3","unstructured":"Beltagy I, Peters ME, Cohan A (2020) Longformer: The long-document transformer. arXiv preprint arXiv:2004.05150"},{"key":"2712_CR4","doi-asserted-by":"crossref","unstructured":"Bender EM, Gebru T, McMillan-Major A, et\u00a0al (2021) On the dangers of stochastic parrots: Can language models be too big? In: Proceedings of the 2021 ACM conference on fairness, accountability, and transparency, pp 610\u2013623","DOI":"10.1145\/3442188.3445922"},{"issue":"9","key":"2712_CR5","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3657631","volume":"56","author":"GM Biancofiore","year":"2024","unstructured":"Biancofiore GM, Deldjoo Y, Noia TD et al (2024) Interactive question answering systems: Literature review. ACM Comput Surv 56(9):1\u201338","journal-title":"ACM Comput Surv"},{"key":"2712_CR6","unstructured":"Cao Y, Fang M, Tao D (2019) Bag: Bi-directional attention entity graph convolutional network for multi-hop reasoning question answering. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp 357\u2013362"},{"key":"2712_CR7","unstructured":"Clark K, Luong MT, Le QV, et\u00a0al (2020) Electra: Pre-training text encoders as discriminators rather than generators. arXiv preprint arXiv:2003.10555"},{"issue":"1","key":"2712_CR8","doi-asserted-by":"publisher","first-page":"135","DOI":"10.1111\/j.1467-1770.1977.tb00297.x","volume":"27","author":"MA Clarke","year":"1977","unstructured":"Clarke MA, Silberstein S (1977) Toward a realization of psycholinguistic principles in the esl reading class 1. Lang Learn 27(1):135\u2013154","journal-title":"Lang Learn"},{"key":"2712_CR9","unstructured":"Deng Z, Zhu Y, Chen Y, et\u00a0al (2022) Prompt-based conservation learning for multi-hop question answering. In: Proceedings of the 29th International Conference on Computational Linguistics, pp 1791\u20131800"},{"key":"2712_CR10","unstructured":"Dhingra B, Zaheer M, Balachandran V, et\u00a0al (2020) Differentiable reasoning over a virtual knowledge base. In: International Conference on Learning Representations, https:\/\/openreview.net\/forum?id=SJxstlHFPH"},{"key":"2712_CR11","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2024.125098","volume":"258","author":"Y Du","year":"2024","unstructured":"Du Y, Yan R, Hou Y et al (2024) Adversarial entity graph convolutional networks for multi-hop inference question answering. Expert Syst Appl 258:125098","journal-title":"Expert Syst Appl"},{"key":"2712_CR12","unstructured":"Dubey A, Jauhri A, Pandey A, et\u00a0al (2024) The llama 3 herd of models. arXiv preprint arXiv:2407.21783"},{"key":"2712_CR13","doi-asserted-by":"crossref","unstructured":"Fang Y, Sun S, Gan Z, et\u00a0al (2020) Hierarchical graph network for multi-hop question answering. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp 8823\u20138838","DOI":"10.18653\/v1\/2020.emnlp-main.710"},{"issue":"14","key":"2712_CR14","doi-asserted-by":"publisher","first-page":"3183","DOI":"10.3390\/electronics12143183","volume":"12","author":"P Gao","year":"2023","unstructured":"Gao P, Gao F, Wang P et al (2023) Cluereader: Heterogeneous graph attention network for multi-hop machine reading comprehension. Electronics 12(14):3183","journal-title":"Electronics"},{"key":"2712_CR15","doi-asserted-by":"crossref","unstructured":"Gao P, Gao F, Ni JC, et\u00a0al (2024) Medical knowledge graph question answering for drug-drug interaction prediction based on multi-hop machine reading comprehension. CAAI Trans Intell Technol","DOI":"10.1049\/cit2.12332"},{"issue":"1","key":"2712_CR16","first-page":"1","volume":"3","author":"Y Gu","year":"2021","unstructured":"Gu Y, Tinn R, Cheng H et al (2021) Domain-specific language model pretraining for biomedical natural language processing. ACM Trans Comput Healthcare (HEALTH) 3(1):1\u201323","journal-title":"ACM Trans Comput Healthcare (HEALTH)"},{"issue":"5669","key":"2712_CR17","doi-asserted-by":"publisher","first-page":"438","DOI":"10.1126\/science.1095455","volume":"304","author":"P Hagoort","year":"2004","unstructured":"Hagoort P, Hald L, Bastiaansen M et al (2004) Integration of word meaning and world knowledge in language comprehension. Science 304(5669):438\u2013441","journal-title":"Science"},{"key":"2712_CR18","doi-asserted-by":"crossref","unstructured":"Ho X, Nguyen AKD, Sugawara S, et\u00a0al (2020) Constructing a multi-hop qa dataset for comprehensive evaluation of reasoning steps. In: Proceedings of the 28th International Conference on Computational Linguistics, pp 6609\u20136625","DOI":"10.18653\/v1\/2020.coling-main.580"},{"key":"2712_CR19","unstructured":"Huang G, Long Y, Luo C, et\u00a0al (2023) LIDA: Lexical-based imbalanced data augmentation for content moderation. In: Huang CR, Harada Y, Kim JB, et\u00a0al (eds) Proceedings of the 37th Pacific Asia Conference on Language, Information and Computation. Association for Computational Linguistics, Hong Kong, China, pp 59\u201369, https:\/\/aclanthology.org\/2023.paclic-1.6"},{"key":"2712_CR20","doi-asserted-by":"crossref","unstructured":"Huang G, Li Y, Jameel S, et\u00a0al (2024a) From explainable to interpretable deep learning for natural language processing in healthcare: How far from reality? Comput Struct Biotechnol J","DOI":"10.1016\/j.csbj.2024.05.004"},{"key":"2712_CR21","unstructured":"Huang G, Long Y, Luo C, et\u00a0al (2024b) Prompting explicit and implicit knowledge for multi-hop question answering based on human reading process. In: Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), pp 13179\u201313189"},{"key":"2712_CR22","unstructured":"Huang G, Long Y, Luo C (2025) Similarity-dissimilarity loss for multi-label supervised contrastive learning. https:\/\/arxiv.org\/abs\/2410.13439, arXiv:2410.13439"},{"key":"2712_CR23","doi-asserted-by":"crossref","unstructured":"Jiang Y, Joshi N, Chen YC, et\u00a0al (2019) Explore, propose, and assemble: An interpretable model for multi-hop reading comprehension. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp 2714\u20132725","DOI":"10.18653\/v1\/P19-1261"},{"key":"2712_CR24","unstructured":"Jin C, Rinard M (2023) Evidence of meaning in language models trained on programs. arXiv preprint arXiv:2305.11169"},{"issue":"2","key":"2712_CR25","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3490238","volume":"55","author":"Q Jin","year":"2022","unstructured":"Jin Q, Yuan Z, Xiong G et al (2022) Biomedical question answering: a survey of approaches and challenges. ACM Comput Surv (CSUR) 55(2):1\u201336","journal-title":"ACM Comput Surv (CSUR)"},{"key":"2712_CR26","doi-asserted-by":"crossref","unstructured":"Kanakarajan KR, Kundumani B, Sankarasubbu M (2021) Bioelectra: pretrained biomedical text encoder using discriminators. In: Proceedings of the 20th workshop on biomedical language processing, pp 143\u2013154","DOI":"10.18653\/v1\/2021.bionlp-1.16"},{"key":"2712_CR27","doi-asserted-by":"crossref","unstructured":"Khalifa M, Logeswaran L, Lee M, et\u00a0al (2023) Few-shot reranking for multi-hop qa via language model prompting. In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp 15882\u201315897","DOI":"10.18653\/v1\/2023.acl-long.885"},{"key":"2712_CR28","unstructured":"Kojima T, Gu SS, Reid M, et\u00a0al (2022) Large language models are zero-shot reasoners. arXiv preprint arXiv:2205.11916"},{"key":"2712_CR29","doi-asserted-by":"crossref","unstructured":"Lester B, Al-Rfou R, Constant N (2021) The power of scale for parameter-efficient prompt tuning. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp 3045\u20133059","DOI":"10.18653\/v1\/2021.emnlp-main.243"},{"key":"2712_CR30","doi-asserted-by":"crossref","unstructured":"Lewis M, Liu Y, Goyal N, et\u00a0al (2020) Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp 7871\u20137880","DOI":"10.18653\/v1\/2020.acl-main.703"},{"key":"2712_CR31","doi-asserted-by":"crossref","unstructured":"Li R, Wang L, Wang S, et\u00a0al (2021) Asynchronous multi-grained graph network for interpretable multi-hop reading comprehension. In: IJCAI, pp 3857\u20133863","DOI":"10.24963\/ijcai.2021\/531"},{"key":"2712_CR32","doi-asserted-by":"crossref","unstructured":"Li XL, Liang P (2021) Prefix-tuning: Optimizing continuous prompts for generation. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp 4582\u20134597","DOI":"10.18653\/v1\/2021.acl-long.353"},{"key":"2712_CR33","first-page":"3730","volume":"2024","author":"Z Lin","year":"2024","unstructured":"Lin Z, Chen W, Song Y et al (2024) Prompting few-shot multi-hop question generation via comprehending type-aware semantics. Find Assoc Comput Linguist: NAACL 2024:3730\u20133740","journal-title":"Find Assoc Comput Linguist: NAACL"},{"key":"2712_CR34","doi-asserted-by":"crossref","unstructured":"Liu X, Ji K, Fu Y, et\u00a0al (2021) P-tuning v2: Prompt tuning can be comparable to fine-tuning universally across scales and tasks. arXiv preprint arXiv:2110.07602","DOI":"10.18653\/v1\/2022.acl-short.8"},{"key":"2712_CR35","unstructured":"Loshchilov I, Hutter F (2018) Decoupled weight decay regularization. In: International Conference on Learning Representations"},{"issue":"6","key":"2712_CR36","doi-asserted-by":"publisher","first-page":"409","DOI":"10.1093\/bib\/bbac409","volume":"23","author":"R Luo","year":"2022","unstructured":"Luo R, Sun L, Xia Y et al (2022) Biogpt: generative pre-trained transformer for biomedical text generation and mining. Brief Bioinform 23(6):409","journal-title":"Brief Bioinform"},{"key":"2712_CR37","doi-asserted-by":"crossref","unstructured":"Min S, Zhong V, Zettlemoyer L, et\u00a0al (2019) Multi-hop reading comprehension through question decomposition and rescoring. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp 6097\u20136109","DOI":"10.18653\/v1\/P19-1613"},{"key":"2712_CR38","unstructured":"Minaee S, Mikolov T, Nikzad N, et\u00a0al (2024) Large language models: A survey. arXiv preprint arXiv:2402.06196"},{"key":"2712_CR39","doi-asserted-by":"crossref","unstructured":"Nishida K, Nishida K, Nagata M, et\u00a0al (2019) Answering while summarizing: Multi-task learning for multi-hop qa with evidence extraction. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp 2335\u20132345","DOI":"10.18653\/v1\/P19-1225"},{"key":"2712_CR40","unstructured":"Nori H, Lee YT, Zhang S, et\u00a0al (2023) Can generalist foundation models outcompete special-purpose tuning? case study in medicine. arXiv preprint arXiv:2311.16452"},{"key":"2712_CR41","doi-asserted-by":"crossref","unstructured":"Perez E, Lewis P, Yih Wt, et\u00a0al (2020) Unsupervised question decomposition for question answering. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp 8864\u20138880","DOI":"10.18653\/v1\/2020.emnlp-main.713"},{"key":"2712_CR42","unstructured":"Plaat A, Wong A, Verberne S, et\u00a0al (2024) Reasoning with large language models, a survey. arXiv preprint arXiv:2407.11511"},{"key":"2712_CR43","doi-asserted-by":"crossref","unstructured":"Qiu L, Xiao Y, Qu Y, et\u00a0al (2019) Dynamically fused graph network for multi-hop reasoning. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp 6140\u20136150","DOI":"10.18653\/v1\/P19-1617"},{"key":"2712_CR44","doi-asserted-by":"crossref","unstructured":"Qorib M, Moon G, Ng HT (2024) Are decoder-only language models better than encoder-only language models in understanding word meaning? Findings of the Association for Computational Linguistics ACL 2024:16339\u201316347","DOI":"10.18653\/v1\/2024.findings-acl.967"},{"key":"2712_CR45","doi-asserted-by":"crossref","unstructured":"Rajpurkar P, Zhang J, Lopyrev K, et\u00a0al (2016) Squad: 100,000+ questions for machine comprehension of text. In: Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pp 2383\u20132392","DOI":"10.18653\/v1\/D16-1264"},{"key":"2712_CR46","unstructured":"Seo M, Kembhavi A, Farhadi A, et\u00a0al (2017) Bidirectional attention flow for machine comprehension. In: International Conference on Learning Representations, https:\/\/openreview.net\/forum?id=HJ0UKP9ge"},{"key":"2712_CR47","doi-asserted-by":"crossref","unstructured":"Shao N, Cui Y, Liu T, et\u00a0al (2020) Is graph structure necessary for multi-hop question answering? In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp 7187\u20137192","DOI":"10.18653\/v1\/2020.emnlp-main.583"},{"key":"2712_CR48","unstructured":"Smith F (1971) Understanding reading: A psycholinguistic analysis of reading and learning to read. Holt, Rinehart and Winston, New York"},{"key":"2712_CR49","doi-asserted-by":"crossref","unstructured":"Tang Y, Ng HT, Tung A (2021) Do multi-hop question answering systems know how to answer the single-hop sub-questions? In: Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, pp 3244\u20133249","DOI":"10.18653\/v1\/2021.eacl-main.283"},{"key":"2712_CR50","doi-asserted-by":"publisher","first-page":"539","DOI":"10.1162\/tacl_a_00475","volume":"10","author":"H Trivedi","year":"2022","unstructured":"Trivedi H, Balasubramanian N, Khot T et al (2022) Musique: Multihop questions via single-hop question composition. Trans Assoc Comput Linguist 10:539\u2013554","journal-title":"Trans Assoc Comput Linguist"},{"key":"2712_CR51","doi-asserted-by":"crossref","unstructured":"Trivedi H, Balasubramanian N, Khot T, et\u00a0al (2023) Interleaving retrieval with chain-of-thought reasoning for knowledge-intensive multi-step questions. In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp 10014\u201310037","DOI":"10.18653\/v1\/2023.acl-long.557"},{"key":"2712_CR52","doi-asserted-by":"crossref","unstructured":"Tu M, Huang K, Wang G, et\u00a0al (2020) Select, answer and explain: Interpretable multi-hop reading comprehension over multiple documents. In: Proceedings of the AAAI conference on artificial intelligence, pp 9073\u20139080","DOI":"10.1609\/aaai.v34i05.6441"},{"key":"2712_CR53","doi-asserted-by":"crossref","unstructured":"Wang B, Deng X, Sun H (2022) Iteratively prompt pre-trained language models for chain of thought. In: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp 2714\u20132730","DOI":"10.18653\/v1\/2022.emnlp-main.174"},{"key":"2712_CR54","doi-asserted-by":"crossref","unstructured":"Wang W, Pan S (2022) Deep inductive logic reasoning for multi-hop reading comprehension. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp 4999\u20135009","DOI":"10.18653\/v1\/2022.acl-long.343"},{"key":"2712_CR55","doi-asserted-by":"crossref","unstructured":"Weber L, Minervini P, M\u00fcnchmeyer J, et\u00a0al (2019) Nlprolog: Reasoning with weak unification for question answering in natural language. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp 6151\u20136161","DOI":"10.18653\/v1\/P19-1618"},{"key":"2712_CR56","unstructured":"Wei J, Wang X, Schuurmans D, et\u00a0al (2022) Chain of thought prompting elicits reasoning in large language models. arXiv preprint arXiv:2201.11903"},{"key":"2712_CR57","doi-asserted-by":"crossref","unstructured":"Weissenborn D, Wiese G, Seiffe L (2017) Making neural qa as simple as possible but not simpler. In: Proceedings of the 21st Conference on Computational Natural Language Learning (CoNLL 2017), pp 271\u2013280","DOI":"10.18653\/v1\/K17-1028"},{"key":"2712_CR58","doi-asserted-by":"publisher","first-page":"287","DOI":"10.1162\/tacl_a_00021","volume":"6","author":"J Welbl","year":"2018","unstructured":"Welbl J, Stenetorp P, Riedel S (2018) Constructing datasets for multi-hop reading comprehension across documents. Trans Assoc Comput Linguist 6:287\u2013302","journal-title":"Trans Assoc Comput Linguist"},{"key":"2712_CR59","doi-asserted-by":"crossref","unstructured":"Wolf T, Debut L, Sanh V, et\u00a0al (2020) Transformers: State-of-the-art natural language processing. In: Proceedings of the 2020 conference on empirical methods in natural language processing: system demonstrations, pp 38\u201345","DOI":"10.18653\/v1\/2020.emnlp-demos.6"},{"key":"2712_CR60","unstructured":"Wu B, Zhang Z, Zhao H (2021) Graph-free multi-hop reading comprehension: A select-to-guide strategy. arXiv preprint arXiv:2107.11823"},{"key":"2712_CR61","unstructured":"Wu J, Yang L, Okumura M, et\u00a0al (2024a) Cofca: A step-wise counterfactual multi-hop qa benchmark. arXiv preprint arXiv:2402.11924"},{"key":"2712_CR62","unstructured":"Wu J, Yu T, Wang R, et\u00a0al (2024b) Infoprompt: Information-theoretic soft prompt tuning for natural language understanding. Advances in Neural Information Processing Systems 36"},{"key":"2712_CR63","doi-asserted-by":"crossref","unstructured":"Yang Z, Qi P, Zhang S, et\u00a0al (2018) Hotpotqa: A dataset for diverse, explainable multi-hop question answering. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp 2369\u20132380","DOI":"10.18653\/v1\/D18-1259"},{"key":"2712_CR64","doi-asserted-by":"crossref","unstructured":"Yin KM (1985) The role of prior knowledge in reading comprehension","DOI":"10.64152\/10125\/66995"},{"key":"2712_CR65","doi-asserted-by":"crossref","unstructured":"Yuan H, Yuan Z, Gan R, et\u00a0al (2022) Biobart: Pretraining and evaluation of a biomedical generative language model. In: Proceedings of the 21st Workshop on Biomedical Language Processing, pp 97\u2013109","DOI":"10.18653\/v1\/2022.bionlp-1.9"},{"key":"2712_CR66","doi-asserted-by":"crossref","unstructured":"Zhang J, Zhang H, Zhang D, et\u00a0al (2024) End-to-end beam retrieval for multi-hop question answering. In: Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), pp 1718\u20131731","DOI":"10.18653\/v1\/2024.naacl-long.96"},{"key":"2712_CR67","unstructured":"Zhang Z, Zhang A, Li M, et\u00a0al (2023) Automatic chain of thought prompting in large language models. In: The Eleventh International Conference on Learning Representations, https:\/\/openreview.net\/forum?id=5NTt8GFjUHkr"},{"key":"2712_CR68","doi-asserted-by":"crossref","unstructured":"Zhu A, Hwang A, Dugan L, et\u00a0al (2024) Fanoutqa: A multi-hop, multi-document question answering benchmark for large language models. In: Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pp 18\u201337","DOI":"10.18653\/v1\/2024.acl-short.2"},{"key":"2712_CR69","unstructured":"Zweigenbaum P (2003) Question answering in biomedicine. Proceedings Workshop on Natural Language Processing for Question Answering. EACL, Citeseer, pp 1\u20134"}],"container-title":["International Journal of Machine Learning and Cybernetics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13042-025-02712-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s13042-025-02712-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13042-025-02712-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,15]],"date-time":"2025-10-15T16:59:52Z","timestamp":1760547592000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s13042-025-02712-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,16]]},"references-count":69,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2025,10]]}},"alternative-id":["2712"],"URL":"https:\/\/doi.org\/10.1007\/s13042-025-02712-y","relation":{},"ISSN":["1868-8071","1868-808X"],"issn-type":[{"value":"1868-8071","type":"print"},{"value":"1868-808X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,6,16]]},"assertion":[{"value":"9 January 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 June 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 June 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Non-financial interests: none.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}