{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T23:46:26Z","timestamp":1740181586465,"version":"3.37.3"},"reference-count":40,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2023,8,3]],"date-time":"2023-08-03T00:00:00Z","timestamp":1691020800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,8,3]],"date-time":"2023-08-03T00:00:00Z","timestamp":1691020800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SN COMPUT. SCI."],"DOI":"10.1007\/s42979-023-02039-x","type":"journal-article","created":{"date-parts":[[2023,8,3]],"date-time":"2023-08-03T14:01:40Z","timestamp":1691071300000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Open-Domain Long-Form Question\u2013Answering Using Transformer-Based Pipeline"],"prefix":"10.1007","volume":"4","author":[{"given":"Aprameya","family":"Dash","sequence":"first","affiliation":[]},{"given":"Mohit","family":"Awachar","sequence":"additional","affiliation":[]},{"given":"Anshul","family":"Patel","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7651-3820","authenticated-orcid":false,"given":"Bhawana","family":"Rudra","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,8,3]]},"reference":[{"key":"2039_CR1","unstructured":"Chen W, Chang M-W, Schlinger E, Wang W, Cohen WW. Open question answering over tables and text. 2020. arXiv preprint arXiv:2010.10439."},{"key":"2039_CR2","doi-asserted-by":"crossref","unstructured":"Krishna K, Roy A, Iyyer M, Hurdles to progress in long-form question answering. 2021. arXiv preprint arXiv:2103.06332.","DOI":"10.18653\/v1\/2021.naacl-main.393"},{"key":"2039_CR3","unstructured":"Nakano R, Hilton J, Balaji S, Wu J, Ouyang L, Kim C, Hesse C, Jain S, Kosaraju V, Saunders W et al. Webgpt: browser-assisted question-answering with human feedback. 2021. arXiv preprint arXiv:2112.09332."},{"key":"2039_CR4","unstructured":"Bui M-Q, Tran V, Nguyen H-T, Le Nguyen M, How state-of-the-art models can deal with long-form question answering. In: Proceedings of the 34th Pacific Asia conference on language, information and computation, 2020; pp. 375\u2013382."},{"key":"2039_CR5","doi-asserted-by":"crossref","unstructured":"Fan A, Jernite Y, Perez E, Grangier D, Weston J, Auli M, Eli5: long form question answering. 2019. arXiv preprint arXiv:1907.09190.","DOI":"10.18653\/v1\/P19-1346"},{"key":"2039_CR6","doi-asserted-by":"crossref","unstructured":"Petroni F, Piktus A, Fan A, Lewis P, Yazdani M, De Cao N, Thorne J, Jernite Y, Karpukhin V, Maillard J et al. Kilt: a benchmark for knowledge intensive language tasks. 2020. arXiv preprint arXiv:2009.02252.","DOI":"10.18653\/v1\/2021.naacl-main.200"},{"key":"2039_CR7","unstructured":"Zhang T, Kishore V, Wu F, Weinberger KQ, Artzi Y. Bertscore: evaluating text generation with bert. 2019. arXiv preprint arXiv:1904.09675."},{"key":"2039_CR8","doi-asserted-by":"crossref","unstructured":"Bahri S, Sumpeno S, Nugroho SMS. An information retrieval approach to finding similar questions in question\u2013answering of Indonesian government e-procurement services using tf* idf and lsi model. In: 2018 10th international conference on information technology and electrical engineering (ICITEE). New York: IEEE; 2018. p. 626\u2013631 .","DOI":"10.1109\/ICITEED.2018.8534856"},{"key":"2039_CR9","doi-asserted-by":"crossref","unstructured":"Huang X, Zhang Y, Wei B, Yao L, A question\u2013answering system over traditional Chinese medicine. In: 2015 IEEE international conference on bioinformatics and biomedicine (BIBM). New York: IEEE; 2015. p. 1737\u20131739.","DOI":"10.1109\/BIBM.2016.7822810"},{"key":"2039_CR10","doi-asserted-by":"crossref","unstructured":"Mohapatra SK, Upadhyay A. Using TF-IDF on Kisan call centre dataset for obtaining query answers. In: 2018 international conference on communication, computing and internet of things (IC3IoT). New York: IEEE; 2018. p. 479\u2013482.","DOI":"10.1109\/IC3IoT.2018.8668134"},{"issue":"8","key":"2039_CR11","doi-asserted-by":"publisher","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","volume":"9","author":"S Hochreiter","year":"1997","unstructured":"Hochreiter S, Schmidhuber J. Long short-term memory. Neural Comput. 1997;9(8):1735\u201380.","journal-title":"Neural Comput"},{"key":"2039_CR12","doi-asserted-by":"crossref","unstructured":"Karimi E, Majidi B, Manzuri MT. Relevant question answering in community based networks using deep lSTM neural networks. In: 2019 7th Iranian joint congress on fuzzy and intelligent systems (CFIS). New York: IEEE; 2019. p. 1\u20135.","DOI":"10.1109\/CFIS.2019.8692168"},{"key":"2039_CR13","doi-asserted-by":"crossref","unstructured":"Chen L, Zeng G, Zhang Q, Chen X, Wu D, Question answering over knowledgebase with attention-based lstm networks and knowledge embeddings. In: 2017 IEEE 16th International Conference on Cognitive Informatics & Cognitive Computing (ICCI* CC). New York: IEEE; 2017. p. 243\u2013246.","DOI":"10.1109\/ICCI-CC.2017.8109757"},{"key":"2039_CR14","doi-asserted-by":"crossref","unstructured":"Luo D, Su J, Yu S, A bert-based approach with relation-aware attention for knowledge base question answering. In: 2020 International Joint Conference on Neural Networks (IJCNN). New York: IEEE; 2020. p. 1\u20138.","DOI":"10.1109\/IJCNN48605.2020.9207186"},{"key":"2039_CR15","unstructured":"Devlin J, Chang M-W, Lee K, Toutanova K. Bert: pre-training of deep bidirectional transformers for language understanding. 2018. arXiv preprint arXiv:1810.04805."},{"key":"2039_CR16","doi-asserted-by":"crossref","unstructured":"Rajpurkar P, Zhang J, Lopyrev K, Liang P. Squad: 100,000+ questions for machine comprehension of text. 2016. arXiv preprint arXiv:1606.05250.","DOI":"10.18653\/v1\/D16-1264"},{"key":"2039_CR17","doi-asserted-by":"crossref","unstructured":"Trischler A, Wang T, Yuan X, Harris J, Sordoni A, Bachman P, Suleman K, Newsqa: a machine comprehension dataset. 2016. arXiv preprint arXiv:1611.09830.","DOI":"10.18653\/v1\/W17-2623"},{"key":"2039_CR18","doi-asserted-by":"publisher","first-page":"317","DOI":"10.1162\/tacl_a_00023","volume":"6","author":"T Ko\u010disk\u1ef3","year":"2018","unstructured":"Ko\u010disk\u1ef3 T, Schwarz J, Blunsom P, Dyer C, Hermann KM, Melis G, Grefenstette E. The narrative QA reading comprehension challenge. Trans Assoc Comput Linguist. 2018;6:317\u201328.","journal-title":"Trans Assoc Comput Linguist"},{"key":"2039_CR19","doi-asserted-by":"publisher","first-page":"249","DOI":"10.1162\/tacl_a_00266","volume":"7","author":"S Reddy","year":"2019","unstructured":"Reddy S, Chen D, Manning CD. COQA: a conversational question answering challenge. Trans Assoc Comput Linguist. 2019;7:249\u201366.","journal-title":"Trans Assoc Comput Linguist"},{"key":"2039_CR20","doi-asserted-by":"crossref","unstructured":"Chen D, Fisch A, Weston J, Bordes A. Reading Wikipedia to answer open-domain questions. 2017. arXiv preprint arXiv:1704.00051.","DOI":"10.18653\/v1\/P17-1171"},{"key":"2039_CR21","doi-asserted-by":"crossref","unstructured":"Tang T, Li J, Zhao WX, Wen J-R. MVP: multi-task supervised pre-training for natural language generation. 2022. arXiv preprint arXiv:2206.12131.","DOI":"10.18653\/v1\/2023.findings-acl.558"},{"key":"2039_CR22","unstructured":"Bhojanapalli S, Chakrabarti A, Veit A, Lukasik M, Jain H, Liu F, Chang Y-W, Kumar S. Leveraging redundancy in attention with reuse transformers. 2021. arXiv preprint arXiv:2110.06821."},{"key":"2039_CR23","doi-asserted-by":"crossref","unstructured":"Riabi A, Scialom T, Keraron R, Sagot B, Seddah D, Staiano J. Synthetic data augmentation for zero-shot cross-lingual question answering. 2020. arXiv preprint arXiv:2010.12643.","DOI":"10.18653\/v1\/2021.emnlp-main.562"},{"key":"2039_CR24","doi-asserted-by":"crossref","unstructured":"Huang Z, Liang D, Xu P, Xiang B. Improve transformer models with better relative position embeddings. 2020. arXiv preprint arXiv:2009.13658.","DOI":"10.18653\/v1\/2020.findings-emnlp.298"},{"key":"2039_CR25","doi-asserted-by":"crossref","unstructured":"Nitish S, Darsini R, Shashank G, Tejas V, Arya A. Bidirectional encoder representation from transformers (bert) variants for procedural long-form answer extraction. In: 2022 12th international conference on cloud computing, data science & engineering (confluence). New York: IEEE; 2022. p. 71\u201376.","DOI":"10.1109\/Confluence52989.2022.9734142"},{"key":"2039_CR26","doi-asserted-by":"crossref","unstructured":"Butler R, Duggirala VD, Banaei-Kashani F. ILFQA: a platform for efficient and accurate long-form question answering. In: Proceedings of the fifteenth ACM international conference on web search and data mining; 2022. p. 1565\u20131568.","DOI":"10.1145\/3488560.3502190"},{"key":"2039_CR27","unstructured":"Guu K, Lee K, Tung Z, Pasupat P, Chang M, Realm: retrieval-augmented language model pre-training. 2020. arXiv preprint arXiv:2002.08909."},{"key":"2039_CR28","doi-asserted-by":"crossref","unstructured":"Karpukhin V, O\u011fuz B, Min S, Lewis P, Wu L, Edunov S, Chen D, Yih W-T. Dense passage retrieval for open-domain question answering. 2020. arXiv preprint arXiv:2004.04906.","DOI":"10.18653\/v1\/2020.emnlp-main.550"},{"key":"2039_CR29","doi-asserted-by":"crossref","unstructured":"Lee K, Chang M-W, Toutanova K. Latent retrieval for weakly supervised open domain question answering. 2019. arXiv preprint arXiv:1906.00300.","DOI":"10.18653\/v1\/P19-1612"},{"key":"2039_CR30","first-page":"9459","volume":"33","author":"P Lewis","year":"2020","unstructured":"Lewis P, Perez E, Piktus A, Petroni F, Karpukhin V, Goyal N, K\u00fcttler H, Lewis M, Yih W-T, Rockt\u00e4schel T. Retrieval-augmented generation for knowledge-intensive NLP tasks. Adv Neural Inf Process Syst. 2020;33:9459\u201374.","journal-title":"Adv Neural Inf Process Syst"},{"key":"2039_CR31","doi-asserted-by":"crossref","unstructured":"Su D, Li X, Zhang J, Shang L, Jiang X, Liu Q, Fung P. Read before generate! Faithful long form question answering with machine reading. 2022. arXiv preprint arXiv:2203.00343.","DOI":"10.18653\/v1\/2022.findings-acl.61"},{"issue":"3","key":"2039_CR32","doi-asserted-by":"publisher","first-page":"535","DOI":"10.1109\/TBDATA.2019.2921572","volume":"7","author":"J Johnson","year":"2019","unstructured":"Johnson J, Douze M, J\u00e9gou H. Billion-scale similarity search with GPUS. IEEE Trans Big Data. 2019;7(3):535\u201347.","journal-title":"IEEE Trans Big Data"},{"key":"2039_CR33","doi-asserted-by":"crossref","unstructured":"Sachan DS, Lewis M, Joshi M, Aghajanyan A, Yih W-T, Pineau J, Zettlemoyer L. Improving passage retrieval with zero-shot question generation. 2022. arXiv preprint arXiv:2204.07496.","DOI":"10.18653\/v1\/2022.emnlp-main.249"},{"key":"2039_CR34","doi-asserted-by":"crossref","unstructured":"Wolf T, Debut L, Sanh V, Chaumond J, Delangue C, Moi A, Cistac P, Rault T, Louf R, Funtowicz M. Transformers: state-of-the-art natural language processing. In: Proceedings of the 2020 conference on empirical methods in natural language processing: system demonstrations; 2020. p. 38\u201345.","DOI":"10.18653\/v1\/2020.emnlp-demos.6"},{"key":"2039_CR35","first-page":"16857","volume":"33","author":"K Song","year":"2020","unstructured":"Song K, Tan X, Qin T, Lu J, Liu T-Y. Mpnet: masked and permuted pre-training for language understanding. Adv Neural Inf Process Syst. 2020;33:16857\u201367.","journal-title":"Adv Neural Inf Process Syst"},{"key":"2039_CR36","unstructured":"Sanh V, Debut L, Chaumond J, Wolf T, Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter. 2019. arXiv preprint arXiv:1910.01108."},{"key":"2039_CR37","unstructured":"Sanh V, Webson A, Raffel C, Bach SH, Sutawika L, Alyafeai Z, Chaffin A, Stiegler A, Scao TL, Raja A et al. Multitask prompted training enables zero-shot task generalization. 2021. arXiv preprint arXiv:2110.08207."},{"key":"2039_CR38","doi-asserted-by":"crossref","unstructured":"Lewis M, Liu Y, Goyal N, Ghazvininejad M, Mohamed A, Levy O, Stoyanov V, Zettlemoyer L, Bart: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. 2019. arXiv preprint arXiv:1910.13461.","DOI":"10.18653\/v1\/2020.acl-main.703"},{"issue":"140","key":"2039_CR39","first-page":"1","volume":"21","author":"C Raffel","year":"2020","unstructured":"Raffel C, Shazeer N, Roberts A, Lee K, Narang S, Matena M, Zhou Y, Li W, Liu PJ. Exploring the limits of transfer learning with a unified text-to-text transformer. J Mach Learn Res. 2020;21(140):1\u201367.","journal-title":"J Mach Learn Res"},{"key":"2039_CR40","unstructured":"Fedus W, Zoph B, Shazeer N. Switch transformers: scaling to trillion parameter models with simple and efficient sparsity. 2021."}],"container-title":["SN Computer Science"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s42979-023-02039-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s42979-023-02039-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s42979-023-02039-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,12,18]],"date-time":"2023-12-18T11:27:22Z","timestamp":1702898842000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s42979-023-02039-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,8,3]]},"references-count":40,"journal-issue":{"issue":"5","published-online":{"date-parts":[[2023,9]]}},"alternative-id":["2039"],"URL":"https:\/\/doi.org\/10.1007\/s42979-023-02039-x","relation":{},"ISSN":["2661-8907"],"issn-type":[{"type":"electronic","value":"2661-8907"}],"subject":[],"published":{"date-parts":[[2023,8,3]]},"assertion":[{"value":"13 March 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 June 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 August 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of Interest"}}],"article-number":"595"}}