{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,13]],"date-time":"2026-03-13T08:59:19Z","timestamp":1773392359476,"version":"3.50.1"},"publisher-location":"Cham","reference-count":42,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031488573","type":"print"},{"value":"9783031488580","type":"electronic"}],"license":[{"start":{"date-parts":[[2023,12,20]],"date-time":"2023-12-20T00:00:00Z","timestamp":1703030400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,12,20]],"date-time":"2023-12-20T00:00:00Z","timestamp":1703030400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-48858-0_25","type":"book-chapter","created":{"date-parts":[[2023,12,19]],"date-time":"2023-12-19T06:02:43Z","timestamp":1702965763000},"page":"313-323","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["Leveraging Large Language Models for Literature Review Tasks - A Case Study Using ChatGPT"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1819-6542","authenticated-orcid":false,"given":"Robert","family":"Zimmermann","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0004-3807-5734","authenticated-orcid":false,"given":"Marina","family":"Staab","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3562-2240","authenticated-orcid":false,"given":"Mehran","family":"Nasseri","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0011-3502","authenticated-orcid":false,"given":"Patrick","family":"Brandtner","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,20]]},"reference":[{"key":"25_CR1","unstructured":"vom Brocke, J., et al.: Reconstructing the giant: on the importance of rigour in documenting the literature search process. In: ECIS 2009 Proceedings (2009)"},{"key":"25_CR2","unstructured":"Jozefowicz, R., Vinyals, O., Schuster, M., Shazeer, N., Wu, Y.: Exploring the limits of language modeling. arXiv (2016)"},{"key":"25_CR3","unstructured":"Uszkoreit, J.: Transformer: A Novel Neural Network Architecture for Language Under-standing \u2013 Google AI Blog (2017). https:\/\/ai.googleblog.com\/2017\/08\/transformer-novel-neural-network.html"},{"key":"25_CR4","unstructured":"Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I.: Language models are unsupervised multitask learners, 1\u20139 (2019)"},{"key":"25_CR5","first-page":"27730","volume":"35","author":"L Ouyang","year":"2022","unstructured":"Ouyang, L., et al.: Training language models to follow instructions with human feedback. Adv. Neural. Inf. Process. Syst. 35, 27730\u201327744 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"25_CR6","unstructured":"Zhang, S., et al.: OPT: open pre-trained transformer language models. arXiv"},{"key":"25_CR7","doi-asserted-by":"crossref","unstructured":"Chakrabarty, T., Padmakumar, V., He, H.: Help me write a poem: instruction tuning as a vehicle for collaborative poetry writing. In: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp. 6848\u20136863 (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.460"},{"key":"25_CR8","unstructured":"Weidinger, L., et al.: Ethical and social risks of harm from Language Models (2021)"},{"key":"25_CR9","doi-asserted-by":"publisher","unstructured":"Weidinger, L., et al.: Taxonomy of risks posed by language models. In: 2022 ACM Conference on Fairness, Accountability, and Transparency, New York, NY, USA, pp. 214\u2013229. ACM (2022). https:\/\/doi.org\/10.1145\/3531146.3533088","DOI":"10.1145\/3531146.3533088"},{"key":"25_CR10","unstructured":"OpenAI: Introducing ChatGPT (2023). https:\/\/openai.com\/blog\/chatgpt"},{"key":"25_CR11","doi-asserted-by":"publisher","unstructured":"Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 4171\u20134186 (2019). https:\/\/doi.org\/10.18653\/v1\/N19-1423","DOI":"10.18653\/v1\/N19-1423"},{"key":"25_CR12","doi-asserted-by":"publisher","DOI":"10.1016\/j.frl.2022.103617","volume":"53","author":"M Leippold","year":"2023","unstructured":"Leippold, M.: Thus spoke GPT-3: interviewing a large-language model on climate finance. Financ. Res. Lett. 53, 103617 (2023). https:\/\/doi.org\/10.1016\/j.frl.2022.103617","journal-title":"Financ. Res. Lett."},{"key":"25_CR13","unstructured":"Brown, T.B., et al.: Language Models are Few-Shot Learners (2020)"},{"key":"25_CR14","doi-asserted-by":"publisher","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998\u20136008 (2017). https:\/\/doi.org\/10.48550\/arXiv.1706.03762","DOI":"10.48550\/arXiv.1706.03762"},{"key":"25_CR15","unstructured":"Liu, Y., et al.: RoBERTa: a robustly optimized BERT pretraining approach. arXiv (2019)"},{"key":"25_CR16","doi-asserted-by":"publisher","unstructured":"Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 2020, 5485\u20135551 (2020). https:\/\/doi.org\/10.5555\/3455716.3455856","DOI":"10.5555\/3455716.3455856"},{"key":"25_CR17","unstructured":"OpenAI: GPT-4 Technical Report (2023)"},{"key":"25_CR18","unstructured":"Kojima, T., Gu, S.S., Reid, M., Matsuo, Y., Iwasawa, Y.: Large Language Models are Zero-Shot Reasoners (2022)"},{"key":"25_CR19","doi-asserted-by":"publisher","unstructured":"Sn\u00e6bjarnarson, V., Einarsson, H.: Cross-Lingual QA as a Stepping Stone for Monolingual Open QA in Icelandic. Proceedings of the Workshop on Multilingual Information Access (MIA), vol. , 29\u201336 (2022). doi: https:\/\/doi.org\/10.18653\/v1\/2022.mia-1.4","DOI":"10.18653\/v1\/2022.mia-1.4"},{"key":"25_CR20","unstructured":"Gao, T., Xia, L.,Yu, D. (eds.): Fine-tuning pre-trained language model with multi-level adaptive learning rates for answer selection, vol. (2019)"},{"key":"25_CR21","unstructured":"DeRosa, D.M., Lepsinger, R.: Virtual Team Success: A Practical Guide for Working and Learning from Distance. Wiley (2010)"},{"key":"25_CR22","unstructured":"Hosseini-Asl, E., Asadi, S., Asemi, A., Lavangani, M.A.Z.: Neural text generation for idea generation: the case of brainstorming. Int. J. Hum.-Comput. Stud. 151 (2021)"},{"key":"25_CR23","unstructured":"Palomaki, J., Kytola, A., Vatanen, T.: Collaborative idea generation with a language model. In: Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, vol. 1\u201312 (2021)"},{"key":"25_CR24","unstructured":"Valvoda, J., Fang, Y., Vandyke, D.: Prompting for a conversation: how to control a dialog model? In: Proceedings of the Second Workshop on When Creative AI Meets Conversational AI, pp. 1\u20138 (2022)"},{"key":"25_CR25","unstructured":"Zeng, Y., Nie, J.-Y.: Open-Domain Dialogue Generation Based on Pre-trained Language Models"},{"key":"25_CR26","unstructured":"Li, D., You, J., Funakoshi, K., Okumura, M.: A-TIP: attribute-aware text Infilling via Pre-trained language model. In: Proceedings of the 29th International Conference on Computational Linguistics, pp. 5857\u20135869 (2022)"},{"key":"25_CR27","doi-asserted-by":"publisher","unstructured":"Rahali, A., Akhloufi, M.A.: End-to-End transformer-based models in textual-based NLP. AI 4, 54\u2013110 (2023). https:\/\/doi.org\/10.3390\/ai4010004","DOI":"10.3390\/ai4010004"},{"key":"25_CR28","doi-asserted-by":"publisher","unstructured":"Ziegler, D.M., et al.: Fine-tuning language models from human preferences (2020). https:\/\/doi.org\/10.48550\/arXiv.1909.08593","DOI":"10.48550\/arXiv.1909.08593"},{"key":"25_CR29","doi-asserted-by":"publisher","first-page":"10840","DOI":"10.1609\/aaai.v36i10.21330","volume":"36","author":"X Jiang","year":"2022","unstructured":"Jiang, X., Liang, Y., Chen, W., Duan, N.: XLM-K: improving cross-lingual language model pre-training with multilingual knowledge. AAAI 36, 10840\u201310848 (2022). https:\/\/doi.org\/10.1609\/aaai.v36i10.21330","journal-title":"AAAI"},{"key":"25_CR30","unstructured":"Dunn, A., et al.: Structured information extraction from complex scientific text with fine-tuned large language models (2022)"},{"key":"25_CR31","doi-asserted-by":"crossref","unstructured":"Wu, T., Shiri, F., Kang, J., Qi, G., Haffari, G., Li, Y.-F.: KC-GEE: knowledge-based conditioning for generative event extraction (2022)","DOI":"10.21203\/rs.3.rs-2190758\/v1"},{"key":"25_CR32","unstructured":"Santosh, T.Y.S.S., Chakraborty, P., Dutta, S., Sanyal, D.K., Das, P.P.: Joint entity and relation extraction from scientific documents: role of linguistic information and entity types (2021). https:\/\/ceur-ws.org\/Vol-3004\/paper2.pdf"},{"key":"25_CR33","doi-asserted-by":"publisher","first-page":"5113","DOI":"10.1007\/s11192-021-03948-5","volume":"126","author":"VK Singh","year":"2021","unstructured":"Singh, V.K., Singh, P., Karmakar, M., Leta, J., Mayr, P.: The journal coverage of Web of science, Scopus and dimensions: a comparative analysis. Scientometrics 126, 5113\u20135142 (2021). https:\/\/doi.org\/10.1007\/s11192-021-03948-5","journal-title":"Scientometrics"},{"key":"25_CR34","doi-asserted-by":"publisher","unstructured":"Haman, M., \u0160koln\u00edk, M.: Using ChatGPT to conduct a literature review. Accountab. Res. 1\u20133 (2023). https:\/\/doi.org\/10.1080\/08989621.2023.2185514","DOI":"10.1080\/08989621.2023.2185514"},{"key":"25_CR35","doi-asserted-by":"publisher","DOI":"10.7759\/cureus.37281","volume":"15","author":"O Temsah","year":"2023","unstructured":"Temsah, O., et al.: Overview of early ChatGPT\u2019s presence in medical literature: insights from a hybrid literature review by ChatGPT and human experts. Cureus 15, e37281 (2023). https:\/\/doi.org\/10.7759\/cureus.37281","journal-title":"Cureus"},{"key":"25_CR36","doi-asserted-by":"publisher","unstructured":"Rahman, M., Terano, H.J.R., Rahman, N., Salamzadeh, A., Rahaman, S.: ChatGPT and academic research: a review and recommendations based on practical examples. J. Educ. Mngt. Dev. Stud. 3, 1\u201312 (2023). https:\/\/doi.org\/10.52631\/jemds.v3i1.175","DOI":"10.52631\/jemds.v3i1.175"},{"key":"25_CR37","doi-asserted-by":"publisher","unstructured":"Gupta, R., et al.: Expanding cosmetic plastic surgery research using ChatGPT. Aesthetic Surgery J. (2023). https:\/\/doi.org\/10.1093\/asj\/sjad069","DOI":"10.1093\/asj\/sjad069"},{"key":"25_CR38","unstructured":"Ouyang, L., et al.: Training language models to follow instructions with human feedback"},{"key":"25_CR39","unstructured":"OpenAI: Best practices for prompt engineering with OpenAI API (2023). https:\/\/help.openai.com\/en\/articles\/6654000-best-practices-for-prompt-engineering-with-openai-api"},{"key":"25_CR40","unstructured":"OpenAI: Models (2023). https:\/\/platform.openai.com\/docs\/models\/overview"},{"key":"25_CR41","unstructured":"BigScience Workshop: BLOOM. Hugging Face (2022)"},{"key":"25_CR42","unstructured":"Touvron, H., et al.: LLaMA: Open and Efficient Foundation Language Models (2023)"}],"container-title":["Communications in Computer and Information Science","Advanced Research in Technologies, Information, Innovation and Sustainability"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-48858-0_25","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,12,19]],"date-time":"2023-12-19T06:12:16Z","timestamp":1702966336000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-48858-0_25"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,20]]},"ISBN":["9783031488573","9783031488580"],"references-count":42,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-48858-0_25","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"value":"1865-0929","type":"print"},{"value":"1865-0937","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,12,20]]},"assertion":[{"value":"20 December 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ARTIIS","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Advanced Research in Technologies, Information, Innovation and Sustainability","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Madrid","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Spain","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"artiis2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/artiis.org","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}