{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T11:18:58Z","timestamp":1772104738466,"version":"3.50.1"},"reference-count":50,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T00:00:00Z","timestamp":1772064000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T00:00:00Z","timestamp":1772064000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-026-21394-4","type":"journal-article","created":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T10:34:23Z","timestamp":1772102063000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Enhanced image captioning with positional and dual attention using deep convolutional long short-term memory and emotional feedback mechanism"],"prefix":"10.1007","volume":"85","author":[{"given":"A.","family":"Ranjith Gnana Suthakar","sequence":"first","affiliation":[]},{"given":"B. J.","family":"Sandesh","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,2,26]]},"reference":[{"key":"21394_CR1","unstructured":"Dataset: https:\/\/www.kaggle.com\/datasets\/awsaf49\/coco-2017-dataset"},{"key":"21394_CR2","unstructured":"https:\/\/www.kaggle.com\/datasets\/adityajn105\/flickr30k"},{"key":"21394_CR3","unstructured":"https:\/\/huggingface.co\/datasets\/lmms-lab\/VizWiz-Caps\/viewer\/default\/test?p=1"},{"key":"21394_CR4","unstructured":"https:\/\/www.kaggle.com\/datasets\/prathamsaraf1389\/senticap"},{"issue":"15","key":"21394_CR5","doi-asserted-by":"publisher","first-page":"1","DOI":"10.3390\/app12157724","volume":"12","author":"M Al Duhayyim","year":"2022","unstructured":"Al Duhayyim M, Alazwari S, Mengash HA, Marzouk R, Alzahrani JS, Mahgoub H, Althukair F, Salama AS (2022) Metaheuristics optimization with deep learning enabled automated image captioning system. Appl Sci (Switz) 12(15):1\u201318. https:\/\/doi.org\/10.3390\/app12157724","journal-title":"Appl Sci (Switz)"},{"key":"21394_CR6","first-page":"23716","volume":"35","author":"JB Alayrac","year":"2022","unstructured":"Alayrac JB, Donahue J, Luc P, Miech A, Barr I, Hasson Y, \u2026 Simonyan K (2022) Flamingo: a visual language model for few-shot learning. Adv Neural Inf Proces Syst 35:23716\u201323736 https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2022\/file\/960a172bc7fbf0177ccccbb411a7d800-Paper-Conference.pdf","journal-title":"Adv Neural Inf Proces Syst"},{"key":"21394_CR7","doi-asserted-by":"publisher","first-page":"1","DOI":"10.5244\/C.27.13","volume-title":"British Machine Vision Conference (BMVC)","author":"PF Alcantarilla","year":"2013","unstructured":"Alcantarilla PF, Nuevo J, Bartoli A (2013) Fast explicit diffusion for accelerated features in nonlinear scale spaces. In: British Machine Vision Conference (BMVC), pp 1\u201312. https:\/\/doi.org\/10.5244\/C.27.13"},{"issue":"4","key":"21394_CR8","doi-asserted-by":"publisher","first-page":"1","DOI":"10.3233\/JIFS-189415","volume":"40","author":"JA Alzubi","year":"2021","unstructured":"Alzubi JA, Jain R, Nagrath P, Satapathy S, Taneja S, Gupta P (2021) Deep image captioning using an ensemble of CNN and LSTM based deep neural networks. J Intell Fuzzy Syst 40(4):1\u20139. https:\/\/doi.org\/10.3233\/JIFS-189415","journal-title":"J Intell Fuzzy Syst"},{"key":"21394_CR9","doi-asserted-by":"publisher","first-page":"218386","DOI":"10.1109\/ACCESS.2020.3042484","volume":"8","author":"S Amirian","year":"2020","unstructured":"Amirian S, Rasheed K, Taha TR, Arabnia HR (2020) Automatic image and video caption generation with deep learning: a concise review and algorithmic overlap. IEEE Access 8:218386\u2013218400. https:\/\/doi.org\/10.1109\/ACCESS.2020.3042484","journal-title":"IEEE Access"},{"key":"21394_CR10","doi-asserted-by":"publisher","first-page":"104633","DOI":"10.1109\/ACCESS.2023.3317276","volume":"11","author":"MA Arasi","year":"2023","unstructured":"Arasi MA, Alshahrani HM, Alruwais N, Motwakel A, Ahmed NA, Mohamed A (2023) Automated image captioning using sparrow search algorithm with improved deep learning model. IEEE Access 11:104633\u2013104642. https:\/\/doi.org\/10.1109\/ACCESS.2023.3317276","journal-title":"IEEE Access"},{"issue":"12","key":"21394_CR11","doi-asserted-by":"publisher","first-page":"1","DOI":"10.3390\/sym14122681","volume":"14","author":"S Ayoub","year":"2022","unstructured":"Ayoub S, Gulzar Y, Reegu FA, Turaev S (2022) Generating image captions using Bahdanau attention mechanism and transfer learning. Symmetry 14(12):1\u201319. https:\/\/doi.org\/10.3390\/sym14122681","journal-title":"Symmetry"},{"key":"21394_CR12","doi-asserted-by":"publisher","DOI":"10.1016\/j.heliyon.2024.e36272","author":"A Bhuiyan","year":"2024","unstructured":"Bhuiyan A, Hossain E, Hoque MM, Dewan MAA (2024) Enhancing image caption generation through context-aware attention mechanism. Heliyon. https:\/\/doi.org\/10.1016\/j.heliyon.2024.e36272","journal-title":"Heliyon"},{"key":"21394_CR13","doi-asserted-by":"publisher","first-page":"33679","DOI":"10.1109\/ACCESS.2022.3161428","volume":"10","author":"R Castro","year":"2022","unstructured":"Castro R, Pineda I, Lim W, Morocho-Cayamcela ME (2022) Deep learning approaches based on transformer architectures for image captioning tasks. IEEE Access 10:33679\u201333694. https:\/\/doi.org\/10.1109\/ACCESS.2022.3161428","journal-title":"IEEE Access"},{"key":"21394_CR14","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2305.18565","author":"X Chen","year":"2023","unstructured":"Chen X, Djolonga J, Padlewski P, Mustafa B, Changpinyo S, Wu J, Ruiz CR, Goodman S, Wang X, Tay Y, Shakeri S, Dehghani M, Salz D, Lucic M, Tschannen M, Nagrani A, Hu H, Joshi M, Pang B, Seyedhosseini M (2023) PaLI-X\u202f: on scaling up a multilingual vision and language model. Arxiv. https:\/\/doi.org\/10.48550\/arXiv.2305.18565","journal-title":"Arxiv"},{"issue":"5","key":"21394_CR15","doi-asserted-by":"publisher","first-page":"278","DOI":"10.14569\/IJACSA.2020.0110537","volume":"11","author":"M Chohan","year":"2020","unstructured":"Chohan M, Khan A, Mahar MS, Hassan S, Ghafoor A, Khan M (2020) Image captioning using deep learning: a systematic literature review. Int J Adv Comput Sci Appl 11(5):278\u2013286. https:\/\/doi.org\/10.14569\/IJACSA.2020.0110537","journal-title":"Int J Adv Comput Sci Appl"},{"key":"21394_CR16","doi-asserted-by":"publisher","DOI":"10.1155\/2020\/8909458","author":"Y Chu","year":"2020","unstructured":"Chu Y, Yue X, Yu L, Sergei M, Wang Z (2020) Automatic image captioning based on ResNet50 and LSTM with soft attention. Wirel Commun Mob Comput. https:\/\/doi.org\/10.1155\/2020\/8909458","journal-title":"Wirel Commun Mob Comput"},{"key":"21394_CR17","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2406.01956","author":"Z Ding","year":"2024","unstructured":"Ding Z, Li P, Yang Q, Li S (2024) Enhance image-to-image generation with LLaVA prompt and negative prompt. Arxiv. https:\/\/doi.org\/10.48550\/arXiv.2406.01956","journal-title":"Arxiv"},{"key":"21394_CR18","unstructured":"Dufera, A. T. (2025). Pre-trained CNN architecture for transformer-based image caption generation model. arXiv preprint arXiv:2509.17365. https:\/\/arxiv.org\/pdf\/2509.17365"},{"issue":"1","key":"21394_CR19","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1088\/1742-6596\/1712\/1\/012015","volume":"1712","author":"G Geetha","year":"2020","unstructured":"Geetha G, Kirthigadevi T, Ponsam GG, Karthik T, Safa M (2020) Image captioning using deep convolutional neural networks (CNNs). J Phys Conf Ser 1712(1):1\u201313. https:\/\/doi.org\/10.1088\/1742-6596\/1712\/1\/012015","journal-title":"J Phys Conf Ser"},{"issue":"24","key":"21394_CR20","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s00521-019-04515-z","volume":"32","author":"N Gupta","year":"2020","unstructured":"Gupta N, Jalal AS (2020) Integration of textual cues for fine-grained image captioning using deep CNN and LSTM. Neural Comput Appl 32(24):1\u201310. https:\/\/doi.org\/10.1007\/s00521-019-04515-z","journal-title":"Neural Comput Appl"},{"issue":"6","key":"21394_CR21","doi-asserted-by":"publisher","first-page":"1","DOI":"10.3390\/s24061796","volume":"24","author":"DA Hafeth","year":"2024","unstructured":"Hafeth DA, Kollias S (2024) Insights into object semantics: leveraging transformer networks for advanced image captioning. Sensors 24(6):1\u201318. https:\/\/doi.org\/10.3390\/s24061796","journal-title":"Sensors"},{"key":"21394_CR22","doi-asserted-by":"publisher","first-page":"64918","DOI":"10.1109\/ACCESS.2021.3075579","volume":"9","author":"MZ Hossain","year":"2021","unstructured":"Hossain MZ, Sohel F, Shiratuddin MF, Laga H, Bennamoun M (2021) Text to image synthesis for improved image captioning. IEEE Access 9:64918\u201364928. https:\/\/doi.org\/10.1109\/ACCESS.2021.3075579","journal-title":"IEEE Access"},{"issue":"12","key":"21394_CR23","doi-asserted-by":"publisher","first-page":"2313","DOI":"10.1007\/s10994-020-05919-y","volume":"109","author":"F Huang","year":"2020","unstructured":"Huang F, Li Z, Wei H, Zhang C, Ma H (2020) Boost image captioning with knowledge reasoning. Mach Learn 109(12):2313\u20132332. https:\/\/doi.org\/10.1007\/s10994-020-05919-y","journal-title":"Mach Learn"},{"key":"21394_CR24","doi-asserted-by":"publisher","first-page":"822","DOI":"10.1109\/DASA51403.2020.9317108","volume-title":"International conference on decision aid sciences and application","author":"AH Kamal","year":"2020","unstructured":"Kamal AH, Jishan MA, Mansoor N (2020) TextMage: the automated Bangla caption generator based on deep learning. In: International conference on decision aid sciences and application, pp 822\u2013826. https:\/\/doi.org\/10.1109\/DASA51403.2020.9317108"},{"issue":"2","key":"21394_CR25","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s00034-019-01306-8","volume":"39","author":"A Khamparia","year":"2020","unstructured":"Khamparia A, Pandey B, Tiwari S, Gupta D, Khanna A, Rodrigues JJPC (2020) An integrated hybrid CNN\u2013RNN model for visual description and generation of captions. Circ Syst Signal Process 39(2):1\u201313. https:\/\/doi.org\/10.1007\/s00034-019-01306-8","journal-title":"Circ Syst Signal Process"},{"key":"21394_CR26","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1905.10296","author":"F Kraus","year":"2020","unstructured":"Kraus F, Dietmayer K (2020) Uncertainty estimation in one-stage object detection. Arxiv. https:\/\/doi.org\/10.48550\/arXiv.1905.10296","journal-title":"Arxiv"},{"key":"21394_CR27","doi-asserted-by":"publisher","unstructured":"Li C, Harrison B (2021) 3M: multi-style image caption generation using multi-modality features under multi-UPDOWN model. arXiv preprint arXiv:2103.11186. https:\/\/doi.org\/10.48550\/arXiv.2103.11186","DOI":"10.48550\/arXiv.2103.11186"},{"key":"21394_CR28","first-page":"19730","volume-title":"International conference on machine learning","author":"J Li","year":"2023","unstructured":"Li J, Li D, Savarese S, Hoi S (2023) Blip-2: bootstrapping language-image pre-training with frozen image encoders and large language models. In: International conference on machine learning, pp 19730\u201319742 https:\/\/proceedings.mlr.press\/v202\/li23q\/li23q.pdf"},{"key":"21394_CR29","first-page":"121","volume-title":"European conference on computer vision","author":"X Li","year":"2020","unstructured":"Li X, Yin X, Li C, Zhang P, Hu X, Zhang L, \u2026 Gao J (2020) Oscar: Object-semantics aligned pre-training for vision-language tasks. In: European conference on computer vision, pp 121\u2013137 https:\/\/arxiv.org\/pdf\/2004.06165"},{"key":"21394_CR30","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.autcon.2020.103334","volume":"119","author":"H Liu","year":"2020","unstructured":"Liu H, Wang G, Huang T, He P, Skitmore M, Luo X (2020) Manifesting construction activity scenes via image captioning. Autom Constr 119:1\u201319. https:\/\/doi.org\/10.1016\/j.autcon.2020.103334","journal-title":"Autom Constr"},{"issue":"2","key":"21394_CR31","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3432246","volume":"20","author":"SK Mishra","year":"2021","unstructured":"Mishra SK, Dhir R, Saha S, Bhattacharyya P (2021) A Hindi image caption generation framework using deep learning. ACM Trans Asian Low-Resour Lang Inf Process 20(2):1\u201319. https:\/\/doi.org\/10.1145\/3432246","journal-title":"ACM Trans Asian Low-Resour Lang Inf Process"},{"issue":"3","key":"21394_CR32","doi-asserted-by":"publisher","first-page":"1","DOI":"10.3390\/math10030288","volume":"10","author":"M Omri","year":"2022","unstructured":"Omri M, Abdel-Khalek S, Khalil EM, Bouslimi J, Joshi GP (2022) Modeling of Hyperparameter tuned deep learning model for automated image captioning. Mathematics 10(3):1\u201320. https:\/\/doi.org\/10.3390\/math10030288","journal-title":"Mathematics"},{"issue":"1","key":"21394_CR33","doi-asserted-by":"publisher","first-page":"20762","DOI":"10.1038\/s41598-024-69664-1","volume":"14","author":"AA Osman","year":"2024","unstructured":"Osman AA, Shalaby MAW, Soliman MM, Elsayed KM (2024) Novel concept-based image captioning models using LSTM and multi-encoder transformer architecture. Sci Rep 14(1):20762. https:\/\/doi.org\/10.1038\/s41598-024-69664-1","journal-title":"Sci Rep"},{"key":"21394_CR34","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2311.18799","author":"A Panagopoulou","year":"2024","unstructured":"Panagopoulou A, Xue L, Yu N, Li J, Li D, Joty S, Xu R, Savarese S, Xiong C, Niebles JC (2024) X-InstructBLIP: a framework for aligning image, 3D, audio, video to LLMs and its emergent cross-modal reasoning. Arxiv. https:\/\/doi.org\/10.48550\/arXiv.2311.18799","journal-title":"Arxiv"},{"key":"21394_CR35","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2023.119774","volume":"223","author":"H Parvin","year":"2023","unstructured":"Parvin H, Naghsh-Nilchi AR, Mohammadi HM (2023) Transformer-based local-global guidance for image captioning. Expert Syst Appl 223:119774. https:\/\/doi.org\/10.1016\/j.eswa.2023.119774","journal-title":"Expert Syst Appl"},{"key":"21394_CR36","doi-asserted-by":"publisher","first-page":"686","DOI":"10.1016\/j.procs.2023.01.049","volume":"218","author":"AK Poddar","year":"2022","unstructured":"Poddar AK, Rani R (2022) Hybrid architecture using CNN and LSTM for image captioning in Hindi language. Procedia Comput Sci 218:686\u2013696. https:\/\/doi.org\/10.1016\/j.procs.2023.01.049","journal-title":"Procedia Comput Sci"},{"key":"21394_CR37","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1155\/2022\/4001460","volume":"2022","author":"B Predic","year":"2022","unstructured":"Predic B, Manic D, Saracevic M, Karabasevic D, Stanujkic D (2022) Automatic image caption generation based on some machine learning algorithms. Math Probl Eng 2022:1\u201311. https:\/\/doi.org\/10.1155\/2022\/4001460","journal-title":"Math Probl Eng"},{"key":"21394_CR38","unstructured":"Rampal H, Mohanty A (2020) Efficient CNN-LSTM based image captioning using neural network compression. ArXiv:1\u20138 http:\/\/arxiv.org\/abs\/2012.09708"},{"key":"21394_CR39","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/ICCECE48148.2020.9223087","volume-title":"International conference on computer, electrical and communication engineering","author":"A Rathi","year":"2020","unstructured":"Rathi A (2020) Deep learning apporch for image captioning in Hindi language. In: International conference on computer, electrical and communication engineering, pp 1\u20138. https:\/\/doi.org\/10.1109\/ICCECE48148.2020.9223087"},{"key":"21394_CR40","doi-asserted-by":"publisher","first-page":"522","DOI":"10.1109\/CONIT55038.2022.9847918","volume-title":"2nd international conference on intelligent technologies","author":"O Sargar","year":"2022","unstructured":"Sargar O, Kinger S (2022) Image captioning methods and metrics. In: 2nd international conference on intelligent technologies, pp 522\u2013526. https:\/\/doi.org\/10.1109\/CONIT55038.2022.9847918"},{"key":"21394_CR41","doi-asserted-by":"publisher","first-page":"325","DOI":"10.1109\/PARC49193.2020.236619","volume-title":"International conference on power electronics and iot applications in renewable energy and its control","author":"H Sharma","year":"2020","unstructured":"Sharma H, Agrahari M, Singh SK, Firoj M, Mishra RK (2020) Image captioning: a comprehensive survey. In: International conference on power electronics and iot applications in renewable energy and its control, pp 325\u2013328. https:\/\/doi.org\/10.1109\/PARC49193.2020.236619"},{"key":"21394_CR42","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1155\/2022\/9638438","volume":"2022","author":"A Singh","year":"2022","unstructured":"Singh A, Raguru JK, Prasad G, Chauhan S, Tiwari PK, Zaguia A, Ullah MA (2022) Medical image captioning using optimized deep learning model. Comput Intell Neurosci 2022:1\u20139. https:\/\/doi.org\/10.1155\/2022\/9638438","journal-title":"Comput Intell Neurosci"},{"key":"21394_CR43","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1155\/2020\/3062706","volume":"2020","author":"H Wang","year":"2020","unstructured":"Wang H, Zhang Y, Yu X (2020) An overview of image caption generation methods. Comput Intell Neurosci 2020:1\u201313. https:\/\/doi.org\/10.1155\/2020\/3062706","journal-title":"Comput Intell Neurosci"},{"key":"21394_CR44","unstructured":"Wang, J., Yang, Z., Hu, X., Li, L., Lin, K., Gan, Z., ... & Wang, L. (2022a). Git: A generative image-to-text transformer for vision and language. arXiv preprint arXiv:2205.14100. https:\/\/arxiv.org\/pdf\/2205.14100"},{"key":"21394_CR45","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2108.10904","author":"Z Wang","year":"2022","unstructured":"Wang Z, Yu J, Yu AW, Dai Z, Tsvetkov Y, Cao Y, Team, B (2022b) SIMVLM: simple visual language model pre- training with weak supervision. Arxiv. https:\/\/doi.org\/10.48550\/arXiv.2108.10904","journal-title":"Arxiv"},{"key":"21394_CR46","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.sigpro.2019.107329","volume":"167","author":"S Yan","year":"2020","unstructured":"Yan S, Xie Y, Wu F, Smith JS, Lu W, Zhang B (2020) Image captioning via hierarchical attention mechanism and policy gradient optimization. Signal Process 167:1\u201312. https:\/\/doi.org\/10.1016\/j.sigpro.2019.107329","journal-title":"Signal Process"},{"issue":"5","key":"21394_CR47","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TPAMI.2020.3042192","volume":"44","author":"X Yang","year":"2022","unstructured":"Yang X, Zhang H, Cai J (2022) Auto-encoding and distilling scene graphs for image captioning. IEEE Trans Pattern Anal Mach Intell 44(5):1\u201314. https:\/\/doi.org\/10.1109\/TPAMI.2020.3042192","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"17","key":"21394_CR48","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s11042-021-10632-6","volume":"80","author":"Z Ye","year":"2021","unstructured":"Ye Z, Khan R, Naqvi N, Islam MS (2021) A novel automatic image caption generation using bidirectional long-short term memory framework. Multimed Tools Appl 80(17):1\u201326. https:\/\/doi.org\/10.1007\/s11042-021-10632-6","journal-title":"Multimed Tools Appl"},{"key":"21394_CR49","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.eswa.2021.115462","volume":"184","author":"J Zhang","year":"2021","unstructured":"Zhang J, Li K, Wang Z, Zhao X, Wang Z (2021a) Visual enhanced gLSTM for image captioning. Expert Syst Appl 184:1\u20139. https:\/\/doi.org\/10.1016\/j.eswa.2021.115462","journal-title":"Expert Syst Appl"},{"key":"21394_CR50","first-page":"5579","volume-title":"Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition","author":"P Zhang","year":"2021","unstructured":"Zhang P, Li X, Hu X, Yang J, Zhang L, Wang L, \u2026 Gao J (2021b) Vinvl: Revisiting visual representations in vision-language models. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 5579\u20135588 https:\/\/openaccess.thecvf.com\/content\/CVPR2021\/papers\/Zhang_VinVL_Revisiting_Visual_Representations_in_Vision-Language_Models_CVPR_2021_paper.pdf"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-026-21394-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-026-21394-4","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-026-21394-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T10:34:27Z","timestamp":1772102067000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-026-21394-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2,26]]},"references-count":50,"journal-issue":{"issue":"3","published-online":{"date-parts":[[2026,3]]}},"alternative-id":["21394"],"URL":"https:\/\/doi.org\/10.1007\/s11042-026-21394-4","relation":{},"ISSN":["1573-7721"],"issn-type":[{"value":"1573-7721","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,2,26]]},"assertion":[{"value":"2 September 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 December 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 January 2026","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 February 2026","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"This article does not contain any studies with human participants or animals performed by any of the authors.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent of publication"}},{"value":"The authors declare that they have no conflict of interest.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"The authors declare that they have no competing interests.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"228"}}