{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T15:07:45Z","timestamp":1775228865190,"version":"3.50.1"},"reference-count":61,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2026,1,29]],"date-time":"2026-01-29T00:00:00Z","timestamp":1769644800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T00:00:00Z","timestamp":1775174400000},"content-version":"vor","delay-in-days":64,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J. King Saud Univ. Comput. Inf. Sci."],"published-print":{"date-parts":[[2026,4]]},"DOI":"10.1007\/s44443-026-00475-2","type":"journal-article","created":{"date-parts":[[2026,1,29]],"date-time":"2026-01-29T11:35:38Z","timestamp":1769686538000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Dual-level interactive learning with symptom-aware attention for automatic radiology report generation"],"prefix":"10.1007","volume":"38","author":[{"given":"Minghao","family":"Tang","sequence":"first","affiliation":[]},{"given":"Weina","family":"Ding","sequence":"additional","affiliation":[]},{"given":"Wei","family":"Cai","sequence":"additional","affiliation":[]},{"given":"Yuan","family":"Jiang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,29]]},"reference":[{"key":"475_CR1","doi-asserted-by":"crossref","unstructured":"Anderson P, He X, Buehler C, Teney D, Johnson M, Gould S, Zhang L (2018) Bottom-up and top-down attention for image captioning and visual question answering. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 6077\u20136086","DOI":"10.1109\/CVPR.2018.00636"},{"key":"475_CR2","unstructured":"Banerjee S, Lavie A (2005) Meteor: an automatic metric for mt evaluation with improved correlation with human judgments. In: Proceedings of the Acl workshop on intrinsic and extrinsic evaluation measures for machine translation and\/or summarization, pp 65\u201372"},{"key":"475_CR3","unstructured":"Bi X, Chen D, Chen G, Chen S, Dai D, Deng C, Ding H, Dong K, Du Q, Fu Z et al (2024) Deepseek llm: scaling open-source language models with longtermism. arXiv:2401.02954"},{"key":"475_CR4","doi-asserted-by":"crossref","unstructured":"Bu S, Li T, Yang Y, Dai Z (2024) Instance-level expert knowledge and aggregate discriminative attention for radiology report generation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 14194\u201314204","DOI":"10.1109\/CVPR52733.2024.01346"},{"key":"475_CR5","unstructured":"Chambon P, Delbrouck J-B, Sounack T, Huang S-C, Chen Z, Varma M, Truong SQ, Langlotz CP et al (2024) Chexpert plus: hundreds of thousands of aligned radiology texts, images and patients. arXiv e-prints, 2405"},{"key":"475_CR6","doi-asserted-by":"crossref","unstructured":"Chen Y, Du C, Li C, Hu J, Shi Y, Xiong S, Zhu XX, Mou L (2024) Unicrossadapter: multimodal adaptation of clip for radiology report generation. In: International workshop on foundation models for general medical AI, pp 113\u2013123. Springer","DOI":"10.1007\/978-3-031-73471-7_12"},{"key":"475_CR7","doi-asserted-by":"crossref","unstructured":"Cheng K, Song W, Ma Z, Zhu W, Zhu Z, Zhang J (2023) Beyond generic: enhancing image captioning with real-world knowledge using vision-language pre-training model. In: Proceedings of the 31st ACM international conference on multimedia, pp 5038\u20135047","DOI":"10.1145\/3581783.3611987"},{"key":"475_CR8","doi-asserted-by":"crossref","unstructured":"Chen J, Guo H, Yi K, Li B, Elhoseiny M (2022) Visualgpt: Data-efficient adaptation of pretrained language models for image captioning. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 18030\u201318040","DOI":"10.1109\/CVPR52688.2022.01750"},{"key":"475_CR9","doi-asserted-by":"crossref","unstructured":"Chen Z, Shen Y, Song Y, Wan X (2021) Cross-modal memory networks for radiology report generation. In: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing (Volume 1: Long Papers), pp 5904\u20135914","DOI":"10.18653\/v1\/2021.acl-long.459"},{"key":"475_CR10","unstructured":"Chen Z, Shen Y, Song Y, Wan X (2022) Cross-modal memory networks for radiology report generation. arXiv:2204.13258"},{"key":"475_CR11","doi-asserted-by":"crossref","unstructured":"Chen Z, Song Y, Chang T-H, Wan X (2020) Generating radiology reports via memory-driven transformer. In: Proceedings of the 2020 conference on Empirical Methods in Natural Language Processing (EMNLP), pp 1439\u20131449","DOI":"10.18653\/v1\/2020.emnlp-main.112"},{"key":"475_CR12","doi-asserted-by":"crossref","unstructured":"Cornia M, Stefanini M, Baraldi L, Cucchiara R (2020) Meshed-memory transformer for image captioning. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 10578\u201310587","DOI":"10.1109\/CVPR42600.2020.01059"},{"issue":"2","key":"475_CR13","doi-asserted-by":"publisher","first-page":"304","DOI":"10.1093\/jamia\/ocv080","volume":"23","author":"D Demner-Fushman","year":"2016","unstructured":"Demner-Fushman D, Kohli MD, Rosenman MB, Shooshan SE, Rodriguez L, Antani S, Thoma GR, McDonald CJ (2016) Preparing a collection of radiology examinations for distribution and retrieval. J Am Med Inform Assoc 23(2):304\u2013310","journal-title":"J Am Med Inform Assoc"},{"issue":"5","key":"475_CR14","doi-asserted-by":"publisher","first-page":"3079","DOI":"10.1109\/JBHI.2024.3371894","volume":"28","author":"P Divya","year":"2024","unstructured":"Divya P, Sravani Y, Vishnu C, Mohan CK, Chen YW (2024) Memory guided transformer with spatio-semantic visual extractor for medical report generation. IEEE J Biomed Health Inform 28(5):3079\u20133089","journal-title":"IEEE J Biomed Health Inform"},{"key":"475_CR15","unstructured":"Dubey A, Jauhri A, Pandey A, Kadian A, Al-Dahle A, Letman A, Mathur A, Schelten A, Yang A, Fan A et al (2024) The llama 3 herd of models. arXiv e-prints, 2407"},{"key":"475_CR16","doi-asserted-by":"publisher","unstructured":"Feng S, Azzollini D, Kim JS, Jin CK, Kim E, Gordon S, Yeoh J, Han MA, Lee A, Patel A, Urschler M, Fong A, Simmers C, Tarr G, Barnard S, Wilson B (2021) CANDID-PTX. Radiology: artificial intelligence. https:\/\/doi.org\/10.17608\/k6.auckland.14173982","DOI":"10.17608\/k6.auckland.14173982"},{"issue":"1","key":"475_CR17","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1111\/1754-9485.12014","volume":"57","author":"SK Goergen","year":"2013","unstructured":"Goergen SK, Pool FJ, Turner TJ, Grimm JE, Appleyard MN, Crock C, Fahey MC, Fay MF, Ferris NJ, Liew SM et al (2013) Evidence-based guideline for the written radiology report: methods, recommendations and implementation challenges. J Med Imaging Radiat Oncol 57(1):1\u20137","journal-title":"J Med Imaging Radiat Oncol"},{"key":"475_CR18","unstructured":"Grattafiori A, Dubey A, Jauhri A, Pandey A, Kadian A, Al-Dahle A, Letman A, Mathur A, Schelten A, Vaughan A et al (2024) The llama 3 herd of models. arXiv:2407.21783"},{"key":"475_CR19","doi-asserted-by":"crossref","unstructured":"Gu D, Gao Y, Zhou Y, Zhou M, Metaxas D (2025) Radalign: advancing radiology report generation with vision-language concept alignment. arXiv e-prints, 2501","DOI":"10.1007\/978-3-032-04981-0_46"},{"key":"475_CR20","doi-asserted-by":"crossref","unstructured":"Hou W, Cheng Y, Xu K, Li H, Hu Y, Li W, Liu J (2025) Radar: enhancing radiology report generation with supplementary knowledge injection. arXiv:2505.14318","DOI":"10.18653\/v1\/2025.acl-long.1279"},{"key":"475_CR21","doi-asserted-by":"crossref","unstructured":"Huang L, Wang W, Chen J, Wei X-Y (2019) Attention on attention for image captioning. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 4634\u20134643","DOI":"10.1109\/ICCV.2019.00473"},{"key":"475_CR22","doi-asserted-by":"crossref","unstructured":"Huang Z, Zhang X, Zhang S (2023) Kiut: knowledge-injected u-transformer for radiology report generation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 19809\u201319818","DOI":"10.1109\/CVPR52729.2023.01897"},{"key":"475_CR23","doi-asserted-by":"crossref","unstructured":"Hu X, Gan Z, Wang J, Yang Z, Liu Z, Lu Y, Wang L (2022) Scaling up vision-language pre-training for image captioning. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 17980\u201317989","DOI":"10.1109\/CVPR52688.2022.01745"},{"key":"475_CR24","doi-asserted-by":"crossref","unstructured":"Irvin J, Rajpurkar P, Ko M, Yu Y, Ciurea-Ilcus S, Chute C, Marklund H, Haghgoo B, Ball R, Shpanskaya K et al (2019) Chexpert: a large chest radiograph dataset with uncertainty labels and expert comparison. In: Proceedings of the AAAI conference on artificial intelligence, vol 33, pp 590\u2013597","DOI":"10.1609\/aaai.v33i01.3301590"},{"key":"475_CR25","doi-asserted-by":"crossref","unstructured":"Jing B, Wang Z, Xing E (2019) Show, describe and conclude: on exploiting the structure information of chest x-ray reports. In: Proceedings of the 57th annual meeting of the association for computational linguistics, pp 6570\u20136580","DOI":"10.18653\/v1\/P19-1657"},{"key":"475_CR26","doi-asserted-by":"crossref","unstructured":"Jing B, Xie P, Xing E (2018) On the automatic generation of medical imaging reports. In: Proceedings of the 56th annual meeting of the association for computational linguistics (Volume 1: Long Papers), pp 2577\u20132586","DOI":"10.18653\/v1\/P18-1240"},{"key":"475_CR27","unstructured":"Johnson AE, Pollard TJ, Greenbaum NR, Lungren MP, Deng C-y, Peng Y, Lu Z, Mark RG, Berkowitz SJ, Horng S (2019) Mimic-cxr-jpg, a large publicly available database of labeled chest radiographs. arXiv:1901.07042"},{"issue":"4","key":"475_CR28","doi-asserted-by":"publisher","first-page":"1234","DOI":"10.1093\/bioinformatics\/btz682","volume":"36","author":"J Lee","year":"2020","unstructured":"Lee J, Yoon W, Kim S, Kim D, Kim S, So CH, Kang J (2020) Biobert: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics 36(4):1234\u20131240","journal-title":"Bioinformatics"},{"key":"475_CR29","doi-asserted-by":"crossref","unstructured":"Li H, Hao Y, Yu J, Zhu B, Wang S, Xu T (2024) Cvlp-navd: contrastive visual-language pre-training models for non-annotated visual description. ACM transactions on multimedia computing, communications and applications","DOI":"10.1145\/3708348"},{"key":"475_CR30","doi-asserted-by":"crossref","unstructured":"Li L, Li H, Ren P (2025) Underwater image captioning via attention mechanism based fusion of visual and textual information. Inf Fusion 103269","DOI":"10.1016\/j.inffus.2025.103269"},{"key":"475_CR31","unstructured":"Lin C-Y (2004) Rouge: a package for automatic evaluation of summaries. In: Text summarization branches out, pp 74\u201381"},{"issue":"2","key":"475_CR32","doi-asserted-by":"publisher","first-page":"102178","DOI":"10.1016\/j.ipm.2019.102178","volume":"57","author":"M Liu","year":"2020","unstructured":"Liu M, Li L, Hu H, Guan W, Tian J (2020) Image caption generation with dual attention mechanism. Inf Process Manag 57(2):102178","journal-title":"Inf Process Manag"},{"key":"475_CR33","doi-asserted-by":"crossref","unstructured":"Liu T, Wang J, Hu Y, Li M, Yi J, Chang X, Gao J, Yin B (2025) Hc-llm: Historical-constrained large language models for radiology report generation. In: Proceedings of the AAAI conference on artificial intelligence, vol 39, pp 5595\u20135603","DOI":"10.1609\/aaai.v39i6.32596"},{"key":"475_CR34","doi-asserted-by":"crossref","unstructured":"Liu F, Wu X, Ge S, Fan W, Zou Y (2021) Exploring and distilling posterior and prior knowledge for radiology report generation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 13753\u201313762","DOI":"10.1109\/CVPR46437.2021.01354"},{"key":"475_CR35","doi-asserted-by":"crossref","unstructured":"Liu F, Yin C, Wu X, Ge S, Zhang P, Sun X (2021) Contrastive attention for automatic chest x-ray report generation. In: Findings of the association for computational linguistics: ACL-IJCNLP 2021, pp 269\u2013280","DOI":"10.18653\/v1\/2021.findings-acl.23"},{"key":"475_CR36","doi-asserted-by":"crossref","unstructured":"Lu J, Xiong C, Parikh D, Socher R (2017) Knowing when to look: adaptive attention via a visual sentinel for image captioning. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 375\u2013383","DOI":"10.1109\/CVPR.2017.345"},{"key":"475_CR37","doi-asserted-by":"crossref","unstructured":"Miura Y, Zhang Y, Tsai E, Langlotz C, Jurafsky D (2021) Improving factual completeness and consistency of image-to-text radiology report generation. In: Proceedings of the 2021 conference of the North American chapter of the association for computational linguistics: human language technologies, pp 5288\u20135304","DOI":"10.18653\/v1\/2021.naacl-main.416"},{"key":"475_CR38","doi-asserted-by":"crossref","unstructured":"Papineni K, Roukos S, Ward T, Zhu W-J (2002) Bleu: a method for automatic evaluation of machine translation. In: Proceedings of the 40th annual meeting of the association for computational linguistics, pp 311\u2013318","DOI":"10.3115\/1073083.1073135"},{"issue":"8","key":"475_CR39","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford A, Wu J, Child R, Luan D, Amodei D, Sutskever I et al (2019) Language models are unsupervised multitask learners. OpenAI Blog 1(8):9","journal-title":"OpenAI Blog"},{"issue":"1","key":"475_CR40","doi-asserted-by":"publisher","first-page":"487","DOI":"10.1038\/s41597-022-01608-8","volume":"9","author":"EP Reis","year":"2022","unstructured":"Reis EP, de Paiva JP, da Silva MC, Ribeiro GA, Paiva VF, Bulgarelli L, Lee HM, Santos PV, Brito VM, Amaral LT et al (2022) Brax, brazilian labeled chest x-ray dataset. Sci Data 9(1):487","journal-title":"Sci Data"},{"key":"475_CR41","unstructured":"Ren S, He K, Girshick R, Sun J (2015) Faster r-cnn: towards real-time object detection with region proposal networks. NIPS, vol 28"},{"issue":"1","key":"475_CR42","doi-asserted-by":"publisher","first-page":"0262209","DOI":"10.1371\/journal.pone.0262209","volume":"17","author":"M Sirshar","year":"2022","unstructured":"Sirshar M, Paracha MFK, Akram MU, Alghamdi NS, Zaidi SZY, Fatima T (2022) Attention based automated radiology report generation using cnn and lstm. PLoS ONE 17(1):0262209","journal-title":"PLoS ONE"},{"key":"475_CR43","doi-asserted-by":"crossref","unstructured":"Sun Y, Khor HG, Wang Y, Wang Z, Zhao H, Zhang Y, Ma L, Zheng Z, Liao H (2024) Continually tuning a large language model for multi-domain radiology report generation. In: International conference on medical image computing and computer-assisted intervention, pp 177\u2013187. Springer","DOI":"10.1007\/978-3-031-72086-4_17"},{"key":"475_CR44","doi-asserted-by":"crossref","unstructured":"Tanida T, M\u00fcller P, Kaissis G, Rueckert D (2023) Interactive and explainable region-guided radiology report generation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 7433\u20137442","DOI":"10.1109\/CVPR52729.2023.00718"},{"key":"475_CR45","doi-asserted-by":"crossref","unstructured":"Tiwari V, Bapat K, Shrimali KR, Singh SK, Tiwari B, Jain S, Sharma HK (2021) Automatic generation of chest x-ray medical imaging reports using lstm-cnn. In: Proceedings of the international conference on data science, machine learning and artificial intelligence, pp 80\u201385","DOI":"10.1145\/3484824.3484918"},{"key":"475_CR46","doi-asserted-by":"crossref","unstructured":"Vedantam R, Lawrence\u00a0Zitnick C, Parikh D (2015) Cider: consensus-based image description evaluation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4566\u20134575","DOI":"10.1109\/CVPR.2015.7299087"},{"key":"475_CR47","doi-asserted-by":"crossref","unstructured":"Vinyals O, Toshev A, Bengio S, Erhan D (2015) Show and tell: a neural image caption generator. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3156\u20133164","DOI":"10.1109\/CVPR.2015.7298935"},{"key":"475_CR48","doi-asserted-by":"crossref","unstructured":"Wang Z, Liu L, Wang L, Zhou L (2023) Metransformer: radiology report generation by transformer with multiple learnable expert tokens. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 11558\u201311567","DOI":"10.1109\/CVPR52729.2023.01112"},{"key":"475_CR49","doi-asserted-by":"crossref","unstructured":"Wang X, Wang F, Li Y, Ma Q, Wang S, Jiang B, Tang J (2025) Cxpmrg-bench: pre-training and benchmarking for x-ray medical report generation on chexpert plus dataset. In: Proceedings of the computer vision and pattern recognition conference, pp 5123\u20135133","DOI":"10.1109\/CVPR52734.2025.00483"},{"key":"475_CR50","doi-asserted-by":"crossref","unstructured":"Wang R, Wu Y, Sheng Z (2025) Clipcap++: an efficient image captioning approach via image encoder optimization and llm fine-tuning. Appl Soft Comput 113469","DOI":"10.1016\/j.asoc.2025.113469"},{"key":"475_CR51","doi-asserted-by":"crossref","unstructured":"Wang Z, Zhou L, Wang L, Li X (2021) A self-boosting framework for automated radiographic report generation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 2433\u20132442","DOI":"10.1109\/CVPR46437.2021.00246"},{"key":"475_CR52","doi-asserted-by":"crossref","unstructured":"Xiao T, Shi L, Liu P, Wang Z, Bai C (2025) Radiology report generation via multi-objective preference optimization. In: Proceedings of the AAAI conference on artificial intelligence, vol 39, pp 8664\u20138672","DOI":"10.1609\/aaai.v39i8.32936"},{"key":"475_CR53","doi-asserted-by":"crossref","unstructured":"Xing Q, Song Z, Zhang Y, Feng N, Yu J, Yang W (2025) Mca-rg: enhancing llms with medical concept alignment for radiology report generation. arXiv e-prints, 2507","DOI":"10.1007\/978-3-032-04971-1_36"},{"key":"475_CR54","unstructured":"Xu K, Ba J, Kiros R, Cho K, Courville A, Salakhudinov R, Zemel R, Bengio Y (2015) Show, attend and tell: Neural image caption generation with visual attention. In: International conference on machine learning, pp 2048\u20132057. PMLR"},{"key":"475_CR55","doi-asserted-by":"publisher","first-page":"127823","DOI":"10.1016\/j.neucom.2024.127823","volume":"593","author":"X Yang","year":"2024","unstructured":"Yang X, Yang Y, Ma S, Li Z, Dong W, Wo\u017aniak M (2024) Samt-generator: a second-attention for image captioning based on multi-stage transformer network. Neurocomputing 593:127823","journal-title":"Neurocomputing"},{"issue":"4","key":"475_CR56","doi-asserted-by":"publisher","first-page":"2152","DOI":"10.1109\/JBHI.2024.3350077","volume":"28","author":"X Yi","year":"2024","unstructured":"Yi X, Fu Y, Liu R, Zhang H, Hua R (2024) Tsget: two-stage global enhanced transformer for automatic radiology report generation. IEEE J Biomed Health Inform 28(4):2152\u20132162","journal-title":"IEEE J Biomed Health Inform"},{"key":"475_CR57","doi-asserted-by":"crossref","unstructured":"Yin C, Qian B, Wei J, Li X, Zhang X, Li Y, Zheng Q (2019) Automatic generation of medical imaging diagnostic report with hierarchical recurrent neural network. In: 2019 IEEE International Conference on Data Mining (ICDM), pp 728\u2013737. IEEE","DOI":"10.1109\/ICDM.2019.00083"},{"key":"475_CR58","doi-asserted-by":"publisher","first-page":"132","DOI":"10.1016\/j.neucom.2018.11.114","volume":"392","author":"X Zeng","year":"2020","unstructured":"Zeng X, Wen L, Liu B, Qi X (2020) Deep learning for ultrasound image caption generation based on object detection. Neurocomputing 392:132\u2013141","journal-title":"Neurocomputing"},{"key":"475_CR59","unstructured":"Zhang X, Acosta JN, Miller J, Huang O, Rajpurkar P (2025) Rexgradient-160k: a large-scale publicly available dataset of chest radiographs with free-text reports. arXiv:2505.00228"},{"key":"475_CR60","doi-asserted-by":"crossref","unstructured":"Zhang X, Sun X, Luo Y, Ji J, Zhou Y, Wu Y, Huang F, Ji R (2021) Rstnet: captioning with adaptive attention on visual and non-visual words. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 15465\u201315474","DOI":"10.1109\/CVPR46437.2021.01521"},{"key":"475_CR61","volume":"127","author":"K Zhao","year":"2024","unstructured":"Zhao K, Xiong W (2024) Exploring region features in remote sensing image captioning. Int J Appl Earth Obs Geoinf 127:103672","journal-title":"Int J Appl Earth Obs Geoinf"}],"container-title":["Journal of King Saud University Computer and Information Sciences"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s44443-026-00475-2","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44443-026-00475-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44443-026-00475-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T14:22:50Z","timestamp":1775226170000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s44443-026-00475-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1,29]]},"references-count":61,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2026,4]]}},"alternative-id":["475"],"URL":"https:\/\/doi.org\/10.1007\/s44443-026-00475-2","relation":{},"ISSN":["1319-1578","2213-1248"],"issn-type":[{"value":"1319-1578","type":"print"},{"value":"2213-1248","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,1,29]]},"assertion":[{"value":"11 October 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 January 2026","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 January 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"We declare that we do not have any commercial or associative interest that represents a conflict of interest in connection with the work submitted.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of Interest"}}],"article-number":"105"}}