{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,5]],"date-time":"2026-02-05T21:46:21Z","timestamp":1770327981144,"version":"3.49.0"},"reference-count":46,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,12,13]],"date-time":"2025-12-13T00:00:00Z","timestamp":1765584000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,12,24]],"date-time":"2025-12-24T00:00:00Z","timestamp":1766534400000},"content-version":"vor","delay-in-days":11,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["32471362"],"award-info":[{"award-number":["32471362"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003453","name":"Natural Science Foundation of Guangdong Province","doi-asserted-by":"publisher","award":["2023A1515012606"],"award-info":[{"award-number":["2023A1515012606"]}],"id":[{"id":"10.13039\/501100003453","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J. King Saud Univ. Comput. Inf. Sci."],"published-print":{"date-parts":[[2026,1]]},"DOI":"10.1007\/s44443-025-00402-x","type":"journal-article","created":{"date-parts":[[2025,12,13]],"date-time":"2025-12-13T18:23:26Z","timestamp":1765650206000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["DMMSA net: a dual multi-modal self-attention network for lung target areas segmentation using chest radiology images and reports"],"prefix":"10.1007","volume":"38","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8870-7750","authenticated-orcid":false,"given":"Longjiang","family":"E","sequence":"first","affiliation":[]},{"given":"Yi","family":"Cai","sequence":"additional","affiliation":[]},{"given":"Baisong","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Mingjie","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Jiangguo","family":"Lin","sequence":"additional","affiliation":[]},{"given":"Tao","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,12,13]]},"reference":[{"key":"402_CR1","doi-asserted-by":"publisher","unstructured":"Azad R, Jia Y, Aghdam EK et al (2023) Enhancing Medical Image Segmentation with TransCeption: A Multi-Scale Feature Fusion Approach. CoRR. abs\/2301.10847 https:\/\/doi.org\/10.48550\/ARXIV.2301.10847","DOI":"10.48550\/ARXIV.2301.10847"},{"key":"402_CR2","doi-asserted-by":"publisher","unstructured":"Boecking B, Usuyama N, Bannur S et al (2022) Making the most of text semantics to improve biomedical vision\u2013language processing. European conference on computer vision. Springer Nature Switzerland,\u00a0Cham. https:\/\/doi.org\/10.1007\/978-3-031-20059-5_1","DOI":"10.1007\/978-3-031-20059-5_1"},{"issue":"2","key":"402_CR3","doi-asserted-by":"publisher","first-page":"577","DOI":"10.1109\/TMI.2013.2290491","volume":"33","author":"S Candemir","year":"2014","unstructured":"Candemir S, Jaeger S, Palaniappan K et al (2014) Lung segmentation in chest radiographs using anatomical atlases with nonrigid registration. IEEE Trans Med Imaging 33(2):577\u2013590. https:\/\/doi.org\/10.1109\/TMI.2013.2290491","journal-title":"IEEE Trans Med Imaging"},{"key":"402_CR4","doi-asserted-by":"publisher","unstructured":"Cao H, Wang Y, Chen J et al (2021) Swin-Unet: Unet-like Pure Transformer for Medical Image Segmentation. arXiv preprint arXiv: https:\/\/doi.org\/10.48550\/arXiv.2105.05537.","DOI":"10.48550\/arXiv.2105.05537"},{"key":"402_CR5","doi-asserted-by":"publisher","unstructured":"Chen J, Lu Y, Yu Q et al (2021) TransUNet: Transformers Make Strong Encoders for Medical Image Segmentation. arXiv preprint arXiv:https:\/\/doi.org\/10.48550\/arXiv.2102.04306.","DOI":"10.48550\/arXiv.2102.04306"},{"issue":"4","key":"402_CR6","doi-asserted-by":"publisher","first-page":"834","DOI":"10.1109\/TPAMI.2017.2699184","volume":"40","author":"LC Chen","year":"2018","unstructured":"Chen LC, Papandreou G, Kokkinos I et al (2018) Deeplab: semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs. IEEE Trans Pattern Anal Mach Intell 40(4):834\u2013848. https:\/\/doi.org\/10.1109\/TPAMI.2017.2699184","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"402_CR7","doi-asserted-by":"publisher","unstructured":"Cheng J, Dong L, Lapata M (2016) Long short-term memory-networks for machine reading. In: Proceedings of the 2016 conference on empirical methods in natural language processing, Austin, Texas, USA. https:\/\/doi.org\/10.18653\/v1\/D16-1053","DOI":"10.18653\/v1\/D16-1053"},{"key":"402_CR8","doi-asserted-by":"publisher","first-page":"3504","DOI":"10.1109\/TASLP.2021.3124365","volume":"29","author":"Y Cui","year":"2021","unstructured":"Cui Y, Che W, Liu T et al (2021) Pre-training with whole word masking for Chinese BERT. IEEE\/ACM Transactions on Audio, Speech, and Language Processing 29:3504\u20133514. https:\/\/doi.org\/10.1109\/TASLP.2021.3124365","journal-title":"IEEE\/ACM Transactions on Audio, Speech, and Language Processing"},{"key":"402_CR9","doi-asserted-by":"publisher","unstructured":"Degerli A, Kiranyaz S, Chowdhury MEH et al (2022) Osegnet: operational segmentation network for covid-19 detection using chest x-ray images. In: 2022 IEEE International Conference on Image Processing (ICIP). https:\/\/doi.org\/10.1109\/ICIP46576.2022.9897412","DOI":"10.1109\/ICIP46576.2022.9897412"},{"key":"402_CR10","doi-asserted-by":"publisher","DOI":"10.1002\/aisy.202400201","author":"Y Ding","year":"2024","unstructured":"Ding Y, Liu J, He Y et al (2024) Fi-net: rethinking feature interactions for medical image segmentation. Adv Intell Syst. https:\/\/doi.org\/10.1002\/aisy.202400201","journal-title":"Adv Intell Syst"},{"issue":"10","key":"402_CR11","doi-asserted-by":"publisher","first-page":"1617","DOI":"10.1002\/ppul.24431","volume":"54","author":"L E","year":"2019","unstructured":"E L, Zhao B, Guo Y et al (2019) Using deep-learning techniques for pulmonary-thoracic segmentations and improvement of pneumonia diagnosis in pediatric chest radiographs. Pediatr Pulmonol 54(10):1617\u20131626. https:\/\/doi.org\/10.1002\/ppul.24431","journal-title":"Pediatr Pulmonol"},{"key":"402_CR12","doi-asserted-by":"publisher","unstructured":"Fu J, Liu J, Tian H et al (2019) Dual attention network for scene segmentation. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA. https:\/\/doi.org\/10.1109\/CVPR.2019.00326","DOI":"10.1109\/CVPR.2019.00326"},{"issue":"9","key":"402_CR13","doi-asserted-by":"publisher","first-page":"3507","DOI":"10.1109\/JBHI.2021.3059453","volume":"25","author":"X Fu","year":"2021","unstructured":"Fu X, Bi L, Kumar A et al (2021) Multimodal spatial attention module for targeting multimodal PET-CT lung tumor segmentation. IEEE J Biomed Health Inform 25(9):3507\u20133516. https:\/\/doi.org\/10.1109\/JBHI.2021.3059453","journal-title":"IEEE J Biomed Health Inform"},{"key":"402_CR14","doi-asserted-by":"publisher","unstructured":"He K, Zhang X, Ren S et al (2016) Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA. https:\/\/doi.org\/10.1109\/CVPR.2016.90","DOI":"10.1109\/CVPR.2016.90"},{"issue":"8","key":"402_CR15","doi-asserted-by":"publisher","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","volume":"9","author":"S Hochreiter","year":"1997","unstructured":"Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735\u20131780. https:\/\/doi.org\/10.1162\/neco.1997.9.8.1735","journal-title":"Neural Comput"},{"key":"402_CR16","doi-asserted-by":"publisher","unstructured":"Huang S-C, Shen L, Lungren MP et al (2021) GLoRIA: a multimodal global-local representation learning framework for label-efficient medical image recognition. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada. https:\/\/doi.org\/10.1109\/ICCV48922.2021.00391","DOI":"10.1109\/ICCV48922.2021.00391"},{"issue":"20","key":"402_CR17","doi-asserted-by":"publisher","first-page":"4362","DOI":"10.3390\/electronics12204362","volume":"12","author":"Y Huo","year":"2023","unstructured":"Huo Y, Gang S, Guan C (2023) FCIHMRT: feature cross-layer interaction hybrid method based on Res2Net and transformer for remote sensing scene classification. Electronics 12(20):4362. https:\/\/doi.org\/10.3390\/electronics12204362","journal-title":"Electronics"},{"issue":"9","key":"402_CR18","doi-asserted-by":"publisher","first-page":"850","DOI":"10.1109\/34.232073","volume":"15","author":"DP Huttenlocher","year":"1993","unstructured":"Huttenlocher DP, Klanderman GA, Rucklidge WJ (1993) Comparing images using the Hausdorff distance. IEEE Trans Pattern Anal Mach Intell 15(9):850\u2013863. https:\/\/doi.org\/10.1109\/34.232073","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"2","key":"402_CR19","doi-asserted-by":"publisher","first-page":"203","DOI":"10.1038\/s41592-020-01008-z","volume":"18","author":"F Isensee","year":"2021","unstructured":"Isensee F, Jaeger PF, Kohl SAA et al (2021) NnU-net: a self-configuring method for deep learning-based biomedical image segmentation. Nat Methods 18(2):203\u2013211. https:\/\/doi.org\/10.1038\/s41592-020-01008-z","journal-title":"Nat Methods"},{"key":"402_CR20","doi-asserted-by":"publisher","unstructured":"Kim J-H, Jun J, Zhang B-T (2018) Bilinear attention networks. In: Proceedings of the 32nd international conference on neural information processing systems, Montr\u00e9al, Canada. https:\/\/doi.org\/10.48550\/arXiv.1805.07932","DOI":"10.48550\/arXiv.1805.07932"},{"key":"402_CR21","doi-asserted-by":"publisher","unstructured":"Krhenb\u00fchl P, Koltun V (2012) Efficient inference in fully connected crfs with gaussian edge potentials. In: The Twenty-sixth annual conference on Neural Information Processing Systems (NIPS), Harrahs and Harveys, Lake Tahoe. https:\/\/doi.org\/10.48550\/arXiv.1210.5644","DOI":"10.48550\/arXiv.1210.5644"},{"issue":"4","key":"402_CR22","doi-asserted-by":"publisher","first-page":"1234","DOI":"10.1093\/bioinformatics\/btz682","volume":"36","author":"J Lee","year":"2020","unstructured":"Lee J, Yoon W, Kim S et al (2020) BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics 36(4):1234\u20131240. https:\/\/doi.org\/10.1093\/bioinformatics\/btz682","journal-title":"Bioinformatics"},{"key":"402_CR23","doi-asserted-by":"publisher","DOI":"10.1109\/TGRS.2024.3522293","author":"S Lei","year":"2025","unstructured":"Lei S, Xiao X, Zhang T et al (2025) Exploring fine-grained image-text alignment for referring remote sensing image segmentation. IEEE Trans Geosci Remote Sens. https:\/\/doi.org\/10.1109\/TGRS.2024.3522293","journal-title":"IEEE Trans Geosci Remote Sens"},{"issue":"1","key":"402_CR24","doi-asserted-by":"publisher","first-page":"96","DOI":"10.1109\/TMI.2023.3291719","volume":"43","author":"Z Li","year":"2024","unstructured":"Li Z, Li Y, Li Q et al (2024) LViT: language meets vision transformer in medical image segmentation. IEEE Trans Med Imaging 43(1):96\u2013107. https:\/\/doi.org\/10.1109\/TMI.2023.3291719","journal-title":"IEEE Trans Med Imaging"},{"key":"402_CR25","doi-asserted-by":"publisher","unstructured":"Liu L, Nie X, Lu H et al (2025) Small but Strong: Lightweight Architecture Improves Lung Nodule Detection and Segmentation. Advanced Intelligent Systems. n\/a (n\/a), 2401093. https:\/\/doi.org\/10.1002\/aisy.202401093","DOI":"10.1002\/aisy.202401093"},{"key":"402_CR26","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2024.102634","volume":"113","author":"X Liu","year":"2025","unstructured":"Liu X, Gao P, Yu T et al (2025b) Cswin-UNet: transformer unet with cross-shaped windows for medical image segmentation. Inf Fusion 113:102634. https:\/\/doi.org\/10.1016\/j.inffus.2024.102634","journal-title":"Inf Fusion"},{"key":"402_CR27","doi-asserted-by":"publisher","unstructured":"Luo R, Xu J, Zhang Y et al (2019) PKUSEG: A Toolkit for Multi-Domain Chinese Word Segmentation. CoRR. abs\/1905.11455 https:\/\/doi.org\/10.48550\/arXiv.1906.11455","DOI":"10.48550\/arXiv.1906.11455"},{"issue":"1","key":"402_CR28","doi-asserted-by":"publisher","first-page":"654","DOI":"10.1038\/s41467-024-44824-z","volume":"15","author":"J Ma","year":"2024","unstructured":"Ma J, He Y, Li F et al (2024) Segment anything in medical images. Nat Commun 15(1):654. https:\/\/doi.org\/10.1038\/s41467-024-44824-z","journal-title":"Nat Commun"},{"key":"402_CR29","doi-asserted-by":"publisher","unstructured":"Muller P, Kaissis G, Zou C et al (2022) Joint learning of localized representations from medical images and reports. In: European Conference on Computer Vision (ECCV), Tel Aviv, Israel. https:\/\/doi.org\/10.1007\/978-3-031-19809-0_39","DOI":"10.1007\/978-3-031-19809-0_39"},{"key":"402_CR30","doi-asserted-by":"publisher","unstructured":"Oktay O, Schlemper J, Folgoc LL et al (2018) Attention U-Net: Learning Where to Look for the Pancreas. arXiv preprint arXiv: https:\/\/doi.org\/10.48550\/arXiv.1804.03999.","DOI":"10.48550\/arXiv.1804.03999"},{"issue":"5","key":"402_CR32","doi-asserted-by":"publisher","first-page":"97","DOI":"10.1007\/s44443-025-00106-2","volume":"37","author":"F Qin","year":"2025","unstructured":"Qin F, Liang Y, Yang C et al (2025) Medical image segmentation network based on multi-scale cross-attention and wavelet transform. J King Saud Univ Comput Inf Sci 37(5):97. https:\/\/doi.org\/10.1007\/s44443-025-00106-2","journal-title":"J King Saud Univ Comput Inf Sci"},{"key":"402_CR33","doi-asserted-by":"publisher","first-page":"149","DOI":"10.48550\/arXiv.1804.11191","volume":"1","author":"Z Qin","year":"2018","unstructured":"Qin Z, Yu F, Liu C et al (2018) How convolutional neural network see the world - a survey of convolutional neural network visualization methods. Math Found Comput 1:149\u2013180. https:\/\/doi.org\/10.48550\/arXiv.1804.11191","journal-title":"Math Found Comput"},{"key":"402_CR34","doi-asserted-by":"publisher","unstructured":"Ronneberger O, Fischer P, Brox T (2015) U-Net: convolutional networks for biomedical image segmentation. International conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"402_CR35","doi-asserted-by":"publisher","DOI":"10.1016\/j.jbi.2023.104482","volume":"146","author":"Z Sun","year":"2023","unstructured":"Sun Z, Lin M, Zhu Q et al (2023) A scoping review on multimodal deep learning in biomedical images and texts. J Biomed Inform 146:104482. https:\/\/doi.org\/10.1016\/j.jbi.2023.104482","journal-title":"J Biomed Inform"},{"key":"402_CR36","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2025.113166","volume":"311","author":"J Tagnamas","year":"2025","unstructured":"Tagnamas J, Ramadan H, Yahyaouy A et al (2025) SCA-inceptionunext: a lightweight spatial-channel-attention-based network for efficient medical image segmentation. Knowledge-Based Syst 311:113166. https:\/\/doi.org\/10.1016\/j.knosys.2025.113166","journal-title":"Knowledge-Based Syst"},{"key":"402_CR37","doi-asserted-by":"publisher","unstructured":"Tang F, Ding J, Quan Q et al. (2024) CMUNEXT: an efficient medical image segmentation network based on large kernel and skip fusion. In: 2024 IEEE International Symposium on Biomedical Imaging (ISBI). https:\/\/doi.org\/10.1109\/ISBI56570.2024.10635609","DOI":"10.1109\/ISBI56570.2024.10635609"},{"issue":"12","key":"402_CR38","doi-asserted-by":"publisher","first-page":"1399","DOI":"10.1038\/s41551-022-00936-9","volume":"6","author":"E Tiu","year":"2022","unstructured":"Tiu E, Talius E, Patel P et al (2022) Expert-level detection of pathologies from unannotated chest x-ray images via self-supervised learning. Nat Biomed Eng 6(12):1399\u20131406. https:\/\/doi.org\/10.1038\/s41551-022-00936-9","journal-title":"Nat Biomed Eng"},{"key":"402_CR39","doi-asserted-by":"publisher","first-page":"151","DOI":"10.1007\/978-3-031-16437-8_15","volume":"13433","author":"NK Tomar","year":"2022","unstructured":"Tomar NK, Jha D, Bagci U et al (2022) TGANet: text-guided attention for improved polyp segmentation. Med Image Comput Comput Assist Interv 13433:151\u2013160. https:\/\/doi.org\/10.1007\/978-3-031-16437-8_15","journal-title":"Med Image Comput Comput Assist Interv"},{"key":"402_CR40","doi-asserted-by":"publisher","unstructured":"Vaswani A, Shazeer N, Parmar N et al (2017) Attention is all you need. Proceedings of the 31st International Conference on Neural Information Processing Systems. https:\/\/doi.org\/10.48550\/arXiv.1706.03762","DOI":"10.48550\/arXiv.1706.03762"},{"key":"402_CR41","doi-asserted-by":"publisher","unstructured":"Wang F, Zhou Y, Wang S et al (2022) Multi-granularity cross-modal alignment for generalized medical visual representation learning. In: 36th conference on Neural Information Processing Systems (NeurIPS 2022). https:\/\/doi.org\/10.48550\/arXiv.2210.06044","DOI":"10.48550\/arXiv.2210.06044"},{"key":"402_CR42","doi-asserted-by":"publisher","unstructured":"Wang Z, Wu Z, Agarwal D et al (2022) MedCLIP: contrastive learning from unpaired medical images and text. Conference on empirical methods in natural language processing, Abu Dhabi, United Arab Emirates, association for computational linguistics. https:\/\/doi.org\/10.18653\/v1\/2022.emnlp-main.256","DOI":"10.18653\/v1\/2022.emnlp-main.256"},{"issue":"1","key":"402_CR43","doi-asserted-by":"publisher","first-page":"13","DOI":"10.1109\/TBDATA.2021.3056564","volume":"7","author":"Q Yan","year":"2021","unstructured":"Yan Q, Wang B, Gong D et al (2021) COVID-19 chest CT image segmentation network by multi-scale fusion and enhancement operations. IEEE Trans Big Data 7(1):13\u201324. https:\/\/doi.org\/10.1109\/TBDATA.2021.3056564","journal-title":"IEEE Trans Big Data"},{"key":"402_CR44","doi-asserted-by":"publisher","unstructured":"Ye L, Rochan M, Liu Z et al (2019) Cross-modal self-attention network for referring image segmentation. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA. https:\/\/doi.org\/10.1109\/CVPR.2019.01075","DOI":"10.1109\/CVPR.2019.01075"},{"key":"402_CR45","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2024.127475","volume":"581","author":"H Zhang","year":"2024","unstructured":"Zhang H, Wang L, Li S et al (2024) Area-keywords cross-modal alignment for referring image segmentation. Neurocomputing 581:127475. https:\/\/doi.org\/10.1016\/j.neucom.2024.127475","journal-title":"Neurocomputing"},{"issue":"1","key":"402_CR46","doi-asserted-by":"publisher","first-page":"27","DOI":"10.1631\/fitee.1700808","volume":"19","author":"Q-S Zhang","year":"2018","unstructured":"Zhang Q-S, Zhu S-C (2018) Visual interpretability for deep learning: a survey. Front Inf Technol Electron Eng 19(1):27\u201339. https:\/\/doi.org\/10.1631\/fitee.1700808","journal-title":"Front Inf Technol Electron Eng"},{"key":"402_CR47","doi-asserted-by":"publisher","unstructured":"Zhang Y, Jiang H, Miura Y et al (2020) Contrastive learning of medical visual representations from paired images and text. Machine learning in health care. https:\/\/doi.org\/10.48550\/arXiv.2010.00747","DOI":"10.48550\/arXiv.2010.00747"}],"container-title":["Journal of King Saud University Computer and Information Sciences"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44443-025-00402-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s44443-025-00402-x","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44443-025-00402-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,5]],"date-time":"2026-02-05T09:51:27Z","timestamp":1770285087000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s44443-025-00402-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,13]]},"references-count":46,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,1]]}},"alternative-id":["402"],"URL":"https:\/\/doi.org\/10.1007\/s44443-025-00402-x","relation":{},"ISSN":["1319-1578","2213-1248"],"issn-type":[{"value":"1319-1578","type":"print"},{"value":"2213-1248","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,12,13]]},"assertion":[{"value":"29 July 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 November 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 December 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}],"article-number":"1"}}