{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,10]],"date-time":"2026-02-10T16:35:46Z","timestamp":1770741346838,"version":"3.49.0"},"reference-count":55,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2026,2,9]],"date-time":"2026-02-09T00:00:00Z","timestamp":1770595200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,2,9]],"date-time":"2026-02-09T00:00:00Z","timestamp":1770595200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-026-21407-2","type":"journal-article","created":{"date-parts":[[2026,2,9]],"date-time":"2026-02-09T22:37:03Z","timestamp":1770676623000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["HAT-UNet: Hybrid attention transformer-based U-Net with multi-source fusion for medical image segmentation"],"prefix":"10.1007","volume":"85","author":[{"ORCID":"https:\/\/orcid.org\/0009-0007-5967-2564","authenticated-orcid":false,"given":"Noura","family":"Bentaher","sequence":"first","affiliation":[]},{"given":"Younes","family":"Kabbadj","sequence":"additional","affiliation":[]},{"given":"Samira","family":"Lafraxo","sequence":"additional","affiliation":[]},{"given":"Mohamed","family":"Ben Salah","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,2,9]]},"reference":[{"key":"21407_CR1","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2024.107200","volume":"101","author":"A Garbaz","year":"2025","unstructured":"Garbaz A, Oukdach Y, Charfi S, El Ansari M, Koutti L, Salihoun M (2025) Dmfc-uformer: Depthwise multi-scale factorized convolution transformer-based unet for medical image segmentation. Biomed Signal Process Control 101:107200","journal-title":"Biomed Signal Process Control"},{"key":"21407_CR2","doi-asserted-by":"publisher","first-page":"52","DOI":"10.1016\/j.ymeth.2024.10.010","volume":"232","author":"A Garbaz","year":"2024","unstructured":"Garbaz A, Oukdach Y, Charfi S, El Ansari M, Koutti L, Salihoun M (2024) Mlfa-unet: A multi-level feature assembly unet for medical image segmentation. Methods 232:52\u201364. https:\/\/doi.org\/10.1016\/j.ymeth.2024.10.010","journal-title":"Methods"},{"issue":"14","key":"21407_CR3","doi-asserted-by":"publisher","first-page":"6299","DOI":"10.3390\/app14146299","volume":"14","author":"C Zhu","year":"2024","unstructured":"Zhu C, Cheng K, Hua X (2024) A medical image segmentation network with multi-scale and dual-branch attention. Appl Sci 14(14):6299","journal-title":"Appl Sci"},{"key":"21407_CR4","doi-asserted-by":"crossref","unstructured":"Ronneberger O, Fischer P, Brox T (2015) U-net: Convolutional networks for biomedical image segmentation. In: Medical image computing and computer-assisted intervention\u2013MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18. Springer, pp 234\u2013241","DOI":"10.1007\/978-3-319-24574-4_28"},{"issue":"6","key":"21407_CR5","doi-asserted-by":"publisher","first-page":"1856","DOI":"10.1109\/TMI.2019.2959609","volume":"39","author":"Z Zhou","year":"2019","unstructured":"Zhou Z, Siddiquee MMR, Tajbakhsh N, Liang J (2019) Unet++: Redesigning skip connections to exploit multiscale features in image segmentation. IEEE Trans Med Imaging 39(6):1856\u20131867","journal-title":"IEEE Trans Med Imaging"},{"key":"21407_CR6","unstructured":"Oktay O, Schlemper J, Folgoc LL, Lee M, Heinrich M, Misawa K, Mori K, McDonagh S, Hammerla NY, Kainz B et al (2018) Attention u-net: Learning where to look for the pancreas. arXiv:1804.03999"},{"key":"21407_CR7","doi-asserted-by":"publisher","first-page":"94","DOI":"10.1016\/j.isprsjprs.2020.01.013","volume":"162","author":"FI Diakogiannis","year":"2020","unstructured":"Diakogiannis FI, Waldner F, Caccetta P, Wu C (2020) Resunet-a: A deep learning framework for semantic segmentation of remotely sensed data. ISPRS J Photogramm Remote Sens 162:94\u2013114","journal-title":"ISPRS J Photogramm Remote Sens"},{"key":"21407_CR8","doi-asserted-by":"crossref","unstructured":"Jha D, Smedsrud PH, Riegler MA, Johansen D, De\u00a0Lange T, Halvorsen P, Johansen HD (2019) Resunet++: An advanced architecture for medical image segmentation. In: 2019 IEEE international symposium on multimedia (ISM). IEEE, pp 225\u20132255","DOI":"10.1109\/ISM46123.2019.00049"},{"issue":"3","key":"21407_CR9","doi-asserted-by":"publisher","first-page":"719","DOI":"10.3390\/life13030719","volume":"13","author":"S Lafraxo","year":"2023","unstructured":"Lafraxo S, Souaidi M, El Ansari M, Koutti L (2023) Semantic segmentation of digestive abnormalities from wce images by using attresu-net architecture. Life 13(3):719","journal-title":"Life"},{"key":"21407_CR10","doi-asserted-by":"crossref","unstructured":"Alom MZ, Hasan M, Yakopcic C, Taha TM, Asari VK (2018) Recurrent residual convolutional neural network based on u-net (r2u-net) for medical image segmentation. arXiv:1802.06955","DOI":"10.1109\/NAECON.2018.8556686"},{"issue":"12","key":"21407_CR11","doi-asserted-by":"publisher","first-page":"2481","DOI":"10.1109\/TPAMI.2016.2644615","volume":"39","author":"V Badrinarayanan","year":"2017","unstructured":"Badrinarayanan V, Kendall A, Cipolla R (2017) Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans Pattern Anal Mach Intell 39(12):2481\u20132495","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"20","key":"21407_CR12","doi-asserted-by":"publisher","first-page":"8589","DOI":"10.3390\/s23208589","volume":"23","author":"A Al Qurri","year":"2023","unstructured":"Al Qurri A, Almekkawy M (2023) Improved unet with attention for medical image segmentation. Sensors 23(20):8589","journal-title":"Sensors"},{"key":"21407_CR13","doi-asserted-by":"crossref","unstructured":"Woo S, Park J, Lee J-Y, Kweon IS (2018) Cbam: Convolutional block attention module. In: Proceedings of the European conference on computer vision (ECCV), pp 3\u201319","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"21407_CR14","doi-asserted-by":"crossref","unstructured":"Bentaher N, Lafraxo S, Kabbadj Y, Ben\u00a0Salah M, El\u00a0Ansari M, Wakrim S (2025) R2a-unet: double attention mechanisms with residual blocks for enhanced mri image segmentation. Multimed Tools Appl 1\u201331","DOI":"10.1007\/s11042-025-20617-4"},{"key":"21407_CR15","unstructured":"Vaswani A (2017) Attention is all you need. Adv Neural Inf Process Syst"},{"key":"21407_CR16","unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S, Uszkoreit J, Houlsby N (2021) An image is worth 16x16 words: Transformers for image recognition at scale. In: International conference on learning representations (ICLR)"},{"key":"21407_CR17","doi-asserted-by":"publisher","first-page":"106731","DOI":"10.1016\/j.bspc.2024.106731","volume":"97","author":"B Wang","year":"2024","unstructured":"Wang B, Qin J, Lv L, Cheng M, Li L, He J, Li D, Xia D, Wang M, Ren H et al (2024) Dsml-unet: Depthwise separable convolution network with multiscale large kernel for medical image segmentation. Biomed Signal Process Control 97:106731","journal-title":"Biomed Signal Process Control"},{"key":"21407_CR18","doi-asserted-by":"crossref","unstructured":"Long J, Shelhamer E, Darrell T (2015) Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3431\u20133440","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"21407_CR19","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2024.107456","volume":"103","author":"M Hu","year":"2025","unstructured":"Hu M, Dong Y, Li J, Jiang L, Zhang P, Ping Y (2025) Lamffnet: Lightweight adaptive multi-layer feature fusion network for medical image segmentation. Biomed Signal Process Control 103:107456","journal-title":"Biomed Signal Process Control"},{"key":"21407_CR20","doi-asserted-by":"publisher","DOI":"10.1016\/j.artmed.2022.102407","volume":"133","author":"AB Hamida","year":"2022","unstructured":"Hamida AB, Devanne M, Weber J, Truntzer C, Derang\u00e8re V, Ghiringhelli F, Forestier G, Wemmert C (2022) Weakly supervised learning using attention gates for colon cancer histopathological image segmentation. Artif Intell Med 133:102407","journal-title":"Artif Intell Med"},{"key":"21407_CR21","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2021.103077","volume":"71","author":"D Maji","year":"2022","unstructured":"Maji D, Sigedar P, Singh M (2022) Attention res-unet with guided decoder for semantic segmentation of brain tumors. Biomed Signal Process Control 71:103077","journal-title":"Biomed Signal Process Control"},{"key":"21407_CR22","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2022.104032","volume":"79","author":"Z Kong","year":"2023","unstructured":"Kong Z, Zhang M, Zhu W, Yi Y, Wang T, Zhang B (2023) Data enhancement based on m2-unet for liver segmentation in computed tomography. Biomed Signal Process Control 79:104032","journal-title":"Biomed Signal Process Control"},{"key":"21407_CR23","unstructured":"Howard AG (2017) Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv:1704.04861"},{"issue":"6","key":"21407_CR24","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1145\/3065386","volume":"60","author":"A Krizhevsky","year":"2017","unstructured":"Krizhevsky A, Sutskever I, Hinton GE (2017) Imagenet classification with deep convolutional neural networks. Commun ACM 60(6):84\u201390","journal-title":"Commun ACM"},{"key":"21407_CR25","doi-asserted-by":"crossref","unstructured":"Valanarasu JMJ, Patel VM (2022) Unext: Mlp-based rapid medical image segmentation network. In: International conference on medical image computing and computer-assisted intervention. Springer, pp 23\u201333","DOI":"10.1007\/978-3-031-16443-9_3"},{"key":"21407_CR26","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2022.109512","volume":"253","author":"Z Han","year":"2022","unstructured":"Han Z, Jian M, Wang G-G (2022) Convunext: An efficient convolution neural network for medical image segmentation. Knowl-Based Syst 253:109512","journal-title":"Knowl-Based Syst"},{"key":"21407_CR27","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2022.104112","volume":"79","author":"B Zhan","year":"2023","unstructured":"Zhan B, Song E, Liu H, Gong Z, Ma G, Hung C-C (2023) Cfnet: A medical image segmentation method using the multi-view attention mechanism and adaptive fusion strategy. Biomed Signal Process Control 79:104112","journal-title":"Biomed Signal Process Control"},{"key":"21407_CR28","doi-asserted-by":"crossref","unstructured":"Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7132\u20137141","DOI":"10.1109\/CVPR.2018.00745"},{"key":"21407_CR29","doi-asserted-by":"crossref","unstructured":"Wang Q, Wu B, Zhu P, Li P, Zuo W, Hu Q (2020) Eca-net: Efficient channel attention for deep convolutional neural networks. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 11534\u201311542","DOI":"10.1109\/CVPR42600.2020.01155"},{"issue":"4","key":"21407_CR30","doi-asserted-by":"publisher","first-page":"2880","DOI":"10.1109\/TII.2020.3022912","volume":"17","author":"H Hu","year":"2020","unstructured":"Hu H, Li Q, Zhao Y, Zhang Y (2020) Parallel deep learning algorithms with hybrid attention mechanism for image segmentation of lung tumors. IEEE Trans Industr Inf 17(4):2880\u20132889","journal-title":"IEEE Trans Industr Inf"},{"key":"21407_CR31","doi-asserted-by":"publisher","DOI":"10.3389\/fninf.2021.781551","volume":"15","author":"Y Xiao","year":"2021","unstructured":"Xiao Y, Yin H, Wang S-H, Zhang Y-D (2021) Trec: transferred resnet and cbam for detecting brain diseases. Front Neuroinform 15:781551","journal-title":"Front Neuroinform"},{"key":"21407_CR32","doi-asserted-by":"publisher","first-page":"605132","DOI":"10.3389\/fbioe.2020.605132","volume":"8","author":"Q Jin","year":"2020","unstructured":"Jin Q, Meng Z, Sun C, Cui H, Su R (2020) Ra-unet: A hybrid deep attention-aware network to extract liver and tumor in ct scans. Front Bioeng Biotechnol 8:605132","journal-title":"Front Bioeng Biotechnol"},{"key":"21407_CR33","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.cag.2020.05.003","volume":"90","author":"C Li","year":"2020","unstructured":"Li C, Tan Y, Chen W, Luo X, He Y, Gao Y, Li F (2020) Anu-net: Attention-based nested u-net to exploit full resolution features for medical image segmentation. Comput Graph 90:11\u201320","journal-title":"Comput Graph"},{"key":"21407_CR34","doi-asserted-by":"publisher","first-page":"10543","DOI":"10.1109\/ACCESS.2020.2965651","volume":"8","author":"YS Hariyani","year":"2020","unstructured":"Hariyani YS, Eom H, Park C (2020) Da-capnet: dual attention deep learning based on u-net for nailfold capillary segmentation. IEEE Access 8:10543\u201310553","journal-title":"IEEE Access"},{"key":"21407_CR35","doi-asserted-by":"crossref","unstructured":"Wang H, Cao P, Wang J, Zaiane OR (2022) Uctransnet: rethinking the skip connections in u-net from a channel-wise perspective with transformer. In: Proceedings of the AAAI conference on artificial intelligence, vol 36, pp 2441\u20132449","DOI":"10.1609\/aaai.v36i3.20144"},{"key":"21407_CR36","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2024.106285","volume":"94","author":"Y Feng","year":"2024","unstructured":"Feng Y, Zhu X, Zhang X, Li Y, Lu H (2024) Pamsnet: A medical image segmentation network based on spatial pyramid and attention mechanism. Biomed Signal Process Control 94:106285","journal-title":"Biomed Signal Process Control"},{"key":"21407_CR37","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2024.106977","volume":"100","author":"Q Chen","year":"2025","unstructured":"Chen Q, Wang W, Wang Z, Jia H, Zhao M (2025) Ds&stm-net: A novel hybrid network of feature mutual fusion for medical image segmentation. Biomed Signal Process Control 100:106977","journal-title":"Biomed Signal Process Control"},{"key":"21407_CR38","doi-asserted-by":"crossref","unstructured":"Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z, Lin S, Guo B (2021) Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 10012\u201310022","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"21407_CR39","unstructured":"Chen J, Lu Y, Yu Q, Luo X, Adeli E, Wang Y, Lu L, Yuille AL, Zhou Y (2021) Transunet: Transformers make strong encoders for medical image segmentation. arXiv:2102.04306"},{"key":"21407_CR40","doi-asserted-by":"crossref","unstructured":"Valanarasu JMJ, Oza P, Hacihaliloglu I, Patel VM (2021) Medical transformer: Gated axial-attention for medical image segmentation. In: Medical image computing and computer assisted intervention\u2013MICCAI 2021: 24th International Conference, Strasbourg, France, September 27\u2013October 1, 2021, Proceedings, Part I 24. Springer, pp 36\u201346","DOI":"10.1007\/978-3-030-87193-2_4"},{"key":"21407_CR41","doi-asserted-by":"crossref","unstructured":"Hatamizadeh A, Tang Y, Nath V, Yang D, Myronenko A, Landman B, Roth HR, Xu D (2022) Unetr: Transformers for 3d medical image segmentation. In: Proceedings of the IEEE\/CVF winter conference on applications of computer vision, pp 574\u2013584","DOI":"10.1109\/WACV51458.2022.00181"},{"key":"21407_CR42","doi-asserted-by":"crossref","unstructured":"Hatamizadeh A, Nath V, Tang Y, Yang D, Roth HR, Xu D (2021) Swin unetr: Swin transformers for semantic segmentation of brain tumors in mri images. In: International MICCAI brainlesion workshop. Springer, pp 272\u2013284","DOI":"10.1007\/978-3-031-08999-2_22"},{"key":"21407_CR43","first-page":"1","volume":"60","author":"X He","year":"2022","unstructured":"He X, Zhou Y, Zhao J, Zhang D, Yao R, Xue Y (2022) Swin transformer embedding unet for remote sensing image semantic segmentation. IEEE Trans Geosci Remote Sens 60:1\u201315","journal-title":"IEEE Trans Geosci Remote Sens"},{"issue":"4","key":"21407_CR44","doi-asserted-by":"publisher","first-page":"5355","DOI":"10.1109\/TNNLS.2022.3204090","volume":"35","author":"H Du","year":"2022","unstructured":"Du H, Wang J, Liu M, Wang Y, Meijering E (2022) Swinpa-net: Swin transformer-based multiscale feature pyramid aggregation network for medical image segmentation. IEEE Trans Neural Netw Learn Syst 35(4):5355\u20135366","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"key":"21407_CR45","doi-asserted-by":"crossref","unstructured":"Zhou H-Y, Guo J, Zhang Y, Han X, Yu L, Wang L, Yu Y (2023) nnformer: Volumetric medical image segmentation via a 3d transformer. IEEE Trans Image Process","DOI":"10.1109\/TIP.2023.3293771"},{"key":"21407_CR46","doi-asserted-by":"crossref","unstructured":"Cao H, Wang Y, Chen J, Jiang D, Zhang X, Tian Q, Wang M (2022) Swin-unet: Unet-like pure transformer for medical image segmentation. In: European conference on computer vision. Springer, pp 205\u2013218","DOI":"10.1007\/978-3-031-25066-8_9"},{"key":"21407_CR47","doi-asserted-by":"crossref","unstructured":"Peiris H, Hayat M, Chen Z, Egan G, Harandi M (2022) A robust volumetric transformer for accurate 3d tumor segmentation. In: International conference on medical image computing and computer-assisted intervention. Springer, pp 162\u2013172","DOI":"10.1007\/978-3-031-16443-9_16"},{"key":"21407_CR48","doi-asserted-by":"crossref","unstructured":"Rahman MM, Marculescu R (2023) Medical image segmentation via cascaded attention decoding. In: Proceedings of the IEEE\/CVF winter conference on applications of computer vision, pp 6222\u20136231","DOI":"10.1109\/WACV56688.2023.00616"},{"key":"21407_CR49","doi-asserted-by":"crossref","unstructured":"Oukdach Y, Garbaz A, Kerkaou Z, El\u00a0Ansari M, Koutti L, El\u00a0Ouafdi AF, Salihoun M (2024) Uvit-seg: An efficient vit and u-net-based framework for accurate colorectal polyp segmentation in colonoscopy and wce images. J Imaging Inf Med 1\u201321","DOI":"10.1007\/s10278-024-01124-8"},{"key":"21407_CR50","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2024.107190","volume":"101","author":"G Chen","year":"2025","unstructured":"Chen G, Zhang M, Zhu J, Meng Y (2025) Pdca-net: Parallel dual-channel attention network for polyp segmentation. Biomed Signal Process Control 101:107190","journal-title":"Biomed Signal Process Control"},{"key":"21407_CR51","doi-asserted-by":"publisher","first-page":"94007","DOI":"10.1109\/ACCESS.2022.3204280","volume":"10","author":"P Chen","year":"2022","unstructured":"Chen P, Huang S, Yue Q (2022) Skin lesion segmentation using recurrent attentional convolutional networks. IEEE Access 10:94007\u201394018","journal-title":"IEEE Access"},{"issue":"1","key":"21407_CR52","doi-asserted-by":"publisher","first-page":"33668","DOI":"10.1038\/s41598-025-18202-8","volume":"15","author":"W El-Shafai","year":"2025","unstructured":"El-Shafai W, Ali AM, Alzaben N, El-Fattah IA (2025) H-fusion seg: dual-branch hyper-attention fusion network with sam integration for robust skin disease segmentation. Sci Rep 15(1):33668","journal-title":"Sci Rep"},{"key":"21407_CR53","doi-asserted-by":"publisher","first-page":"1935","DOI":"10.7717\/peerj-cs.1935","volume":"10","author":"Y Li","year":"2024","unstructured":"Li Y, Tian T, Hu J, Yuan C (2024) Sutrans-net: a hybrid transformer approach to skin lesion segmentation. PeerJ Comput Sci 10:1935","journal-title":"PeerJ Comput Sci"},{"key":"21407_CR54","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2024.107343","volume":"102","author":"M Golia\u0161","year":"2025","unstructured":"Golia\u0161 M, \u0160ikudov\u00e1 E (2025) Retinal blood vessel segmentation and inpainting networks with multi-level self-attention. Biomed Signal Process Control 102:107343","journal-title":"Biomed Signal Process Control"},{"key":"21407_CR55","doi-asserted-by":"publisher","first-page":"3438","DOI":"10.1038\/s41598-024-84255-w","volume":"15","author":"GB Kande","year":"2025","unstructured":"Kande GB, Nalluri MR, Manikandan R, Cho J, Veerappampalayam Easwaramoorthy S (2025) Multi scale multi attention network for blood vessel segmentation in fundus images. Sci Rep 15:3438","journal-title":"Sci Rep"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-026-21407-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-026-21407-2","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-026-21407-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,9]],"date-time":"2026-02-09T22:37:07Z","timestamp":1770676627000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-026-21407-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2,9]]},"references-count":55,"journal-issue":{"issue":"2","published-online":{"date-parts":[[2026,2]]}},"alternative-id":["21407"],"URL":"https:\/\/doi.org\/10.1007\/s11042-026-21407-2","relation":{},"ISSN":["1573-7721"],"issn-type":[{"value":"1573-7721","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,2,9]]},"assertion":[{"value":"20 July 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 January 2026","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 January 2026","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 February 2026","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that no funds, grants, or other support was received during the preparation of this manuscript.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Funding"}},{"value":"The authors have no relevant financial or nonfinancial interests to disclose.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing Interests"}},{"value":"The authors declare that they have no conflict of interest.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of Interest"}},{"value":"This article does not contain studies involving human participants or animals performed by any of the authors.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical Approval"}},{"value":"Informed consent was obtained from all individual participants included in the study.","order":6,"name":"Ethics","group":{"name":"EthicsHeading","label":"Informed Consent"}}],"article-number":"157"}}