{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T15:23:07Z","timestamp":1772119387226,"version":"3.50.1"},"reference-count":44,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,12,4]],"date-time":"2025-12-04T00:00:00Z","timestamp":1764806400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,12,4]],"date-time":"2025-12-04T00:00:00Z","timestamp":1764806400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62202416"],"award-info":[{"award-number":["62202416"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62162068"],"award-info":[{"award-number":["62162068"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2026,2]]},"DOI":"10.1007\/s00530-025-02069-w","type":"journal-article","created":{"date-parts":[[2025,12,4]],"date-time":"2025-12-04T07:21:33Z","timestamp":1764832893000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Enhancing infrared-visible image fusion via text-guided adaptive feature integration"],"prefix":"10.1007","volume":"32","author":[{"given":"Jundong","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Yanan","family":"Guo","sequence":"additional","affiliation":[]},{"given":"Kangjian","family":"He","sequence":"additional","affiliation":[]},{"given":"Dan","family":"Xu","sequence":"additional","affiliation":[]},{"given":"SongHan","family":"Zheng","sequence":"additional","affiliation":[]},{"given":"WenCheng","family":"Mei","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,12,4]]},"reference":[{"issue":"2","key":"2069_CR1","doi-asserted-by":"publisher","first-page":"770","DOI":"10.1109\/TCSVT.2023.3289170","volume":"34","author":"S Park","year":"2024","unstructured":"Park, S., Vien, A.G., Lee, C.: Cross-modal transformers for infrared and visible image fusion. IEEE Trans. Circuits Syst. Video Technol. 34(2), 770\u2013785 (2024). https:\/\/doi.org\/10.1109\/TCSVT.2023.3289170","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"2069_CR2","doi-asserted-by":"publisher","first-page":"199","DOI":"10.1016\/j.optcom.2014.12.032","volume":"341","author":"G Cui","year":"2015","unstructured":"Cui, G., Feng, H., Xu, Z., Li, Q., Chen, Y.: Detail preserved fusion of visible and infrared images using regional saliency extraction and multi-scale image decomposition. Opt. Commun. 341, 199\u2013209 (2015). https:\/\/doi.org\/10.1016\/j.optcom.2014.12.032","journal-title":"Opt. Commun."},{"issue":"2","key":"2069_CR3","doi-asserted-by":"publisher","first-page":"1335","DOI":"10.1007\/s00371-024-03423-1","volume":"41","author":"Y Liu","year":"2025","unstructured":"Liu, Y., Li, X., Zhong, W.: Simplifusion: a simplified infrared and visible image fusion network. Vis. Comput. 41(2), 1335\u20131350 (2025). https:\/\/doi.org\/10.1007\/s00371-024-03423-1","journal-title":"Vis. Comput."},{"issue":"5","key":"2069_CR4","doi-asserted-by":"publisher","first-page":"288","DOI":"10.1007\/s00530-024-01500-y","volume":"30","author":"B Liu","year":"2024","unstructured":"Liu, B., Wu, Z., Li, X.: Msadrcn metalearning based joint superresolution fusion of infrared and visible images. Multimedia Syst. 30(5), 288 (2024). https:\/\/doi.org\/10.1007\/s00530-024-01500-y","journal-title":"Multimedia Syst."},{"issue":"1","key":"2069_CR5","doi-asserted-by":"publisher","DOI":"10.1155\/2013\/720979","volume":"2013","author":"Z-J Feng","year":"2013","unstructured":"Feng, Z.-J., Zhang, X.-L., Yuan, L.-Y., Wang, J.-N.: Infrared target detection and location for visual surveillance using fusion scheme of visible and infrared images. Math. Probl. Eng. 2013(1), 720979 (2013). https:\/\/doi.org\/10.1155\/2013\/720979","journal-title":"Math. Probl. Eng."},{"key":"2069_CR6","doi-asserted-by":"publisher","unstructured":"Liu, J., Liu, Z., Wu, G., Ma, L., Liu, R., Zhong, W., Luo, Z., Fan, X.: Multi-interactive feature learning and a full-time multi-modality benchmark for image fusion and segmentation. In: 2023 IEEE\/CVF International Conference on Computer Vision (ICCV), 8081\u20138090 (2023). https:\/\/doi.org\/10.1109\/ICCV51070.2023.00745","DOI":"10.1109\/ICCV51070.2023.00745"},{"key":"2069_CR7","doi-asserted-by":"publisher","first-page":"661","DOI":"10.1016\/j.isprsjprs.2025.01.022","volume":"220","author":"S Yi","year":"2025","unstructured":"Yi, S., Chen, M., Yuan, X., Guo, S., Wang, J.: An interactive fusion attention-guided network for ground surface hot spring fluids segmentation in dual-spectrum uav images. ISPRS J. Photogramm. Remote. Sens. 220, 661\u2013691 (2025). https:\/\/doi.org\/10.1016\/j.isprsjprs.2025.01.022","journal-title":"ISPRS J. Photogramm. Remote. Sens."},{"key":"2069_CR8","doi-asserted-by":"publisher","unstructured":"Singh, R., Vatsa, M., Noore, A.: Integrated multilevel image fusion and match score fusion of visible and infrared face images for robust face recognition. Pattern Recognit. 41(3), 880\u2013893 (2008). https:\/\/doi.org\/10.1016\/j.patcog.2007.06.022","DOI":"10.1016\/j.patcog.2007.06.022"},{"key":"2069_CR9","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1016\/j.ins.2019.08.066","volume":"508","author":"J Chen","year":"2020","unstructured":"Chen, J., Li, X., Luo, L., Mei, X., Ma, J.: Infrared and visible image fusion based on target-enhanced multiscale transform decomposition. Inf. Sci. 508, 64\u201378 (2020). https:\/\/doi.org\/10.1016\/j.ins.2019.08.066","journal-title":"Inf. Sci."},{"key":"2069_CR10","doi-asserted-by":"publisher","first-page":"94","DOI":"10.1016\/j.infrared.2013.07.010","volume":"61","author":"J Adu","year":"2013","unstructured":"Adu, J., Gan, J., Wang, Y., Huang, J.: Image fusion based on nonsubsampled contourlet transform for infrared and visible light image. Infrared Phys. Technol. 61, 94\u2013100 (2013). https:\/\/doi.org\/10.1016\/j.infrared.2013.07.010","journal-title":"Infrared Phys. Technol."},{"issue":"4","key":"2069_CR11","doi-asserted-by":"publisher","first-page":"1456","DOI":"10.1109\/TCE.2007.4429237","volume":"53","author":"D Drajic","year":"2007","unstructured":"Drajic, D., Cvejic, N.: Adaptive fusion of multimodal surveillance image sequences in visual sensor networks. IEEE Trans. Consum. Electron. 53(4), 1456\u20131462 (2007). https:\/\/doi.org\/10.1109\/TCE.2007.4429237","journal-title":"IEEE Trans. Consum. Electron."},{"issue":"12","key":"2069_CR12","doi-asserted-by":"publisher","first-page":"1882","DOI":"10.1109\/LSP.2016.2618776","volume":"23","author":"Y Liu","year":"2016","unstructured":"Liu, Y., Chen, X., Ward, R.K., Jane Wang, Z.: Image fusion with convolutional sparse representation. IEEE Signal Process. Lett. 23(12), 1882\u20131886 (2016). https:\/\/doi.org\/10.1109\/LSP.2016.2618776","journal-title":"IEEE Signal Process. Lett."},{"key":"2069_CR13","doi-asserted-by":"publisher","first-page":"8","DOI":"10.1016\/j.infrared.2017.02.005","volume":"82","author":"J Ma","year":"2017","unstructured":"Ma, J., Zhou, Z., Wang, B., Zong, H.: Infrared and visible image fusion based on visual saliency map and weighted least square optimization. Infrared Phys. Technol. 82, 8\u201317 (2017). https:\/\/doi.org\/10.1016\/j.infrared.2017.02.005","journal-title":"Infrared Phys. Technol."},{"key":"2069_CR14","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1016\/j.inffus.2021.12.004","volume":"82","author":"L Tang","year":"2022","unstructured":"Tang, L., Yuan, J., Ma, J.: Image fusion in the loop of high-level vision tasks: a semantic-aware real-time infrared and visible image fusion network. Inf. Fusion 82, 28\u201342 (2022). https:\/\/doi.org\/10.1016\/j.inffus.2021.12.004","journal-title":"Inf. Fusion"},{"key":"2069_CR15","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2023.101828","volume":"98","author":"D Wang","year":"2023","unstructured":"Wang, D., Liu, J., Liu, R., Fan, X.: An interactively reinforced paradigm for joint infrared-visible image fusion and saliency object detection. Inf. Fusion 98, 101828 (2023). https:\/\/doi.org\/10.1016\/j.inffus.2023.101828","journal-title":"Inf. Fusion"},{"key":"2069_CR16","unstructured":"Radford, A., Kim, J.W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., Krueger, G., Sutskever, I.: Learning transferable visual models from natural language supervision (2021). arXiv:2103.00020"},{"key":"2069_CR17","unstructured":"Li, J., Li, D., Savarese, S., Hoi, S.: BLIP-2: bootstrapping language-image pre-training with frozen image encoders and large language models (2023). arXiv:2301.12597"},{"key":"2069_CR18","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2024.109192","volume":"137","author":"C Wang","year":"2024","unstructured":"Wang, C., Pu, Y., Zhao, Z., Nie, R., Cao, J., Xu, D.: FCLFusion: a frequency-aware and collaborative learning for infrared and visible image fusion. Eng. Appl. Artif. Intell. 137, 109192 (2024). https:\/\/doi.org\/10.1016\/j.engappai.2024.109192","journal-title":"Eng. Appl. Artif. Intell."},{"key":"2069_CR19","doi-asserted-by":"publisher","first-page":"385","DOI":"10.1109\/TCI.2024.3369398","volume":"10","author":"H Tang","year":"2024","unstructured":"Tang, H., Liu, G., Qian, Y., Wang, J., Xiong, J.: Egefusion: towards edge gradient enhancement in infrared and visible image fusion with multi-scale transform. IEEE Trans. Comput. Imaging 10, 385\u2013398 (2024). https:\/\/doi.org\/10.1109\/TCI.2024.3369398","journal-title":"IEEE Trans. Comput. Imaging"},{"key":"2069_CR20","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2023.101851","volume":"99","author":"B Yang","year":"2023","unstructured":"Yang, B., Jiang, Z., Pan, D., Yu, H., Gui, W.: Detail-aware near infrared and visible fusion with multi-order hyper-laplacian priors. Inf. Fusion 99, 101851 (2023). https:\/\/doi.org\/10.1016\/j.inffus.2023.101851","journal-title":"Inf. Fusion"},{"issue":"5","key":"2069_CR21","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2019","unstructured":"Li, H., Wu, X.-J.: Densefuse: a fusion approach to infrared and visible images. IEEE Trans. Image Process. 28(5), 2614\u20132623 (2019). https:\/\/doi.org\/10.1109\/TIP.2018.2887342","journal-title":"IEEE Trans. Image Process."},{"key":"2069_CR22","doi-asserted-by":"publisher","DOI":"10.1016\/j.optlaseng.2023.107925","volume":"173","author":"Y Qian","year":"2024","unstructured":"Qian, Y., Liu, G., Tang, H., Xing, M., Chang, R.: Btsfusion: fusion of infrared and visible image via a mechanism of balancing texture and salience. Opt. Lasers Eng. 173, 107925 (2024). https:\/\/doi.org\/10.1016\/j.optlaseng.2023.107925","journal-title":"Opt. Lasers Eng."},{"key":"2069_CR23","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2024.111041","volume":"158","author":"X Wang","year":"2025","unstructured":"Wang, X., Fang, L., Zhao, J., Pan, Z., Li, H., Li, Y.: Mmae: a universal image fusion method via mask attention mechanism. Pattern Recognit. 158, 111041 (2025). https:\/\/doi.org\/10.1016\/j.patcog.2024.111041","journal-title":"Pattern Recognit."},{"issue":"6","key":"2069_CR24","doi-asserted-by":"publisher","first-page":"509","DOI":"10.1016\/j.vrih.2023.06.003","volume":"5","author":"M Wang","year":"2023","unstructured":"Wang, M., Meng, M., Liu, J., Wu, J.: Adequate alignment and interaction for cross-modal retrieval. Virtual Reality Intell. Hardware 5(6), 509\u2013522 (2023). https:\/\/doi.org\/10.1016\/j.vrih.2023.06.003","journal-title":"Virtual Reality Intell. Hardware"},{"key":"2069_CR25","doi-asserted-by":"publisher","first-page":"50","DOI":"10.1109\/TMM.2021.3120873","volume":"25","author":"X Lin","year":"2023","unstructured":"Lin, X., Sun, S., Huang, W., Sheng, B., Li, P., Feng, D.D.: Eapt: efficient attention pyramid transformer for image processing. IEEE Trans. Multimedia 25, 50\u201361 (2023). https:\/\/doi.org\/10.1109\/TMM.2021.3120873","journal-title":"IEEE Trans. Multimedia"},{"key":"2069_CR26","doi-asserted-by":"publisher","first-page":"2226","DOI":"10.1109\/TMM.2022.3144890","volume":"25","author":"N Jiang","year":"2023","unstructured":"Jiang, N., Sheng, B., Li, P., Lee, T.-Y.: Photohelper: portrait photographing guidance via deep feature retrieval and fusion. IEEE Trans. Multimedia 25, 2226\u20132238 (2023). https:\/\/doi.org\/10.1109\/TMM.2022.3144890","journal-title":"IEEE Trans. Multimedia"},{"key":"2069_CR27","unstructured":"Zhu, X., Su, W., Lu, L., Li, B., Wang, X., Dai, J.: Deformable DETR: deformable transformers for end-to-end object detection (2021). arXiv:2010.04159"},{"key":"2069_CR28","doi-asserted-by":"publisher","unstructured":"Vs, V., Jose\u00a0Valanarasu, J.M., Oza, P., Patel, V.M.: Image fusion transformer. In: 2022 IEEE International Conference on Image Processing (ICIP), pp. 3566\u20133570 (2022). https:\/\/doi.org\/10.1109\/ICIP46576.2022.9897280","DOI":"10.1109\/ICIP46576.2022.9897280"},{"key":"2069_CR29","doi-asserted-by":"publisher","unstructured":"Tang, W., He, F., Liu, Y.: Itfuse: an interactive transformer for infrared and visible image fusion. Pattern Recognit. 156, 110822 (2024). https:\/\/doi.org\/10.1016\/j.patcog.2024.110822","DOI":"10.1016\/j.patcog.2024.110822"},{"key":"2069_CR30","doi-asserted-by":"publisher","unstructured":"Liu, J., Fan, X., Huang, Z., Wu, G., Liu, R., Zhong, W., Luo, Z.: Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5792\u20135801 (2022). https:\/\/doi.org\/10.1109\/CVPR52688.2022.00571","DOI":"10.1109\/CVPR52688.2022.00571"},{"issue":"3","key":"2069_CR31","doi-asserted-by":"publisher","first-page":"2257","DOI":"10.1002\/cav.2257","volume":"35","author":"L Sun","year":"2024","unstructured":"Sun, L., Wang, Y., Qin, W.: A language-directed virtual human motion generation approach based on musculoskeletal models. Comput. Anim. Virtual Worlds 35(3), 2257 (2024). https:\/\/doi.org\/10.1002\/cav.2257","journal-title":"Comput. Anim. Virtual Worlds"},{"issue":"3","key":"2069_CR32","doi-asserted-by":"publisher","first-page":"2268","DOI":"10.1002\/cav.2268","volume":"35","author":"Y Bao","year":"2024","unstructured":"Bao, Y., Liu, X., Qi, Y., Liu, R., Li, H.: Adaptive information fusion network for multi-modal personality recognition. Comput. Anim. Virtual Worlds 35(3), 2268 (2024). https:\/\/doi.org\/10.1002\/cav.2268","journal-title":"Comput. Anim. Virtual Worlds"},{"key":"2069_CR33","unstructured":"Zhao, Z., Deng, L., Bai, H., Cui, Y., Zhang, Z., Zhang, Y., Qin, H., Chen, D., Zhang, J., Wang, P., Gool, L.V.: Image fusion via vision-language model (2024). arXiv:2402.02235"},{"key":"2069_CR34","unstructured":"Zhang, H., Cao, L., Ma, J.: Text-DiFuse: an interactive multi-modal image fusion framework based on text-modulated diffusion model (2024). arXiv:2410.23905"},{"key":"2069_CR35","doi-asserted-by":"publisher","unstructured":"Wang, Y., Miao, L., Zhou, Z., Zhang, L., Qiao, Y.: Infrared and visible image fusion with language-driven loss in clip embedding space (2024). https:\/\/doi.org\/10.48550\/ARXIV.2402.16267","DOI":"10.48550\/ARXIV.2402.16267"},{"key":"2069_CR36","unstructured":"Dong, Y., Li, G., Tao, Y., Jiang, X., Zhang, K., Li, J., Deng, J., Su, J., Zhang, J., Xu, J.: FAN: fourier analysis networks (2025). arXiv:2410.02675"},{"key":"2069_CR37","doi-asserted-by":"crossref","unstructured":"Tian, R., Wu, Z., Dai, Q., Hu, H., Qiao, Y., Jiang, Y.-G.: ResFormer: scaling ViTs with multi-resolution training (2023). arXiv:2212.00776","DOI":"10.1109\/CVPR52729.2023.02176"},{"key":"2069_CR38","doi-asserted-by":"crossref","unstructured":"Zhu, J., Chen, X., He, K., LeCun, Y., Liu, Z.: Transformers without normalization (2025). arXiv:2503.10622","DOI":"10.1109\/CVPR52734.2025.01388"},{"key":"2069_CR39","doi-asserted-by":"publisher","unstructured":"Li, J., Wen, Y., He, L.: Scconv: Spatial and channel reconstruction convolution for feature redundancy. In: 2023 IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp. 6153\u20136162 (2023). https:\/\/doi.org\/10.1109\/CVPR52729.2023.00596","DOI":"10.1109\/CVPR52729.2023.00596"},{"key":"2069_CR40","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2024.3374294","volume":"73","author":"G Xiao","year":"2024","unstructured":"Xiao, G., Tang, Z., Guo, H., Yu, J., Shen, H.T.: Fafusion: learning for infrared and visible image fusion via frequency awareness. IEEE Trans. Instrum. Meas. 73, 1\u201311 (2024). https:\/\/doi.org\/10.1109\/TIM.2024.3374294","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"2069_CR41","doi-asserted-by":"publisher","first-page":"79","DOI":"10.1016\/j.inffus.2022.03.007","volume":"83\u201384","author":"L Tang","year":"2022","unstructured":"Tang, L., Yuan, J., Zhang, H., Jiang, X., Ma, J.: Piafusion: a progressive infrared and visible image fusion network based on illumination aware. Inf. Fusion 83\u201384, 79\u201392 (2022). https:\/\/doi.org\/10.1016\/j.inffus.2022.03.007","journal-title":"Inf. Fusion"},{"issue":"3","key":"2069_CR42","doi-asserted-by":"publisher","first-page":"1712","DOI":"10.1109\/TCSVT.2023.3296745","volume":"34","author":"Y Zhao","year":"2024","unstructured":"Zhao, Y., Zheng, Q., Zhu, P., Zhang, X., Ma, W.: Tufusion: a transformer-based universal fusion algorithm for multimodal images. IEEE Trans. Circuits Syst. Video Technol. 34(3), 1712\u20131725 (2024). https:\/\/doi.org\/10.1109\/TCSVT.2023.3296745","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"2069_CR43","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2024.128116","volume":"600","author":"G Yang","year":"2024","unstructured":"Yang, G., Li, J., Lei, H., Gao, X.: A multi-scale information integration framework for infrared and visible image fusion. Neurocomputing 600, 128116 (2024). https:\/\/doi.org\/10.1016\/j.neucom.2024.128116","journal-title":"Neurocomputing"},{"key":"2069_CR44","doi-asserted-by":"publisher","first-page":"188","DOI":"10.1016\/j.neunet.2021.01.021","volume":"137","author":"C Peng","year":"2021","unstructured":"Peng, C., Tian, T., Chen, C., Guo, X., Ma, J.: Bilateral attention decoder: a lightweight decoder for real-time semantic segmentation. Neural Netw. 137, 188\u2013199 (2021). https:\/\/doi.org\/10.1016\/j.neunet.2021.01.021","journal-title":"Neural Netw."}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-02069-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-025-02069-w","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-02069-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,11]],"date-time":"2026-02-11T04:19:45Z","timestamp":1770783585000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-025-02069-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,4]]},"references-count":44,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,2]]}},"alternative-id":["2069"],"URL":"https:\/\/doi.org\/10.1007\/s00530-025-02069-w","relation":{"has-preprint":[{"id-type":"doi","id":"10.21203\/rs.3.rs-6809147\/v1","asserted-by":"object"}]},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,12,4]]},"assertion":[{"value":"3 June 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 October 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 December 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no Conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"12"}}