{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,7]],"date-time":"2025-11-07T01:09:19Z","timestamp":1762477759726,"version":"build-2065373602"},"reference-count":54,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T00:00:00Z","timestamp":1755734400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T00:00:00Z","timestamp":1755734400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2025,10]]},"DOI":"10.1007\/s00530-025-01903-5","type":"journal-article","created":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T10:48:28Z","timestamp":1755773308000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Multi-focus image fusion based on re-parameterized large kernel convolution and edge information fusion"],"prefix":"10.1007","volume":"31","author":[{"given":"Qing","family":"Li","sequence":"first","affiliation":[]},{"given":"Hao","family":"Zhai","sequence":"additional","affiliation":[]},{"given":"You","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Xiaoning","family":"Sun","sequence":"additional","affiliation":[]},{"given":"Long","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,8,21]]},"reference":[{"key":"1903_CR1","doi-asserted-by":"crossref","first-page":"100","DOI":"10.1016\/j.inffus.2016.05.004","volume":"33","author":"S Li","year":"2017","unstructured":"Li, S., Kang, X., Fang, L., Jianwen, H., Yin, H.: Pixel-level image fusion: A survey of the state of the art. Inf. Fusion 33, 100\u2013112 (2017)","journal-title":"Inf. Fusion"},{"key":"1903_CR2","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1109\/TGRS.2021.3090410","volume":"60","author":"UA Bhatti","year":"2022","unstructured":"Bhatti, U.A., Yu, Z., Chanussot, J., Zeeshan, Z., Yuan, L., Luo, W., Nawaz, S.A., Bhatti, M.A., Ain, Q.U., Mehmood, A.: Local similarity-based spatial-spectral fusion hyperspectral image classification with deep cnn and gabor filtering. IEEE Trans. Geosci. Remote Sens. 60, 1\u201315 (2022)","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"issue":"9","key":"1903_CR3","doi-asserted-by":"crossref","first-page":"1855","DOI":"10.1016\/j.patcog.2004.03.010","volume":"37","author":"Gonzalo Pajares and Jes\u00fas Manuel de la Cruz","year":"2004","unstructured":"Gonzalo Pajares and Jes\u00fas Manuel de la Cruz: A wavelet-based image fusion tutorial. Pattern Recogn. 37(9), 1855\u20131872 (2004)","journal-title":"Pattern Recogn."},{"key":"1903_CR4","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2022.116737","volume":"197","author":"D Tang","year":"2022","unstructured":"Tang, D., Xiong, Q., Yin, H., Zhu, Z., Li, Y.: A novel sparse representation based fusion approach for multi-focus images. Expert Syst. Appl. 197, 116737 (2022)","journal-title":"Expert Syst. Appl."},{"issue":"2","key":"1903_CR5","doi-asserted-by":"crossref","first-page":"136","DOI":"10.1016\/j.inffus.2012.01.007","volume":"14","author":"I De","year":"2013","unstructured":"De, I., Chanda, B.: Multi-focus image fusion using a morphology-based focus measure in a quad-tree structure. Inf. Fusion 14(2), 136\u2013146 (2013)","journal-title":"Inf. Fusion"},{"issue":"16","key":"1903_CR6","doi-asserted-by":"crossref","first-page":"1948","DOI":"10.1016\/j.patrec.2006.05.004","volume":"27","author":"M Li","year":"2006","unstructured":"Li, M., Cai, W., Tan, Z.: A region-based multi-sensor image fusion scheme using pulse-coupled neural network. Pattern Recogn. Lett. 27(16), 1948\u20131956 (2006)","journal-title":"Pattern Recogn. Lett."},{"key":"1903_CR7","doi-asserted-by":"crossref","first-page":"125","DOI":"10.1016\/j.ins.2017.12.043","volume":"433\u2013434","author":"H Tang","year":"2018","unstructured":"Tang, H., Xiao, B., Li, W., Wang, G.: Pixel convolutional neural network for multi-focus image fusion. Inf. Sci. 433\u2013434, 125\u2013141 (2018)","journal-title":"Inf. Sci."},{"issue":"2","key":"1903_CR8","doi-asserted-by":"crossref","first-page":"285","DOI":"10.1109\/TMM.2019.2928516","volume":"22","author":"B Xiao","year":"2020","unstructured":"Xiao, B., Ou, G., Tang, H., Bi, X., Li, W.: Multi-focus image fusion by hessian matrix based decomposition. IEEE Trans. Multimedia 22(2), 285\u2013297 (2020)","journal-title":"IEEE Trans. Multimedia"},{"key":"1903_CR9","doi-asserted-by":"crossref","first-page":"139","DOI":"10.1016\/j.inffus.2014.05.004","volume":"23","author":"Yu Liu","year":"2015","unstructured":"Liu, Yu., Liu, S., Wang, Z.: Multi-focus image fusion with dense sift. Inf. Fusion 23, 139\u2013155 (2015)","journal-title":"Inf. Fusion"},{"key":"1903_CR10","doi-asserted-by":"crossref","first-page":"62","DOI":"10.1016\/j.optlastec.2018.07.045","volume":"110","author":"W Yin","year":"2019","unstructured":"Yin, W., Zhao, W., You, D., Wang, D.: Local binary pattern metric-based multi-focus image fusion. Opt. Laser Technol 110, 62\u201368 (2019)","journal-title":"Opt. Laser Technol"},{"key":"1903_CR11","volume":"100","author":"S Liu","year":"2022","unstructured":"Liu, S., Ma, J., Yang, Y., Qiu, T., Li, H., Shaohai, H., Zhang, Y.: A multi-focus color image fusion algorithm based on low vision image reconstruction and focused feature extraction. Signal Process. Image Commun 100, 116533 (2022)","journal-title":"Signal Process. Image Commun"},{"key":"1903_CR12","doi-asserted-by":"crossref","first-page":"40","DOI":"10.1016\/j.inffus.2020.08.022","volume":"66","author":"H Zhang","year":"2021","unstructured":"Zhang, H., Le, Z., Shao, Z., Han, X., Ma, J.: Mff-gan: An unsupervised generative adversarial network with adaptive and gradient joint constraints for multi-focus image fusion. Inf. Fusion 66, 40\u201353 (2021)","journal-title":"Inf. Fusion"},{"issue":"7","key":"1903_CR13","doi-asserted-by":"crossref","first-page":"1200","DOI":"10.1109\/JAS.2022.105686","volume":"9","author":"J Ma","year":"2022","unstructured":"Ma, J., Tang, L., Fan, F., Huang, J., Mei, X., Ma, Y.: Swinfusion: Cross-domain long-range learning for general image fusion via swin transformer. IEEE\/CAA J. Autom. Sin 9(7), 1200\u20131217 (2022)","journal-title":"IEEE\/CAA J. Autom. Sin"},{"key":"1903_CR14","doi-asserted-by":"crossref","first-page":"191","DOI":"10.1016\/j.inffus.2016.12.001","volume":"36","author":"Yu Liu","year":"2017","unstructured":"Liu, Yu., Xun, C., Hu, P., Zengfu, W.: Multi-focus image fusion with a deep convolutional neural network. Inf. Fusion 36, 191\u2013207 (2017)","journal-title":"Inf. Fusion"},{"key":"1903_CR15","doi-asserted-by":"crossref","first-page":"2915","DOI":"10.1109\/TIP.2023.3276330","volume":"32","author":"O Bouzos","year":"2023","unstructured":"Bouzos, O., Andreadis, I., Mitianoudis, N.: A convolutional neural network-based conditional random field model for structured multi-focus image fusion robust to noise. IEEE Trans. Image Process. 32, 2915\u20132930 (2023)","journal-title":"IEEE Trans. Image Process."},{"key":"1903_CR16","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, \u0141, Polosukhin, I.: Attention is all you need. Adv. Neural Inf. Process Syst. 30, (2017)"},{"key":"1903_CR17","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Sylvain: An image is worth $$16\\times 16$$ words: Transformers for image recognition at scale (2020)"},{"key":"1903_CR18","volume":"76","author":"W Pan","year":"2023","unstructured":"Pan, W., Jiang, L., Hua, Z., Li, J.: Multi-focus image fusion: Transformer and shallow feature attention matters. Displays 76, 102353 (2023)","journal-title":"Displays"},{"key":"1903_CR19","first-page":"2023","volume":"1\u201316","author":"H Zhai","year":"2023","unstructured":"Zhai, H., Pan, X., Yang, Y., Jiang, J., Li, Q.: Two-stage focus measurement network with joint boundary refinement for multifocus image fusion. Int. J. Intell. Syst. 1\u201316, 2023 (2023)","journal-title":"Int. J. Intell. Syst."},{"key":"1903_CR20","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2023.121156","volume":"235","author":"Z Duan","year":"2024","unstructured":"Duan, Z., Luo, X., Zhang, T.: Combining transformers with cnn for multi-focus image fusion. Expert Syst. Appl. 235, 121156 (2024)","journal-title":"Expert Syst. Appl."},{"issue":"2","key":"1903_CR21","first-page":"358","volume":"23","author":"N Kanopoulos","year":"1988","unstructured":"Kanopoulos, N., Vasanthavada, N., Robert, L.: Solid-state circuits Baker. Design of an image edge detection filter using the sobel operator 23(2), 358\u2013367 (1988)","journal-title":"Design of an image edge detection filter using the sobel operator"},{"issue":"1","key":"1903_CR22","first-page":"15","volume":"10","author":"MS Judith","year":"1970","unstructured":"Judith, M.S.: Object enhancement and extraction. 10(1), 15\u201319 (1970)","journal-title":"Object enhancement and extraction."},{"issue":"11","key":"1903_CR23","doi-asserted-by":"crossref","first-page":"1985","DOI":"10.1109\/LSP.2015.2449878","volume":"22","author":"SK Md","year":"2015","unstructured":"Md, S.K., Appina, B., Channappayya, S.S.: Full-reference stereo image quality assessment using natural stereo scene statistics. IEEE Signal Process. Lett. 22(11), 1985\u20139 (2015)","journal-title":"IEEE Signal Process. Lett."},{"issue":"6","key":"1903_CR24","doi-asserted-by":"crossref","first-page":"1391","DOI":"10.1109\/TGRS.2005.846874","volume":"43","author":"Z Wang","year":"2005","unstructured":"Wang, Z., Ziou, D., Armenakis, C., Li, D.: A comparative analysis of image fusion methods. IEEE Trans. Geosci. Remote Sens. 43(6), 1391\u20131402 (2005)","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"1903_CR25","doi-asserted-by":"crossref","first-page":"4","DOI":"10.1016\/j.inffus.2013.12.002","volume":"19","author":"AP James","year":"2014","unstructured":"James, A.P., Dasarathy, B.V.: Medical image fusion: A survey of the state of the art. Inf. Fusion 19, 4\u201319 (2014)","journal-title":"Inf. Fusion"},{"issue":"7","key":"1903_CR26","doi-asserted-by":"crossref","first-page":"2864","DOI":"10.1109\/TIP.2013.2244222","volume":"22","author":"L Shutao","year":"2013","unstructured":"Shutao, L., Xudong, K., Jianwen, H.: Image fusion with guided filtering. IEEE Trans. Image Process. 22(7), 2864\u20132875 (2013)","journal-title":"IEEE Trans. Image Process."},{"key":"1903_CR27","volume":"237","author":"Yu Zhang","year":"2024","unstructured":"Zhang, Yu., Chen, J., Ma, X., Wang, G., Bhatti, U.A., Huang, M.: Interactive medical image annotation using improved attention u-net with compound geodesic distance. Expert Syst. Appl. 237, 121282 (2024)","journal-title":"Expert Syst. Appl."},{"key":"1903_CR28","unstructured":"Ding, X., Zhang, X., Han, J., Ding G.: Scaling up your kernels to 31x31: Revisiting large kernel design in cnns. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pages 11963\u201311975"},{"key":"1903_CR29","unstructured":"Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., Lin, S., Guo, B.: Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF international conference on computer vision, pages 10012\u201310022"},{"issue":"4","key":"1903_CR30","doi-asserted-by":"crossref","first-page":"733","DOI":"10.1007\/s41095-023-0364-2","volume":"9","author":"M-H Guo","year":"2023","unstructured":"Guo, M.-H., Cheng-Ze, L., Liu, Z.-N., Cheng, M.-M., Shi-Min, H.: Visual attention network. Comput. visual media 9(4), 733\u2013752 (2023)","journal-title":"Visual attention network. Comput. visual media"},{"key":"1903_CR31","unstructured":"Liu, S., Chen, T., Chen, X., Chen, X., Xiao, Q., Wu, B., K\u00e4rkk\u00e4inen, T., Pechenizkiy, M., Mocanu, D., Wang, Z.: More convnets in the 2020s: Scaling up kernels beyond 51x51 using sparsity. arXiv:2207.03620 (2022)"},{"key":"1903_CR32","doi-asserted-by":"crossref","unstructured":"Ding, X., Zhang, Y., Ge, Y., Zhao, S., Song, L., Yue, X., Unireplknet, Y.: A universal perception large-kernel convnet for audio, video, point cloud, time-series and image recognition (2023)","DOI":"10.1109\/CVPR52733.2024.00527"},{"issue":"4","key":"1903_CR33","doi-asserted-by":"crossref","first-page":"308","DOI":"10.1049\/el:20000267","volume":"36","author":"CS Xydeas","year":"2000","unstructured":"Xydeas, C.S.: Vladimir: Objective image fusion performance measure. Electr. Lett. 36(4), 308\u2013309 (2000)","journal-title":"Electr. Lett."},{"issue":"10","key":"1903_CR34","doi-asserted-by":"crossref","first-page":"1421","DOI":"10.1016\/j.imavis.2007.12.002","volume":"27","author":"Y Chen","year":"2009","unstructured":"Chen, Y., Blum, R.S.: A new automated quality assessment algorithm for image fusion. Image Vision Comput. 27(10), 1421\u20131432 (2009)","journal-title":"Image Vision Comput."},{"issue":"3\u20134","key":"1903_CR35","first-page":"287","volume":"200","author":"Q Wang","year":"2005","unstructured":"Wang, Q., Shen, Y., Qiu, J.: A nonlinear correlation measure for multivariable data set. Physica D: Nonlinear Phenomena. 200(3\u20134), 287\u2013295 (2005)","journal-title":"Physica D: Nonlinear Phenomena."},{"key":"1903_CR36","doi-asserted-by":"crossref","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error measurement to structural similarity. IEEE Trans. Image Process. 13, 600\u2013613 (2004)","journal-title":"IEEE Trans. Image Process."},{"issue":"18","key":"1903_CR37","doi-asserted-by":"crossref","first-page":"1066","DOI":"10.1049\/el:20081754","volume":"44","author":"M Hossny","year":"2008","unstructured":"Hossny, M., Nahavandi, S., Creighton, D.: Comments on \u2018information measure for performance of image fusion\u2019. Electr Lett. 44(18), 1066\u20131067 (2008)","journal-title":"Electr Lett."},{"key":"1903_CR38","unstructured":"Xydeas, C. S., Petrovic, V. S.: Objective pixel-level image fusion performance measure. In: Sensor Fusion: Architectures, Algorithms, and Applications IV, volume 4051, pages 89\u201398. SPIE"},{"issue":"4","key":"1903_CR39","doi-asserted-by":"crossref","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004)","journal-title":"IEEE Trans. Image Process."},{"issue":"2","key":"1903_CR40","doi-asserted-by":"crossref","first-page":"127","DOI":"10.1016\/j.inffus.2011.08.002","volume":"14","author":"Yu Han","year":"2013","unstructured":"Han, Yu., Cai, Y., Cao, Y., Xu, X.: A new image fusion performance metric based on visual information fidelity. Inf. Fusion. 14(2), 127\u2013135 (2013)","journal-title":"Inf. Fusion."},{"key":"1903_CR41","doi-asserted-by":"crossref","first-page":"35","DOI":"10.1016\/j.image.2018.12.004","volume":"72","author":"X Qiu","year":"2019","unstructured":"Qiu, X., Li, M., Zhang, L., Yuan, X.I.C.: Guided filter-based multi-focus image fusion through focus region detection. Signal Process. Image Commun. 72, 35\u201346 (2019)","journal-title":"Signal Process. Image Commun."},{"key":"1903_CR42","doi-asserted-by":"crossref","first-page":"5576","DOI":"10.1007\/s00034-019-01131-z","volume":"38","author":"DP Bavirisetti","year":"2019","unstructured":"Bavirisetti, D.P., Xiao, G., Zhao, J., Dhuli, R., Liu, G.: Multi-scale guided image and video fusion: A fast and efficient approach. Circuits Syst. Signal Process. 38, 5576\u20135605 (2019)","journal-title":"Circuits Syst. Signal Process."},{"key":"1903_CR43","doi-asserted-by":"crossref","first-page":"4816","DOI":"10.1109\/TIP.2020.2976190","volume":"29","author":"J Li","year":"2020","unstructured":"Li, J., Guo, X., Guangming, L., Zhang, B., Yong, X., Feng, W., Drpl, D.: Deep regression pair learning for multi-focus image fusion. IEEE Trans. Image Process. 29, 4816\u20134831 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"1903_CR44","doi-asserted-by":"crossref","first-page":"99","DOI":"10.1016\/j.inffus.2019.07.011","volume":"54","author":"Y Zhang","year":"2020","unstructured":"Zhang, Y., Liu, Y., Sun, P., Yan, H., Zhao, X., Zhang, L.: IFCNN: A general image fusion framework based on convolutional neural network. Inf. Fusion 54, 99\u2013118 (2020)","journal-title":"Inf. Fusion"},{"issue":"1","key":"1903_CR45","first-page":"502","volume":"44","author":"X Han","year":"2020","unstructured":"Han, X., Ma, J., Jiang, J., Guo, X., Ling, H.: U2fusion: A unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 44(1), 502\u2013518 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1903_CR46","doi-asserted-by":"crossref","first-page":"26316","DOI":"10.1109\/ACCESS.2020.2971137","volume":"8","author":"H Xu","year":"2020","unstructured":"Xu, H., Fan, F., Zhang, H., Le, Z., Huang, J.: A deep model for multi-focus image fusion based on gradients and connected regions. IEEE Access 8, 26316\u201326327 (2020)","journal-title":"IEEE Access"},{"key":"1903_CR47","doi-asserted-by":"crossref","first-page":"5793","DOI":"10.1007\/s00521-020-05358-9","volume":"33","author":"B Ma","year":"2021","unstructured":"Ma, B., Zhu, Y., Yin, X., Ban, X., Huang, H., Michele, M.: Sesf-fuse: An unsupervised deep model for multi-focus image fusion. Neural Comput. Appl. 33, 5793\u20135804 (2021)","journal-title":"Neural Comput. Appl."},{"key":"1903_CR48","doi-asserted-by":"crossref","first-page":"204","DOI":"10.1016\/j.neucom.2021.10.115","volume":"470","author":"B Ma","year":"2022","unstructured":"Ma, B., Yin, X., Wu, D., Shen, H., Ban, X., Wang, Y.: End-to-end learning for simultaneously generating decision map and multi-focus image fusion result. Neurocomputing 470, 204\u2013216 (2022)","journal-title":"Neurocomputing"},{"key":"1903_CR49","doi-asserted-by":"crossref","first-page":"127","DOI":"10.1016\/j.inffus.2022.11.014","volume":"92","author":"H Xingyu","year":"2023","unstructured":"Xingyu, H., Jiang, J., Liu, X., Zmff, J.: Zero-shot multi-focus image fusion. Inf. Fusion 92, 127\u2013138 (2023)","journal-title":"Inf. Fusion"},{"key":"1903_CR50","volume":"238","author":"M Li","year":"2024","unstructured":"Li, M., Pei, R., Zheng, T., Zhang, Y., Fu, W.: Fusiondiff: Multi-focus image fusion using denoising diffusion probabilistic models. Expert Syst. Appl. 238, 121664 (2024)","journal-title":"Expert Syst. Appl."},{"issue":"9","key":"1903_CR51","first-page":"4819","volume":"44","author":"X Zhang","year":"2021","unstructured":"Zhang, X.: Deep learning-based multi-focus image fusion: A survey and a comparative study. IEEE Trans. Pattern Anal. Mach. Intell. 44(9), 4819\u201338 (2021)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1903_CR52","first-page":"12638","volume":"33","author":"Q Cai","year":"2020","unstructured":"Cai, Q., Wang, Y., Yao, Y.P., Mei, T.: Joint contrastive learning with infinite possibilities. Adv. Neural. Inf. Process. Syst. 33, 12638\u201312648 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"1903_CR53","doi-asserted-by":"crossref","unstructured":"Wu, H., Xiao, B., Codella, N., Liu, M., Dai, X., Yuan, L., Zhang, L.: Cvt: Introducing convolutions to vision transformers. In: Proceedings of the IEEE\/CVF international conference on computer vision (pp. 22\u201331) (2021)","DOI":"10.1109\/ICCV48922.2021.00009"},{"issue":"2","key":"1903_CR54","doi-asserted-by":"crossref","first-page":"1489","DOI":"10.1109\/TPAMI.2022.3164083","volume":"45","author":"Y Li","year":"2023","unstructured":"Li, Y., Yao, T., Pan, Y., Mei, T.: Contextual transformer networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 45(2), 1489\u20131500 (2023)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-01903-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-025-01903-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-01903-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,7]],"date-time":"2025-11-07T01:03:26Z","timestamp":1762477406000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-025-01903-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,21]]},"references-count":54,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2025,10]]}},"alternative-id":["1903"],"URL":"https:\/\/doi.org\/10.1007\/s00530-025-01903-5","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"type":"print","value":"0942-4962"},{"type":"electronic","value":"1432-1882"}],"subject":[],"published":{"date-parts":[[2025,8,21]]},"assertion":[{"value":"24 September 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 June 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 August 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"All the authors of this paper have no Conflict of interest, financial or otherwise.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"326"}}