{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,31]],"date-time":"2026-03-31T11:42:35Z","timestamp":1774957355844,"version":"3.50.1"},"reference-count":54,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2023,6,9]],"date-time":"2023-06-09T00:00:00Z","timestamp":1686268800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,6,9]],"date-time":"2023-06-09T00:00:00Z","timestamp":1686268800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61801190"],"award-info":[{"award-number":["61801190"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"National Key Research and Development Project of China","award":["2019YFC0409105"],"award-info":[{"award-number":["2019YFC0409105"]}]},{"name":"Thirteenth Five-Year Plan Scientific Research Planning Project of Education Department of Jilin Province","award":["JJKH20200997KJ"],"award-info":[{"award-number":["JJKH20200997KJ"]}]},{"name":"Thirteenth Five-Year Plan Scientific Research Planning Project of Education Department of Jilin Province","award":["JJKH20200678KJ"],"award-info":[{"award-number":["JJKH20200678KJ"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Comput Vis"],"published-print":{"date-parts":[[2023,10]]},"DOI":"10.1007\/s11263-023-01806-w","type":"journal-article","created":{"date-parts":[[2023,6,9]],"date-time":"2023-06-09T07:02:01Z","timestamp":1686294121000},"page":"2529-2552","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":37,"title":["When Multi-Focus Image Fusion Networks Meet Traditional Edge-Preservation Technology"],"prefix":"10.1007","volume":"131","author":[{"given":"Zeyu","family":"Wang","sequence":"first","affiliation":[]},{"given":"Xiongfei","family":"Li","sequence":"additional","affiliation":[]},{"given":"Libo","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Haoran","family":"Duan","sequence":"additional","affiliation":[]},{"given":"Shidong","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Hao","family":"Liu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8412-4956","authenticated-orcid":false,"given":"Xiaoli","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,6,9]]},"reference":[{"key":"1806_CR1","doi-asserted-by":"crossref","unstructured":"Agustsson, E., Timofte, R. (2017). Ntire 2017 challenge on single image super-resolution: Dataset and study. In: The IEEE conference on computer vision and pattern recognition (CVPR) Workshops, pp 1\u20138.","DOI":"10.1109\/CVPRW.2017.150"},{"key":"1806_CR2","doi-asserted-by":"publisher","first-page":"105","DOI":"10.1016\/j.inffus.2014.05.003","volume":"22","author":"X Bai","year":"2015","unstructured":"Bai, X., Zhang, Y., Zhou, F., et al. (2015). Quadtree-based multi-focus image fusion using a weighted focus-measure. Information Fusion, 22, 105\u2013118.","journal-title":"Information Fusion"},{"issue":"6","key":"1806_CR3","doi-asserted-by":"publisher","first-page":"844","DOI":"10.1109\/TPAMI.2002.1008390","volume":"24","author":"D Barash","year":"2002","unstructured":"Barash, D. (2002). Fundamental relationship between bilateral filtering, adaptive smoothing, and the nonlinear diffusion equation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 24(6), 844\u2013847.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"issue":"8","key":"1806_CR4","doi-asserted-by":"publisher","first-page":"5735","DOI":"10.1007\/s10462-021-09961-7","volume":"54","author":"S Bhat","year":"2021","unstructured":"Bhat, S., & Koundal, D. (2021). Multi-focus image fusion techniques: a survey. Artificial Intelligence Review, 54(8), 5735\u20135787.","journal-title":"Artificial Intelligence Review"},{"issue":"10","key":"1806_CR5","doi-asserted-by":"publisher","first-page":"1421","DOI":"10.1016\/j.imavis.2007.12.002","volume":"27","author":"Y Chen","year":"2009","unstructured":"Chen, Y., & Blum, R. S. (2009). A new automated quality assessment algorithm for image fusion. Image and Vision Computing, 27(10), 1421\u20131432.","journal-title":"Image and Vision Computing"},{"key":"1806_CR6","doi-asserted-by":"crossref","unstructured":"Chen, Z., Wang, D., Gong, S., & et\u00a0al. (2017). Application of multi-focus image fusion in visual power patrol inspection. In: 2017 IEEE 2nd Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), IEEE, pp 1688\u20131692.","DOI":"10.1109\/IAEAC.2017.8054302"},{"issue":"11","key":"1806_CR7","doi-asserted-by":"publisher","first-page":"626","DOI":"10.1049\/el:20060693","volume":"42","author":"N Cvejic","year":"2006","unstructured":"Cvejic, N., Canagarajah, C., & Bull, D. (2006). Image fusion metric based on mutual information and tsallis entropy. Electronics Letters, 42(11), 626\u2013627.","journal-title":"Electronics Letters"},{"key":"1806_CR8","doi-asserted-by":"publisher","first-page":"96","DOI":"10.1016\/j.inffus.2018.01.009","volume":"45","author":"MS Farid","year":"2019","unstructured":"Farid, M. S., Mahmood, A., & Al-Maadeed, S. A. (2019). Multi-focus image fusion using content adaptive blurring. Information Fusion, 45, 96\u2013112.","journal-title":"Information Fusion"},{"key":"1806_CR9","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2010324.1964964","volume":"1","author":"ES Gastal","year":"2011","unstructured":"Gastal, E. S., & Oliveira, M. M. (2011). Domain transform for edge-aware image and video processing. ACM SIGGRAPH, 1, 1\u201312.","journal-title":"ACM SIGGRAPH"},{"key":"1806_CR10","doi-asserted-by":"crossref","unstructured":"Kim, J., Lee, JK., Lee, KM. (2016). Accurate image super-resolution using very deep convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1646\u20131654.","DOI":"10.1109\/CVPR.2016.182"},{"key":"1806_CR11","doi-asserted-by":"publisher","first-page":"4816","DOI":"10.1109\/TIP.2020.2976190","volume":"29","author":"J Li","year":"2020","unstructured":"Li, J., Guo, X., Lu, G., et al. (2020). Drpl: Deep regression pair learning for multi-focus image fusion. IEEE Transactions on Image Processing, 29, 4816\u20134831.","journal-title":"IEEE Transactions on Image Processing"},{"key":"1806_CR12","doi-asserted-by":"publisher","first-page":"100","DOI":"10.1016\/j.inffus.2016.05.004","volume":"33","author":"S Li","year":"2017","unstructured":"Li, S., Kang, X., Fang, L., et al. (2017). Pixel-level image fusion: A survey of the state of the art. Information Fusion, 33, 100\u2013112.","journal-title":"Information Fusion"},{"key":"1806_CR13","doi-asserted-by":"publisher","first-page":"147","DOI":"10.1016\/j.inffus.2014.09.004","volume":"24","author":"Y Liu","year":"2015","unstructured":"Liu, Y., Liu, S., & Wang, Z. (2015). A general framework for image fusion based on multi-scale transform and sparse representation. Information Fusion, 24, 147\u2013164.","journal-title":"Information Fusion"},{"key":"1806_CR14","doi-asserted-by":"publisher","first-page":"191","DOI":"10.1016\/j.inffus.2016.12.001","volume":"36","author":"Y Liu","year":"2017","unstructured":"Liu, Y., Chen, X., Peng, H., et al. (2017). Multi-focus image fusion with a deep convolutional neural network. Information Fusion, 36, 191\u2013207.","journal-title":"Information Fusion"},{"key":"1806_CR15","doi-asserted-by":"publisher","first-page":"158","DOI":"10.1016\/j.inffus.2017.10.007","volume":"42","author":"Y Liu","year":"2018","unstructured":"Liu, Y., Chen, X., Wang, Z., et al. (2018). Deep learning for pixel-level image fusion: Recent advances and future prospects. Information Fusion, 42, 158\u2013173.","journal-title":"Information Fusion"},{"key":"1806_CR16","doi-asserted-by":"publisher","first-page":"71","DOI":"10.1016\/j.inffus.2020.06.013","volume":"64","author":"Y Liu","year":"2020","unstructured":"Liu, Y., Wang, L., Cheng, J., et al. (2020). Multi-focus image fusion: A survey of the state of the art. Information Fusion, 64, 71\u201391.","journal-title":"Information Fusion"},{"issue":"1","key":"1806_CR17","doi-asserted-by":"publisher","first-page":"94","DOI":"10.1109\/TPAMI.2011.109","volume":"34","author":"Z Liu","year":"2011","unstructured":"Liu, Z., Blasch, E., Xue, Z., et al. (2011). Objective assessment of multiresolution image fusion algorithms for context enhancement in night vision: a comparative study. IEEE Transactions on Pattern Analysis and Machine Intelligence, 34(1), 94\u2013109.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"1806_CR18","doi-asserted-by":"publisher","first-page":"102","DOI":"10.1016\/j.inffus.2016.09.007","volume":"35","author":"Z Liu","year":"2017","unstructured":"Liu, Z., Chai, Y., Yin, H., et al. (2017). A novel multi-focus image fusion approach based on image decomposition. Information Fusion, 35, 102\u2013116.","journal-title":"Information Fusion"},{"issue":"168","key":"1806_CR19","first-page":"914","volume":"258","author":"Y Luo","year":"2022","unstructured":"Luo, Y., He, K., Xu, D., et al. (2022). Infrared and visible image fusion based on visibility enhancement and hybrid multiscale decomposition. Optik, 258(168), 914.","journal-title":"Optik"},{"issue":"11","key":"1806_CR20","doi-asserted-by":"publisher","first-page":"5793","DOI":"10.1007\/s00521-020-05358-9","volume":"33","author":"B Ma","year":"2021","unstructured":"Ma, B., Zhu, Y., Yin, X., et al. (2021). Sesf-fuse: An unsupervised deep model for multi-focus image fusion. Neural Computing and Applications, 33(11), 5793\u20135804.","journal-title":"Neural Computing and Applications"},{"key":"1806_CR21","doi-asserted-by":"publisher","first-page":"204","DOI":"10.1016\/j.neucom.2021.10.115","volume":"470","author":"B Ma","year":"2022","unstructured":"Ma, B., Yin, X., Wu, D., et al. (2022). End-to-end learning for simultaneously generating decision map and multi-focus image fusion result. Neurocomputing, 470, 204\u2013216.","journal-title":"Neurocomputing"},{"key":"1806_CR22","doi-asserted-by":"publisher","first-page":"8668","DOI":"10.1109\/TIP.2020.3018261","volume":"29","author":"H Ma","year":"2020","unstructured":"Ma, H., Liao, Q., Zhang, J., et al. (2020). An $$\\alpha $$-matte boundary defocus model-based cascaded network for multi-focus image fusion. IEEE Transactions on Image Processing, 29, 8668\u20138679.","journal-title":"IEEE Transactions on Image Processing"},{"key":"1806_CR23","doi-asserted-by":"publisher","first-page":"119","DOI":"10.1016\/j.inffus.2018.07.010","volume":"48","author":"B Meher","year":"2019","unstructured":"Meher, B., Agrawal, S., Panda, R., et al. (2019). A survey on region based image fusion methods. Information Fusion, 48, 119\u2013132.","journal-title":"Information Fusion"},{"key":"1806_CR24","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.inffus.2014.10.004","volume":"25","author":"M Nejati","year":"2015","unstructured":"Nejati, M., Samavi, S., & Shirani, S. (2015). Multi-focus image fusion using dictionary-based sparse representation. Information Fusion, 25, 72\u201384.","journal-title":"Information Fusion"},{"issue":"107","key":"1806_CR25","first-page":"404","volume":"106","author":"X Qin","year":"2020","unstructured":"Qin, X., Zhang, Z., Huang, C., et al. (2020). U2-net: Going deeper with nested u-structure for salient object detection. Pattern Recognition, 106(107), 404.","journal-title":"Pattern Recognition"},{"key":"1806_CR26","first-page":"35","volume":"72","author":"X Qiu","year":"2019","unstructured":"Qiu, X., Li, M., Zhang, L., et al. (2019). Guided filter-based multi-focus image fusion through focus region detection. Signal Processing: Image Communication, 72, 35\u201346.","journal-title":"Signal Processing: Image Communication"},{"issue":"7","key":"1806_CR27","doi-asserted-by":"publisher","first-page":"313","DOI":"10.1049\/el:20020212","volume":"38","author":"G Qu","year":"2002","unstructured":"Qu, G., Zhang, D., & Yan, P. (2002). Information measure for performance of image fusion. Electronics Letters, 38(7), 313\u2013315.","journal-title":"Electronics Letters"},{"key":"1806_CR28","first-page":"52","volume":"2","author":"B Rajalingam","year":"2018","unstructured":"Rajalingam, B., & Priya, R. (2018). Hybrid multimodality medical image fusion technique for feature enhancement in medical diagnosis. International Journal of Engineering Science Invention, 2, 52\u201360.","journal-title":"International Journal of Engineering Science Invention"},{"issue":"11","key":"1806_CR29","doi-asserted-by":"publisher","first-page":"2629","DOI":"10.1049\/ipr2.12251","volume":"15","author":"AA Smadi","year":"2021","unstructured":"Smadi, A. A., Yang, S., Mehmood, A., et al. (2021). Smart pansharpening approach using kernel-based image filtering. IET Image Processing, 15(11), 2629\u20132642.","journal-title":"IET Image Processing"},{"key":"1806_CR30","doi-asserted-by":"publisher","first-page":"295","DOI":"10.1016\/j.procs.2021.12.243","volume":"198","author":"AA Smadi","year":"2022","unstructured":"Smadi, A. A., Abugabah, A., Mehmood, A., et al. (2022). Brain image fusion approach based on side window filtering. Procedia Computer Science, 198, 295\u2013300.","journal-title":"Procedia Computer Science"},{"issue":"102","key":"1806_CR31","first-page":"280","volume":"64","author":"W Tan","year":"2021","unstructured":"Tan, W., Thit\u00f8n, W., Xiang, P., et al. (2021). Multi-modal brain image fusion based on multi-level edge-preserving filtering. Biomedical Signal Processing and Control, 64(102), 280.","journal-title":"Biomedical Signal Processing and Control"},{"key":"1806_CR32","doi-asserted-by":"publisher","first-page":"125","DOI":"10.1016\/j.ins.2017.12.043","volume":"433","author":"H Tang","year":"2018","unstructured":"Tang, H., Xiao, B., Li, W., et al. (2018). Pixel convolutional neural network for multi-focus image fusion. Information Sciences, 433, 125\u2013141.","journal-title":"Information Sciences"},{"key":"1806_CR33","doi-asserted-by":"crossref","unstructured":"Wang, L., Lu, H., Wang, Y., et\u00a0al. (2017). Learning to detect salient objects with image-level supervision. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 136\u2013145.","DOI":"10.1109\/CVPR.2017.404"},{"key":"1806_CR34","unstructured":"Wang, Pw., Liu, B. (2008). A novel image fusion metric based on multi-scale analysis. In: 2008 9th International Conference on Signal Processing, IEEE, pp 965\u2013968."},{"key":"1806_CR35","doi-asserted-by":"crossref","unstructured":"Wang, Y., Wang, L., Yang, J., et\u00a0al. (2019a). Flickr1024: A large-scale dataset for stereo image super-resolution. In: International Conference on Computer Vision Workshops, pp 3852\u20133857.","DOI":"10.1109\/ICCVW.2019.00478"},{"issue":"108","key":"1806_CR36","first-page":"254","volume":"189","author":"Y Wang","year":"2021","unstructured":"Wang, Y., Li, X., Zhu, R., et al. (2021). A multi-focus image fusion framework based on multi-scale sparse representation in gradient domain. Signal Processing, 189(108), 254.","journal-title":"Signal Processing"},{"issue":"24","key":"1806_CR37","doi-asserted-by":"publisher","first-page":"34483","DOI":"10.1007\/s11042-019-08070-6","volume":"78","author":"Z Wang","year":"2019","unstructured":"Wang, Z., Li, X., Duan, H., et al. (2019). Multifocus image fusion using convolutional neural networks in the discrete wavelet transform domain. Multimedia Tools and Applications, 78(24), 34483\u201334512.","journal-title":"Multimedia Tools and Applications"},{"key":"1806_CR38","doi-asserted-by":"publisher","first-page":"4527","DOI":"10.1109\/TIP.2022.3184250","volume":"31","author":"Z Wang","year":"2022","unstructured":"Wang, Z., Li, X., Duan, H., et al. (2022). A self-supervised residual feature learning model for multifocus image fusion. IEEE Transactions on Image Processing, 31, 4527\u20134542.","journal-title":"IEEE Transactions on Image Processing"},{"key":"1806_CR39","unstructured":"Wang, Z., Li, X., Yu, S., et\u00a0al. (2022b). Vsp-fuse: Multifocus image fusion model using the knowledge transferred from visual salience priors. IEEE Transactions on Circuits and Systems for Video Technology pp 1\u201315"},{"key":"1806_CR40","doi-asserted-by":"publisher","first-page":"163","DOI":"10.1109\/TIP.2020.3033158","volume":"30","author":"B Xiao","year":"2020","unstructured":"Xiao, B., Xu, B., Bi, X., et al. (2020). Global-feature encoding u-net (geu-net) for multi-focus image fusion. IEEE Transactions on Image Processing, 30, 163\u2013175.","journal-title":"IEEE Transactions on Image Processing"},{"key":"1806_CR41","unstructured":"Xu, H., Ma, J., Jiang, J., et\u00a0al .(2020a). U2fusion: A unified unsupervised image fusion network. IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"1806_CR42","unstructured":"Xu, S., Wei, X., Zhang, C., et\u00a0al. (2020b). Mffw: A new dataset for multi-focus image fusion. arXiv preprint arXiv:2002.04780"},{"key":"1806_CR43","unstructured":"Xydeas, C. S., & Petrovic, V. S. (2000). Objective pixel-level image fusion performance measure. Sensor Fusion: Architectures, Algorithms, and Applications IV, 4051, 89\u201398."},{"key":"1806_CR44","doi-asserted-by":"crossref","unstructured":"Yang, C., Zhang, J. Q., Wang, X. R., et al. (2008). A novel similarity based quality metric for image fusion. Information Fusion, 9(2), 156\u2013160.","DOI":"10.1016\/j.inffus.2006.09.001"},{"key":"1806_CR45","doi-asserted-by":"crossref","unstructured":"Yu, S., Li, X., Ma, M., et al. (2021). Multi-focus image fusion based on l1 image transform. Multimedia Tools and Applications, 80(4), 5673\u20135700.","DOI":"10.1007\/s11042-020-09877-4"},{"key":"1806_CR46","first-page":"1","volume":"70","author":"Y Zang","year":"2021","unstructured":"Zang, Y., Zhou, D., Wang, C., et al. (2021). Ufa-fuse: A novel deep supervised and hybrid model for multifocus image fusion. IEEE Transactions on Instrumentation and Measurement, 70, 1\u201317.","journal-title":"IEEE Transactions on Instrumentation and Measurement"},{"key":"1806_CR47","doi-asserted-by":"publisher","first-page":"733","DOI":"10.1016\/j.neucom.2015.09.092","volume":"174","author":"B Zhang","year":"2016","unstructured":"Zhang, B., Lu, X., Pei, H., et al. (2016). Multi-focus image fusion algorithm based on focused region extraction. Neurocomputing, 174, 733\u2013748.","journal-title":"Neurocomputing"},{"issue":"10","key":"1806_CR48","doi-asserted-by":"publisher","first-page":"2761","DOI":"10.1007\/s11263-021-01501-8","volume":"129","author":"H Zhang","year":"2021","unstructured":"Zhang, H., & Ma, J. (2021). Sdnet: A versatile squeeze-and-decomposition network for real-time image fusion. International Journal of Computer Vision, 129(10), 2761\u20132785.","journal-title":"International Journal of Computer Vision"},{"key":"1806_CR49","unstructured":"Zhang, H., Xu, H., Xiao, Y., et\u00a0al. (2020a). Rethinking the image fusion: A fast unified image fusion network based on proportional maintenance of gradient and intensity. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp 12,797\u201312,804."},{"key":"1806_CR50","doi-asserted-by":"publisher","first-page":"323","DOI":"10.1016\/j.inffus.2021.06.008","volume":"76","author":"H Zhang","year":"2021","unstructured":"Zhang, H., Xu, H., Tian, X., et al. (2021). Image fusion meets deep learning: A survey and perspective. Information Fusion, 76, 323\u2013336.","journal-title":"Information Fusion"},{"issue":"5","key":"1806_CR51","doi-asserted-by":"publisher","first-page":"2045","DOI":"10.1109\/TIP.2016.2524212","volume":"25","author":"Q Zhang","year":"2016","unstructured":"Zhang, Q., & Levine, M. D. (2016). Robust multi-focus image fusion using multi-task sparse representation and spatial context. IEEE Transactions on Image Processing, 25(5), 2045\u20132058.","journal-title":"IEEE Transactions on Image Processing"},{"key":"1806_CR52","first-page":"4819","volume":"44","author":"X Zhang","year":"2021","unstructured":"Zhang, X. (2021). Deep learning-based multi-focus image fusion: A survey and a comparative study. IEEE Transactions on Pattern Analysis and Machine Intelligence online., 44, 4819.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence online."},{"key":"1806_CR53","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Liu, Y., Sun, P., et al. (2020). Ifcnn: A general image fusion framework based on convolutional neural network. Information Fusion, 54, 99\u2013118.","DOI":"10.1016\/j.inffus.2019.07.011"},{"key":"1806_CR54","doi-asserted-by":"crossref","unstructured":"Zhi-guo, J., Dong-bing, H., Jin, C., et\u00a0al. (2004). A wavelet based algorithm for multi-focus micro-image fusion. In: Third International Conference on Image and Graphics (ICIG\u201904), IEEE, pp 176\u2013179.","DOI":"10.1109\/ICIG.2004.29"}],"container-title":["International Journal of Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-023-01806-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11263-023-01806-w\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-023-01806-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,8,19]],"date-time":"2023-08-19T02:03:47Z","timestamp":1692410627000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11263-023-01806-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,9]]},"references-count":54,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2023,10]]}},"alternative-id":["1806"],"URL":"https:\/\/doi.org\/10.1007\/s11263-023-01806-w","relation":{},"ISSN":["0920-5691","1573-1405"],"issn-type":[{"value":"0920-5691","type":"print"},{"value":"1573-1405","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,6,9]]},"assertion":[{"value":"14 February 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 April 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 June 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}