{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,15]],"date-time":"2026-03-15T23:01:59Z","timestamp":1773615719763,"version":"3.50.1"},"reference-count":44,"publisher":"Allerton Press","issue":"2","license":[{"start":{"date-parts":[[2024,4,1]],"date-time":"2024-04-01T00:00:00Z","timestamp":1711929600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,4,1]],"date-time":"2024-04-01T00:00:00Z","timestamp":1711929600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Aut. Control Comp. Sci."],"published-print":{"date-parts":[[2024,4]]},"DOI":"10.3103\/s0146411624700068","type":"journal-article","created":{"date-parts":[[2024,5,6]],"date-time":"2024-05-06T10:02:20Z","timestamp":1714989740000},"page":"166-176","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["BCNN: An Effective Multifocus Image fusion Method Based on the Hierarchical Bayesian and Convolutional Neural Networks"],"prefix":"10.3103","volume":"58","author":[{"family":"ChunXiang Liu","sequence":"first","affiliation":[]},{"given":"Yuwei","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Lei","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Tianqi","family":"Cheng","sequence":"additional","affiliation":[]},{"given":"Xinping","family":"Guo","sequence":"additional","affiliation":[]}],"member":"1627","published-online":{"date-parts":[[2024,5,6]]},"reference":[{"key":"7692_CR1","doi-asserted-by":"publisher","first-page":"71","DOI":"10.1016\/j.inffus.2020.06.013","volume":"64","author":"Yu. Liu","year":"2020","unstructured":"Liu, Yu., Wang, L., Cheng, J., Li, C., and Chen, X., Multi-focus image fusion: A survey of the state of the art, Inf. Fusion, 2020, vol. 64, pp. 71\u201391. https:\/\/doi.org\/10.1016\/j.inffus.2020.06.013","journal-title":"Inf. Fusion"},{"key":"7692_CR2","doi-asserted-by":"publisher","first-page":"4425","DOI":"10.1007\/s11831-021-09540-7","volume":"28","author":"H. Kaur","year":"2021","unstructured":"Kaur, H., Koundal, D., and Kadyan, V., Image fusion techniques: A survey, Arch. Comput. Methods Eng., 2021, vol. 28, no. 7, pp. 4425\u20134447. https:\/\/doi.org\/10.1007\/s11831-021-09540-7","journal-title":"Arch. Comput. Methods Eng."},{"key":"7692_CR3","doi-asserted-by":"publisher","unstructured":"Stathaki, T., Image fusion: Algorithms and applications, Sensor Rev., 2009, vol. 29, no. 3. https:\/\/doi.org\/10.1108\/sr.2009.08729cae.001","DOI":"10.1108\/sr.2009.08729cae.001"},{"key":"7692_CR4","doi-asserted-by":"publisher","first-page":"104048","DOI":"10.1016\/j.compbiomed.2020.104048","volume":"126","author":"J. Fu","year":"2020","unstructured":"Fu, J., Li, W., and Du, J., Multimodal medical image fusion via Laplacian pyramid and convolutional neural network reconstruction with local gradient energy strategy, Comput. Biol. Med., 2020, vol. 126, p. 104048. https:\/\/doi.org\/10.1016\/j.compbiomed.2020.104048","journal-title":"Comput. Biol. Med."},{"key":"7692_CR5","doi-asserted-by":"publisher","first-page":"108996","DOI":"10.1016\/j.sigpro.2023.108996","volume":"208","author":"L. Sun","year":"2023","unstructured":"Sun, L., Li, Yu., Zheng, M., Zhong, Z., and Zhang, Ya., MCnet: Multiscale visible image and infrared image fusion network, Signal Process., 2023, vol. 208, p. 108996. https:\/\/doi.org\/10.1016\/j.sigpro.2023.108996","journal-title":"Signal Process."},{"key":"7692_CR6","doi-asserted-by":"publisher","first-page":"108542","DOI":"10.1016\/j.asoc.2022.108542","volume":"118","author":"Z. Chao","year":"2022","unstructured":"Chao, Z., Duan, X., Jia, S., Guo, X., Liu, H., and Jia, F., Medical image fusion via discrete stationary wavelet transform and an enhanced radial basis function neural network, Appl. Soft Comput., 2022, vol. 118, p. 108542. https:\/\/doi.org\/10.1016\/j.asoc.2022.108542","journal-title":"Appl. Soft Comput."},{"key":"7692_CR7","doi-asserted-by":"publisher","first-page":"107307","DOI":"10.1016\/j.asoc.2021.107307","volume":"106","author":"S. Bhat","year":"2021","unstructured":"Bhat, S. and Koundal, D., Multi-focus image fusion using neutrosophic based wavelet transform, Appl. Soft Comput., 2021, vol. 106, p. 107307. https:\/\/doi.org\/10.1016\/j.asoc.2021.107307","journal-title":"Appl. Soft Comput."},{"key":"7692_CR8","doi-asserted-by":"publisher","first-page":"268","DOI":"10.1016\/j.neucom.2015.01.050","volume":"159","author":"L. Dong","year":"2015","unstructured":"Dong, L., Yang, Q., Wu, H., Xiao, H., and Xu, M., High quality multi-spectral and panchromatic image fusion technologies based on Curvelet transform, Neurocomputing, 2015, vol. 159, pp. 268\u2013274. https:\/\/doi.org\/10.1016\/j.neucom.2015.01.050","journal-title":"Neurocomputing"},{"key":"7692_CR9","doi-asserted-by":"publisher","first-page":"108062","DOI":"10.1016\/j.sigpro.2021.108062","volume":"184","author":"X. Li","year":"2021","unstructured":"Li, X., Zhou, F., Tan, H., Chen, Yu., and Zuo, W., Multi-focus image fusion based on nonsubsampled contourlet transform and residual removal, Signal Process., 2021, vol. 184, p. 108062. https:\/\/doi.org\/10.1016\/j.sigpro.2021.108062","journal-title":"Signal Process."},{"key":"7692_CR10","doi-asserted-by":"publisher","first-page":"107793","DOI":"10.1016\/j.sigpro.2020.107793","volume":"178","author":"B. Li","year":"2021","unstructured":"Li, B., Peng, H., and Wang, J., A novel fusion method based on dynamic threshold neural P systems and nonsubsampled contourlet transform for multi-modality medical images, Signal Process., 2021, vol. 178, p. 107793. https:\/\/doi.org\/10.1016\/j.sigpro.2020.107793","journal-title":"Signal Process."},{"key":"7692_CR11","doi-asserted-by":"publisher","first-page":"104353","DOI":"10.1016\/j.bspc.2022.104353","volume":"80","author":"X. Li","year":"2023","unstructured":"Li, X., Wan, W., Zhou, F., Cheng, X., Jie, Yu., and Tan, H., Medical image fusion based on sparse representation and neighbor energy activity, Biomed. Signal Process. Control, 2023, vol. 80, p. 104353. https:\/\/doi.org\/10.1016\/j.bspc.2022.104353","journal-title":"Biomed. Signal Process. Control"},{"key":"7692_CR12","doi-asserted-by":"publisher","first-page":"119909","DOI":"10.1016\/j.eswa.2023.119909","volume":"223","author":"L. Qu","year":"2023","unstructured":"Qu, L., Yin, S., Liu, S., Liu, X., Wang, M., and Song, Z., AIM-MEF: Multi-exposure image fusion based on adaptive information mining in both spatial and frequency domains, Expert Syst. Appl., 2023, vol. 223, p. 119909. https:\/\/doi.org\/10.1016\/j.eswa.2023.119909","journal-title":"Expert Syst. Appl."},{"key":"7692_CR13","doi-asserted-by":"publisher","first-page":"116135","DOI":"10.1016\/j.eswa.2021.116135","volume":"189","author":"T. Kurban","year":"2022","unstructured":"Kurban, T., Region based multi-spectral fusion method for remote sensing images using differential search algorithm and IHS transform, Expert Syst. Appl., 2022, vol. 189, p. 116135. https:\/\/doi.org\/10.1016\/j.eswa.2021.116135","journal-title":"Expert Syst. Appl."},{"key":"7692_CR14","doi-asserted-by":"publisher","first-page":"104178","DOI":"10.1016\/j.infrared.2022.104178","volume":"123","author":"Z. Guo","year":"2022","unstructured":"Guo, Z., Yu, X., and Du, Q., Infrared and visible image fusion based on saliency and fast guided filtering, Infrared Phys. Technol., 2022, vol. 123, p. 104178. https:\/\/doi.org\/10.1016\/j.infrared.2022.104178","journal-title":"Infrared Phys. Technol."},{"key":"7692_CR15","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.inffus.2014.10.004","volume":"25","author":"N. Mansour","year":"2015","unstructured":"Mansour, N., Samavi, S., and Shirani, Sh., Multi-focus image fusion using dictionary-based sparse representation, Inf. Fusion, 2015, vol. 25, pp. 72\u201384. https:\/\/doi.org\/10.1016\/j.inffus.2014.10.004","journal-title":"Inf. Fusion"},{"key":"7692_CR16","doi-asserted-by":"publisher","first-page":"295","DOI":"10.1016\/j.jvcir.2019.06.002","volume":"62","author":"N. Hayat","year":"2019","unstructured":"Hayat, N. and Imran, M., Ghost-free multi exposure image fusion technique using dense SIFT descriptor and guided filter, J. Visual Commun. Image Representation, 2019, vol. 62, pp. 295\u2013308. https:\/\/doi.org\/10.1016\/j.jvcir.2019.06.002","journal-title":"J. Visual Commun. Image Representation"},{"key":"7692_CR17","doi-asserted-by":"publisher","first-page":"715","DOI":"10.1016\/j.jestch.2019.01.004","volume":"22","author":"J. Jinju","year":"2019","unstructured":"Jinju, J., Santhi, N., Ramar, K., and Sathya Bama, B., Spatial frequency discrete wavelet transform image fusion technique for remote sensing applications, Eng. Sci. Technol., \n               Int. J., 2019, vol. 22, no. 3, pp. 715\u2013726. https:\/\/doi.org\/10.1016\/j.jestch.2019.01.004","journal-title":"Int. J."},{"key":"7692_CR18","doi-asserted-by":"publisher","first-page":"191","DOI":"10.1016\/j.inffus.2016.12.001","volume":"36","author":"Yu. Liu","year":"2017","unstructured":"Liu, Yu., Chen, X., Peng, H., and Wang, Z., Multi-focus image fusion with a deep convolutional neural network, Inf. Fusion, 2017, vol. 36, pp. 191\u2013207. https:\/\/doi.org\/10.1016\/j.inffus.2016.12.001","journal-title":"Inf. Fusion"},{"key":"7692_CR19","doi-asserted-by":"publisher","first-page":"125","DOI":"10.1016\/j.ins.2017.12.043","volume":"433\u2013434","author":"H. Tang","year":"2018","unstructured":"Tang, H., Xiao, B., Li, W., and Wang, G., Pixel convolutional neural network for multi-focus image fusion, Inf. Sci., 2018, vols. 433\u2013434, pp. 125\u2013141. https:\/\/doi.org\/10.1016\/j.ins.2017.12.043","journal-title":"Inf. Sci."},{"key":"7692_CR20","doi-asserted-by":"publisher","first-page":"201","DOI":"10.1016\/j.inffus.2019.02.003","volume":"51","author":"M. Amin-Naji","year":"2019","unstructured":"Amin-Naji, M., Aghagolzadeh, A., and Ezoji, M., Ensemble of CNN for multi-focus image fusion, Inf. Fusion, 2019, vol. 51, pp. 201\u2013214. https:\/\/doi.org\/10.1016\/j.inffus.2019.02.003","journal-title":"Inf. Fusion"},{"key":"7692_CR21","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1016\/j.inffus.2019.07.011","volume":"54","author":"Yu. Zhang","year":"2020","unstructured":"Zhang, Yu., Liu, Yu., Sun, P., Yan, H., Zhao, X., and Zhang, L., IFCNN: A general image fusion framework based on convolutional neural network, Inf. Fusion, 2020, vol. 54, pp. 99\u2013118. https:\/\/doi.org\/10.1016\/j.inffus.2019.07.011","journal-title":"Inf. Fusion"},{"key":"7692_CR22","doi-asserted-by":"publisher","first-page":"107681","DOI":"10.1016\/j.sigpro.2020.107681","volume":"176","author":"D. Gai","year":"2020","unstructured":"Gai, D., Shen, X., Chen, H., and Su, P., Multi-focus image fusion method based on two stage of convolutional neural network, Signal Process., 2020, vol. 176, p. 107681. https:\/\/doi.org\/10.1016\/j.sigpro.2020.107681","journal-title":"Signal Process."},{"key":"7692_CR23","doi-asserted-by":"publisher","first-page":"107174","DOI":"10.1016\/j.compeleceng.2021.107174","volume":"92","author":"Z. Yang","year":"2021","unstructured":"Yang, Z., Yang, X., Zhang, R., Liu, K., Anisetti, M., and Jeon, G., Gradient-based multi-focus image fusion method using convolution neural network, Comput. Electr. Eng., 2021, vol. 92, no. 4, p. 107174. https:\/\/doi.org\/10.1016\/j.compeleceng.2021.107174","journal-title":"Comput. Electr. Eng."},{"key":"7692_CR24","doi-asserted-by":"publisher","first-page":"5793","DOI":"10.1007\/s00521-020-05358-9","volume":"33","author":"B. Ma","year":"2021","unstructured":"Ma, B., Zhu, Yu., Yin, X., Ban, X., Huang, H., and Mukeshimana, M., SESF-Fuse: An unsupervised deep model for multi-focus image fusion, Neural Comput. Appl., 2021, vol. 33, no. 11, pp. 5793\u20135804. https:\/\/doi.org\/10.1007\/s00521-020-05358-9","journal-title":"Neural Comput. Appl."},{"key":"7692_CR25","doi-asserted-by":"publisher","first-page":"40","DOI":"10.1016\/j.inffus.2020.08.022","volume":"66","author":"H. Zhang","year":"2021","unstructured":"Zhang, H., Le, Z., Shao, Z., Xu, H., and Ma, J., MFF-GAN: An unsupervised generative adversarial network with adaptive and gradient joint constraints for multi-focus image fusion, Inf. Fusion, 2021, vol. 66, pp. 40\u201353. https:\/\/doi.org\/10.1016\/j.inffus.2020.08.022","journal-title":"Inf. Fusion"},{"key":"7692_CR26","doi-asserted-by":"publisher","first-page":"108282","DOI":"10.1016\/j.sigpro.2021.108282","volume":"189","author":"R. Zhao","year":"2021","unstructured":"Zhao, R., Zhang, T., Luo, X., and Tan, J., DCKN: Multi-focus image fusion via dynamic convolutional kernel network, Signal Process., 2021, vol. 189, p. 108282. https:\/\/doi.org\/10.1016\/j.sigpro.2021.108282","journal-title":"Signal Process."},{"key":"7692_CR27","doi-asserted-by":"publisher","first-page":"110291","DOI":"10.1016\/j.knosys.2023.110291","volume":"263","author":"G. Yang","year":"2023","unstructured":"Yang, G., Wu, X., and Zhang, J., A dynamic balanced quadtree for real-time streaming data, Knowl.-Based Syst., 2023, vol. 263, p. 110291. https:\/\/doi.org\/10.1016\/j.knosys.2023.110291","journal-title":"Knowl.-Based Syst."},{"key":"7692_CR28","unstructured":"ImageNet. https:\/\/image-net.org\/. Cited January 11, 2023."},{"key":"7692_CR29","unstructured":"Lytro Multi-Focus Image Dataset. http:\/\/mansournejati.ece.iut.ac.ir\/content\/lytro-multi-focus-dataset. Cited January 20, 2023."},{"key":"7692_CR30","doi-asserted-by":"publisher","first-page":"52","DOI":"10.1016\/j.isprsjprs.2021.10.001","volume":"182","author":"H. Cheng","year":"2021","unstructured":"Cheng, H., Wu, H., Zheng, J., Qi, K., and Liu, W., A hierarchical self-attention augmented Laplacian pyramid expanding network for change detection in high-resolution remote sensing images, ISPRS J. Photogrammetry Remote Sensing, 2021, vol. 182, pp. 52\u201366. https:\/\/doi.org\/10.1016\/j.isprsjprs.2021.10.001","journal-title":"ISPRS J. Photogrammetry Remote Sensing"},{"key":"7692_CR31","doi-asserted-by":"publisher","first-page":"60","DOI":"10.1016\/j.inffus.2013.11.005","volume":"20","author":"Z. Zhou","year":"2014","unstructured":"Zhou, Z., Li, S., and Wang, B., Multi-scale weighted gradient-based fusion for multi-focus images, Inf. Fusion, 2014, vol. 20, pp. 60\u201372. https:\/\/doi.org\/10.1016\/j.inffus.2013.11.005","journal-title":"Inf. Fusion"},{"key":"7692_CR32","doi-asserted-by":"publisher","first-page":"1334","DOI":"10.1016\/j.sigpro.2009.01.012","volume":"89","author":"Q. Zhang","year":"2009","unstructured":"Zhang, Q. and Guo, B.-L., Multifocus image fusion using the nonsubsampled contourlet transform, Signal Process., 2009, vol. 89, no. 7, pp. 1334\u20131346. https:\/\/doi.org\/10.1016\/j.sigpro.2009.01.012","journal-title":"Signal Process."},{"key":"7692_CR33","doi-asserted-by":"publisher","unstructured":"Borwonwatanadelok, P., Rattanapitak, W., and Udomhunsakul, S., Multi-focus image fusion based on stationary wavelet transform and extended spatial frequency measurement, 2009 Int. Conf. on Electronic Computer Technology, Macau, China, 2009, IEEE, 2009, pp. 77\u201381. https:\/\/doi.org\/10.1109\/icect.2009.94","DOI":"10.1109\/icect.2009.94"},{"key":"7692_CR34","doi-asserted-by":"publisher","first-page":"35","DOI":"10.1016\/j.image.2018.12.004","volume":"72","author":"X. Qiu","year":"2019","unstructured":"Qiu, X., Li, M., Zhang, L., and Yuan, X., Guided filter-based multi-focus image fusion through focus region detection, Signal Process.: Image Commun., 2019, vol. 72, pp. 35\u201346. https:\/\/doi.org\/10.1016\/j.image.2018.12.004","journal-title":"Signal Process.: Image Commun."},{"key":"7692_CR35","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1016\/j.inffus.2014.05.004","volume":"23","author":"Yu. Liu","year":"2015","unstructured":"Liu, Yu., Liu, S., and Wang, Z., Multi-focus image fusion with dense SIFT, Inf. Fusion, 2015, vol. 23, pp. 139\u2013155. https:\/\/doi.org\/10.1016\/j.inffus.2014.05.004","journal-title":"Inf. Fusion"},{"key":"7692_CR36","doi-asserted-by":"publisher","first-page":"103214","DOI":"10.1016\/j.bspc.2021.103214","volume":"71","author":"S. Goyal","year":"2022","unstructured":"Goyal, S., Singh, V., Rani, A., and Yadav, N., Multimodal image fusion and denoising in NSCT domain using CNN and FOTGV, Biomed. Signal Process. Control, 2022, vol. 71, p. 103214. https:\/\/doi.org\/10.1016\/j.bspc.2021.103214","journal-title":"Biomed. Signal Process. Control"},{"key":"7692_CR37","doi-asserted-by":"publisher","first-page":"685","DOI":"10.1016\/j.ins.2021.08.030","volume":"579","author":"T. Zhang","year":"2021","unstructured":"Zhang, T., Waqas, M., Liu, Z., Tu, S., Halim, Z., Rehman, S.U., Li, Yu., and Han, Z., A fusing framework of shortcut convolutional neural networks, Inf. Sci., 2021, vol. 579, pp. 685\u2013699. https:\/\/doi.org\/10.1016\/j.ins.2021.08.030","journal-title":"Inf. Sci."},{"key":"7692_CR38","doi-asserted-by":"publisher","first-page":"127","DOI":"10.1016\/j.inffus.2011.08.002","volume":"14","author":"Yu. Han","year":"2013","unstructured":"Han, Yu., Cai, Yu., Cao, Yi., and Xu, X., A new image fusion performance metric based on visual information fidelity, Inf. Fusion, 2013, vol. 14, no. 2, pp. 127\u2013135. https:\/\/doi.org\/10.1016\/j.inffus.2011.08.002","journal-title":"Inf. Fusion"},{"key":"7692_CR39","doi-asserted-by":"publisher","first-page":"454","DOI":"10.1016\/j.inffus.2023.02.021","volume":"95","author":"Zh. Chang","year":"2023","unstructured":"Chang, Zh., Yang, Sh., Feng, Zh., Gao, Q., Wang, Sh., and Cui, Yu., Semantic-relation transformer for visible and infrared fused image quality assessment, Inf. Fusion, 2023, vol. 95, pp. 454\u2013470. https:\/\/doi.org\/10.1016\/j.inffus.2023.02.021","journal-title":"Inf. Fusion"},{"key":"7692_CR40","doi-asserted-by":"publisher","first-page":"1890","DOI":"10.1016\/j.aeue.2015.09.004","volume":"69","author":"V. Aslantas","year":"2015","unstructured":"Aslantas, V. and Bendes, E., A new image quality metric for image fusion: The sum of the correlations of differences, AEU Int. J. Electron. Commun., 2015, vol. 69, no. 12, pp. 1890\u20131896. https:\/\/doi.org\/10.1016\/j.aeue.2015.09.004","journal-title":"AEU Int. J. Electron. Commun."},{"key":"7692_CR41","doi-asserted-by":"publisher","first-page":"56","DOI":"10.1016\/j.cviu.2007.04.003","volume":"109","author":"Z. Liu","year":"2008","unstructured":"Liu, Z., Forsyth, D.S., and Lagani\u00e8re, R., A feature-based metric for the quantitative evaluation of pixel-level image fusion, Comput. Vision Image Understanding, 2008, vol. 109, no. 1, pp. 56\u201368. https:\/\/doi.org\/10.1016\/j.cviu.2007.04.003","journal-title":"Comput. Vision Image Understanding"},{"key":"7692_CR42","doi-asserted-by":"publisher","first-page":"103328","DOI":"10.1016\/j.jvcir.2021.103328","volume":"81","author":"X. Ma","year":"2021","unstructured":"Ma, X., Wang, Zh., and Hu, S., Multi-focus image fusion based on multi-scale sparse representation, J. Visual Commun. Image Representation, 2021, vol. 81, p. 103328. https:\/\/doi.org\/10.1016\/j.jvcir.2021.103328","journal-title":"J. Visual Commun. Image Representation"},{"key":"7692_CR43","doi-asserted-by":"publisher","first-page":"45","DOI":"10.1016\/j.patrec.2020.11.014","volume":"141","author":"H. Li","year":"2021","unstructured":"Li, H., Zhang, L., Jiang, M., and Li, Yu., Multi-focus image fusion algorithm based on supervised learning for fully convolutional neural network, Pattern Recognit. Lett., 2021, vol. 141, pp. 45\u201353. https:\/\/doi.org\/10.1016\/j.patrec.2020.11.014","journal-title":"Pattern Recognit. Lett."},{"key":"7692_CR44","doi-asserted-by":"publisher","first-page":"40","DOI":"10.1016\/j.inffus.2020.08.022","volume":"66","author":"H. Zhang","year":"2021","unstructured":"Zhang, H., Le, Zh., Shao, Zh., Xu, H., and Ma, J., MFF-GAN: An unsupervised generative adversarial network with adaptive and gradient joint constraints for multi-focus image fusion, Inf. Fusion, 2021, vol. 66, pp. 40\u201353. https:\/\/doi.org\/10.1016\/j.inffus.2020.08.022","journal-title":"Inf. Fusion"}],"container-title":["Automatic Control and Computer Sciences"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.3103\/S0146411624700068.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.3103\/S0146411624700068","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.3103\/S0146411624700068.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,15]],"date-time":"2026-03-15T22:03:41Z","timestamp":1773612221000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.3103\/S0146411624700068"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,4]]},"references-count":44,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2024,4]]}},"alternative-id":["7692"],"URL":"https:\/\/doi.org\/10.3103\/s0146411624700068","relation":{},"ISSN":["0146-4116","1558-108X"],"issn-type":[{"value":"0146-4116","type":"print"},{"value":"1558-108X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,4]]},"assertion":[{"value":"14 April 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 July 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 July 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 May 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"The authors of this work declare that they have no conflicts of interest.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"CONFLICT OF INTEREST"}}]}}