{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T14:16:16Z","timestamp":1740147376592,"version":"3.37.3"},"reference-count":50,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2023,11,16]],"date-time":"2023-11-16T00:00:00Z","timestamp":1700092800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,11,16]],"date-time":"2023-11-16T00:00:00Z","timestamp":1700092800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62203224"],"award-info":[{"award-number":["62203224"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2024,3]]},"DOI":"10.1007\/s11760-023-02860-0","type":"journal-article","created":{"date-parts":[[2023,11,16]],"date-time":"2023-11-16T18:01:51Z","timestamp":1700157711000},"page":"1443-1453","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Environment enhanced fusion of infrared and visible images based on saliency assignment"],"prefix":"10.1007","volume":"18","author":[{"given":"Jiebang","family":"Wang","sequence":"first","affiliation":[]},{"given":"Gang","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Xiangbo","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Haojie","family":"Tang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,11,16]]},"reference":[{"key":"2860_CR1","doi-asserted-by":"publisher","first-page":"153","DOI":"10.1016\/j.inffus.2018.02.004","volume":"45","author":"J Ma","year":"2019","unstructured":"Ma, J., Ma, Y., Li, C.: Infrared and visible image fusion methods and applications: a survey. Inf. Fusion 45, 153\u2013178 (2019)","journal-title":"Inf. Fusion"},{"issue":"5","key":"2860_CR2","doi-asserted-by":"publisher","first-page":"3645","DOI":"10.1007\/s11831-020-09518-x","volume":"28","author":"S Singh","year":"2021","unstructured":"Singh, S., Mittal, N., Singh, H.: Review of various image fusion algorithms and image fusion performance metric. Arch. Comput, Methods Eng. 28(5), 3645\u20133659 (2021)","journal-title":"Arch. Comput, Methods Eng."},{"issue":"2","key":"2860_CR3","doi-asserted-by":"publisher","first-page":"74","DOI":"10.1016\/j.inffus.2010.03.002","volume":"12","author":"S Li","year":"2011","unstructured":"Li, S., Yang, B., Jianwen, H.: Performance comparison of different multi-resolution transforms for image fusion. Inf. Fusion 12(2), 74\u201384 (2011)","journal-title":"Inf. Fusion"},{"key":"2860_CR4","doi-asserted-by":"publisher","first-page":"134","DOI":"10.1016\/j.infrared.2014.02.013","volume":"64","author":"H Jin","year":"2014","unstructured":"Jin, H., Wang, Y.: A fusion method for visible and infrared images based on contrast pyramid with teaching learning based optimization. Infrared Phys. Technol. 64, 134\u2013142 (2014)","journal-title":"Infrared Phys. Technol."},{"key":"2860_CR5","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.infrared.2017.10.004","volume":"88","author":"X Jin","year":"2018","unstructured":"Jin, X., Jiang, Q., Yao, S., Zhou, D., Nie, R., Lee, S.-J., He, K.: Infrared and visual image fusion method based on discrete cosine transform and local spatial frequency in discrete stationary wavelet transform domain. Infrared Phys. Technol. 88, 1\u201312 (2018)","journal-title":"Infrared Phys. Technol."},{"key":"2860_CR6","doi-asserted-by":"publisher","first-page":"154","DOI":"10.1016\/j.infrared.2018.06.002","volume":"92","author":"T Ma","year":"2018","unstructured":"Ma, T., Ma, J., Fang, B., Fangyu, H., Quan, S., Huajun, D.: Multi-scale decomposition based fusion of infrared and visible image via total variation and saliency analysis. Infrared Phys. Technol. 92, 154\u2013162 (2018)","journal-title":"Infrared Phys. Technol."},{"issue":"10","key":"2860_CR7","doi-asserted-by":"publisher","first-page":"8137","DOI":"10.1007\/s00521-022-06900-7","volume":"34","author":"S Singh","year":"2022","unstructured":"Singh, S., Mittal, N., Singh, H.: A feature level image fusion for IR and visible image using MNMRA based segmentation. Neural Comput. Appl. 34(10), 8137\u20138154 (2022)","journal-title":"Neural Comput. Appl."},{"key":"2860_CR8","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma, J., Wei, Yu., Liang, P., Li, C., Jiang, J.: Fusiongan: a generative adversarial network for infrared and visible image fusion. Inf. Fusion 48, 11\u201326 (2019)","journal-title":"Inf. Fusion"},{"key":"2860_CR9","doi-asserted-by":"publisher","first-page":"40","DOI":"10.1016\/j.inffus.2020.08.022","volume":"66","author":"H Zhang","year":"2021","unstructured":"Zhang, H., Le, Z., Shao, Z., Han, X., Ma, J.: MFF-GAN: an unsupervised generative adversarial network with adaptive and gradient joint constraints for multi-focus image fusion. Inf. Fusion 66, 40\u201353 (2021)","journal-title":"Inf. Fusion"},{"key":"2860_CR10","doi-asserted-by":"publisher","DOI":"10.1016\/j.infrared.2021.103914","volume":"120","author":"LL Tang","year":"2022","unstructured":"Tang, L.L., Liu, G., Xiao, G., Bavirisetti, D.P., Zhang, X.B.: Infrared and visible image fusion based on guided hybrid model and generative adversarial network. Infrared Phys. Technol. 120, 103914 (2022)","journal-title":"Infrared Phys. Technol."},{"issue":"3","key":"2860_CR11","doi-asserted-by":"publisher","first-page":"376","DOI":"10.3390\/e23030376","volume":"23","author":"J Hou","year":"2021","unstructured":"Hou, J., Zhang, D., Wei, W., Ma, J., Zhou, H.: A generative adversarial network for infrared and visible image fusion based on semantic segmentation. Entropy 23(3), 376 (2021)","journal-title":"Entropy"},{"key":"2860_CR12","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1016\/j.inffus.2021.12.004","volume":"82","author":"L Tang","year":"2022","unstructured":"Tang, L., Yuan, J., Ma, J.: Image fusion in the loop of high-level vision tasks: a semantic-aware real-time infrared and visible image fusion network. Inf. Fusion 82, 28\u201342 (2022)","journal-title":"Inf. Fusion"},{"issue":"1","key":"2860_CR13","doi-asserted-by":"publisher","first-page":"04020131","DOI":"10.1061\/(ASCE)CF.1943-5509.0001541","volume":"35","author":"S Pozzer","year":"2021","unstructured":"Pozzer, S., Rezazadeh\u00a0Azar, E., Dalla\u00a0Rosa, F., Chamberlain\u00a0Pravia, Z.M.: Semantic segmentation of defects in infrared thermographic images of highly damaged concrete structures. J. Perform. Constr. Facil. 35(1), 04020131 (2021)","journal-title":"J. Perform. Constr. Facil."},{"key":"2860_CR14","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/LGRS.2022.3141584","volume":"19","author":"Q Hou","year":"2021","unstructured":"Hou, Q., Wang, Z., Tan, F., Zhao, Y., Zheng, H., Zhang, W.: RISTDNET: robust infrared small target detection network. IEEE Geosci. Remote Sens. Lett. 19, 1\u20135 (2021)","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"issue":"1","key":"2860_CR15","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s10921-020-00740-y","volume":"40","author":"D M\u00fcller","year":"2021","unstructured":"M\u00fcller, D., Ehlen, A., Valeske, B.: Convolutional neural networks for semantic segmentation as a tool for multiclass face analysis in thermal infrared. J. Nondestr. Eval. 40(1), 1\u201310 (2021)","journal-title":"J. Nondestr. Eval."},{"issue":"3","key":"2860_CR16","doi-asserted-by":"publisher","first-page":"559","DOI":"10.1109\/TKDE.2017.2763620","volume":"30","author":"X Chen","year":"2017","unstructured":"Chen, X., Fang, Y., Yang, M., Nie, F., Zhao, Z., Huang, J.Z.: Purtreeclust: a clustering algorithm for customer segmentation from massive customer transaction data. IEEE Trans. Knowl. Data Eng. 30(3), 559\u2013572 (2017)","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"2860_CR17","first-page":"281","volume":"1","author":"JB MacQueen","year":"1967","unstructured":"MacQueen, J.B.: Some methods for classification and analysis of multivariate observations. Berkeley Sympos. Math. Stat. Probab. 1, 281\u2013297 (1967)","journal-title":"Berkeley Sympos. Math. Stat. Probab."},{"issue":"5","key":"2860_CR18","doi-asserted-by":"publisher","first-page":"1328","DOI":"10.1109\/TIP.2010.2040763","volume":"19","author":"S Krinidis","year":"2010","unstructured":"Krinidis, S., Chatzis, V.: A robust fuzzy local information c-means clustering algorithm. IEEE Trans. Image Process. 19(5), 1328\u20131337 (2010)","journal-title":"IEEE Trans. Image Process."},{"key":"2860_CR19","doi-asserted-by":"publisher","first-page":"332","DOI":"10.1109\/TFUZZ.2020.3037972","volume":"30","author":"F Yang","year":"2020","unstructured":"Yang, F., Liu, Z., Bai, X., Zhang, Y.: An improved intuitionistic fuzzy c-means for ship segmentation in infrared images. IEEE Trans. Fuzzy Syst. 30, 332\u2013344 (2020)","journal-title":"IEEE Trans. Fuzzy Syst."},{"issue":"3","key":"2860_CR20","doi-asserted-by":"publisher","first-page":"569","DOI":"10.1109\/TPAMI.2014.2345401","volume":"37","author":"M-M Cheng","year":"2014","unstructured":"Cheng, M.-M., Mitra, N.J., Huang, X., Torr, P.H.S., Shi-Min, H.: Global contrast based salient region detection. IEEE Trans. Pattern Anal. Mach. Intell. 37(3), 569\u2013582 (2014)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2860_CR21","doi-asserted-by":"crossref","unstructured":"Achanta, R., Estrada, F., Wils, P., S\u00fcsstrunk, S.: Salient region detection and segmentation. In: International Conference on Computer Vision Systems, pp. 66\u201375. Springer, Berlin (2008)","DOI":"10.1007\/978-3-540-79547-6_7"},{"key":"2860_CR22","doi-asserted-by":"crossref","unstructured":"Zhai, Y., Shah, M.: Visual attention detection in video sequences using spatiotemporal cues. In: Proceedings of the 14th ACM International Conference on Multimedia, pp. 815\u2013824 (2006)","DOI":"10.1145\/1180639.1180824"},{"issue":"12","key":"2860_CR23","doi-asserted-by":"publisher","first-page":"2259","DOI":"10.1016\/S0031-3203(00)00149-7","volume":"34","author":"H-D Cheng","year":"2001","unstructured":"Cheng, H.-D., Jiang, X.-H., Sun, Y., Wang, J.: Color image segmentation: advances and prospects. Pattern Recogn. 34(12), 2259\u20132281 (2001)","journal-title":"Pattern Recogn."},{"key":"2860_CR24","doi-asserted-by":"crossref","unstructured":"Judd, T., Ehinger, K., Durand, F., Torralba, A.: Learning to predict where humans look. In: 2009 IEEE 12th International Conference on Computer Vision, pp. 2106\u20132113. IEEE (2009)","DOI":"10.1109\/ICCV.2009.5459462"},{"key":"2860_CR25","doi-asserted-by":"publisher","first-page":"89","DOI":"10.1016\/j.inffus.2019.12.007","volume":"57","author":"S Bianco","year":"2020","unstructured":"Bianco, S., Buzzelli, M., Ciocca, G., Schettini, R.: Neural architecture search for image saliency fusion. Inf. Fusion 57, 89\u2013101 (2020)","journal-title":"Inf. Fusion"},{"key":"2860_CR26","unstructured":"Su, S.L., Durand, F., Agrawala, M.: An inverted saliency model for display enhancement. In: Proceedings of 2004 MIT Student Oxygen Workshop, Ashland, MA. Citeseer (2004)"},{"key":"2860_CR27","doi-asserted-by":"crossref","unstructured":"Wong, L.-K., Low, K.-L.: Saliency retargeting: An approach to enhance image aesthetics. In: 2011 IEEE Workshop on Applications of Computer Vision (WACV), pp. 73\u201380. IEEE (2011)","DOI":"10.1109\/WACV.2011.5711486"},{"key":"2860_CR28","doi-asserted-by":"crossref","unstructured":"Mejjati, Y.A., Gomez, C.F., Kim, K.I., Shechtman, E., Bylinskii, Z.: Look here! a parametric learning based approach to redirect visual attention. In: European Conference on Computer Vision, pp. 343\u2013361. Springer, Berlin (2020)","DOI":"10.1007\/978-3-030-58592-1_21"},{"issue":"3","key":"2860_CR29","doi-asserted-by":"publisher","first-page":"355","DOI":"10.1016\/S0734-189X(87)80186-X","volume":"39","author":"SM Pizer","year":"1987","unstructured":"Pizer, S.M., Amburn, E.P., Austin, J.D., Cromartie, R., Geselowitz, A., Greer, T., Haar\u00a0Romeny, B., Zimmerman, J.B., Zuiderveld, K.: Adaptive histogram equalization and its variations. Comput. Vision Graph. Image Process. 39(3), 355\u2013368 (1987)","journal-title":"Comput. Vision Graph. Image Process."},{"key":"2860_CR30","doi-asserted-by":"publisher","first-page":"164","DOI":"10.1016\/j.infrared.2018.03.010","volume":"90","author":"S Li","year":"2018","unstructured":"Li, S., Jin, W., Li, L., Li, Y.: An improved contrast enhancement algorithm for infrared images based on adaptive double plateaus histogram equalization. Infrared Phys. Technol. 90, 164\u2013174 (2018)","journal-title":"Infrared Phys. Technol."},{"key":"2860_CR31","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijleo.2021.167887","volume":"247","author":"A Paul","year":"2021","unstructured":"Paul, A., Sutradhar, T., Bhattacharya, P., Maity, S.P.: Infrared images enhancement using fuzzy dissimilarity histogram equalization. Optik 247, 167887 (2021)","journal-title":"Optik"},{"key":"2860_CR32","doi-asserted-by":"publisher","first-page":"164","DOI":"10.1016\/j.infrared.2018.04.003","volume":"91","author":"M Wan","year":"2018","unstructured":"Wan, M., Guohua, G., Qian, W., Ren, K., Chen, Q., Maldague, X.: Particle swarm optimization-based local entropy weighted histogram equalization for infrared image enhancement. Infrared Phys. Technol. 91, 164\u2013181 (2018)","journal-title":"Infrared Phys. Technol."},{"key":"2860_CR33","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijleo.2021.166273","volume":"230","author":"UK Acharya","year":"2021","unstructured":"Acharya, U.K., Kumar, S.: Genetic algorithm based adaptive histogram equalization (GAAHE) technique for medical image enhancement. Optik 230, 166273 (2021)","journal-title":"Optik"},{"key":"2860_CR34","unstructured":"Srinivasan, S., Balram, N.: Adaptive contrast enhancement using local region stretching. In: Proceedings of the 9th Asian symposium on information display, pp. 152\u2013155. Citeseer (2006)"},{"issue":"2\u20133","key":"2860_CR35","doi-asserted-by":"publisher","first-page":"191","DOI":"10.1016\/0098-3004(84)90020-7","volume":"10","author":"JC Bezdek","year":"1984","unstructured":"Bezdek, J.C., Ehrlich, R., Full, W.: FCM: the fuzzy c-means clustering algorithm. Comput. Geosci. 10(2\u20133), 191\u2013203 (1984)","journal-title":"Comput. Geosci."},{"key":"2860_CR36","doi-asserted-by":"crossref","unstructured":"Bavirisetti, D.P., Xiao, G., Liu, G.: Multi-sensor image fusion based on fourth order partial differential equations. In: International Conference on Information Fusion 7 (2017)","DOI":"10.23919\/ICIF.2017.8009719"},{"key":"2860_CR37","doi-asserted-by":"publisher","first-page":"4733","DOI":"10.1109\/TIP.2020.2975984","volume":"29","author":"H Li","year":"2020","unstructured":"Li, H., Xiao-Jun, W., Kittler, J.: MDLATLRR: a novel decomposition method for infrared and visible image fusion. IEEE Trans. Image Process. 29, 4733\u20134746 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"2860_CR38","doi-asserted-by":"publisher","first-page":"227","DOI":"10.1016\/j.infrared.2017.05.007","volume":"83","author":"Yu Zhang","year":"2017","unstructured":"Zhang, Yu., Zhang, L., Bai, X., Zhang, L.: Infrared and visual image fusion through infrared feature extraction and visual information preservation. Infrared Phys. Technol. 83, 227\u2013237 (2017)","journal-title":"Infrared Phys. Technol."},{"key":"2860_CR39","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.inffus.2021.02.023","volume":"73","author":"H Li","year":"2021","unstructured":"Li, H., Xiao-Jun, W., Kittler, J.: RFN-nest: An end-to-end residual fusion network for infrared and visible images. Inf. Fusion 73, 72\u201386 (2021)","journal-title":"Inf. Fusion"},{"issue":"5","key":"2860_CR40","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2018","unstructured":"Li, H., Wu, X.-J.: Densefuse: a fusion approach to infrared and visible images. IEEE Trans. Image Process. 28(5), 2614\u20132623 (2018)","journal-title":"IEEE Trans. Image Process."},{"key":"2860_CR41","doi-asserted-by":"crossref","unstructured":"Ram Prabhakar, K., Sai Srikar, V., Venkatesh Babu, R.: Deepfuse: a deep unsupervised approach for exposure fusion with extreme exposure image pairs. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 4714\u20134722 (2017)","DOI":"10.1109\/ICCV.2017.505"},{"key":"2860_CR42","doi-asserted-by":"publisher","first-page":"4980","DOI":"10.1109\/TIP.2020.2977573","volume":"29","author":"J Ma","year":"2020","unstructured":"Ma, J., Han, X., Jiang, J., Mei, X., Zhang, X.-P.: DDCGAN: a dual-discriminator conditional generative adversarial network for multi-resolution image fusion. IEEE Trans. Image Process. 29, 4980\u20134995 (2020)","journal-title":"IEEE Trans. Image Process."},{"issue":"1","key":"2860_CR43","first-page":"502","volume":"44","author":"X Han","year":"2020","unstructured":"Han, X., Ma, J., Jiang, J., Guo, X., Ling, H.: U2FUSION: a unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 44(1), 502\u2013518 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"1","key":"2860_CR44","first-page":"1","volume":"2","author":"W Roberts","year":"2008","unstructured":"Roberts, W., van Aardt, J., Ahmed, F.: Assessment of image fusion procedures using entropy, image quality, and multispectral classification. J. Appl. Remote Sens. 2(1), 1\u201328 (2008)","journal-title":"J. Appl. Remote Sens."},{"key":"2860_CR45","doi-asserted-by":"publisher","first-page":"1266","DOI":"10.1049\/el.2010.1778","volume":"46","author":"M Hossnym","year":"2010","unstructured":"Hossnym, M., Nahavandi, S., Creighton, D., Bhatti, A.: image fusion performance metric based on mutual information and entropy driven quadtree decomposition. Electron. Lett. 46, 1266\u20131268 (2010)","journal-title":"Electron. Lett."},{"key":"2860_CR46","doi-asserted-by":"publisher","first-page":"127","DOI":"10.1016\/j.inffus.2011.08.002","volume":"14","author":"H Yu","year":"2013","unstructured":"Yu, H., Cai, Y., Cao, Y.: Xu, X: a new image fusion performance metric based on visual information fidelity. Inf. Fusion 14, 127\u2013135 (2013)","journal-title":"Inf. Fusion"},{"key":"2860_CR47","doi-asserted-by":"publisher","first-page":"199","DOI":"10.1016\/j.optcom.2014.12.032","volume":"341","author":"G Cui","year":"2015","unstructured":"Cui, G., Feng, H., Xu, Z., Li, Q., Chen, Y.: Detail preserved fusion of visible and infrared images using regional saliency extraction and multi-scale image decomposition. Opt. Commun. 341, 199\u2013209 (2015)","journal-title":"Opt. Commun."},{"key":"2860_CR48","doi-asserted-by":"crossref","unstructured":"Aslantas, V., Bendes, E.: A new image quality metric for image fusion: the sum of the correlations of differences. AEU-Int. J. Electron. Commun. 69, 1890\u20131896 (2015)","DOI":"10.1016\/j.aeue.2015.09.004"},{"issue":"8","key":"2860_CR49","doi-asserted-by":"publisher","first-page":"141","DOI":"10.1117\/1.2009764","volume":"44","author":"C Xydeas","year":"2005","unstructured":"Xydeas, C., Petrovic, V.: Objective evaluation of signal-level image fusion performance. Opt. Eng. 44(8), 141\u2013155 (2005)","journal-title":"Opt. Eng."},{"key":"2860_CR50","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1016\/j.infrared.2013.11.008","volume":"62","author":"J Zhao","year":"2014","unstructured":"Zhao, J., Chen, Y., Feng, H., Zhihai, X., Li, Q.: Infrared image enhancement through saliency feature analysis based on multi-scale decomposition. Infrared Phys. Technol. 62, 86\u201393 (2014)","journal-title":"Infrared Phys. Technol."}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-023-02860-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-023-02860-0\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-023-02860-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,2,20]],"date-time":"2024-02-20T07:13:36Z","timestamp":1708413216000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-023-02860-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,11,16]]},"references-count":50,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2024,3]]}},"alternative-id":["2860"],"URL":"https:\/\/doi.org\/10.1007\/s11760-023-02860-0","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"type":"print","value":"1863-1703"},{"type":"electronic","value":"1863-1711"}],"subject":[],"published":{"date-parts":[[2023,11,16]]},"assertion":[{"value":"31 August 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"24 September 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 October 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 November 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Declaration on ethical approval is not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}]}}