{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,19]],"date-time":"2026-02-19T09:46:41Z","timestamp":1771494401805,"version":"3.50.1"},"reference-count":68,"publisher":"Springer Science and Business Media LLC","issue":"11","license":[{"start":{"date-parts":[[2021,7,14]],"date-time":"2021-07-14T00:00:00Z","timestamp":1626220800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,7,14]],"date-time":"2021-07-14T00:00:00Z","timestamp":1626220800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2022,11]]},"DOI":"10.1007\/s00371-021-02222-2","type":"journal-article","created":{"date-parts":[[2021,7,14]],"date-time":"2021-07-14T03:52:13Z","timestamp":1626234733000},"page":"3803-3817","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":13,"title":["CGAN: closure-guided attention network for salient object detection"],"prefix":"10.1007","volume":"38","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-3328-7883","authenticated-orcid":false,"given":"Dibyendu Kumar","family":"Das","sequence":"first","affiliation":[]},{"given":"Sahadeb","family":"Shit","sequence":"additional","affiliation":[]},{"given":"Dip Narayan","family":"Ray","sequence":"additional","affiliation":[]},{"given":"Somajyoti","family":"Majumder","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,7,14]]},"reference":[{"key":"2222_CR1","first-page":"438","volume":"2012","author":"A Borji","year":"2012","unstructured":"Borji, A.: Boosting bottom-up and top-down visual features for saliency estimation. IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2012, 438\u2013445 (2012)","journal-title":"IEEE Conference on Computer Vision and Pattern Recognition (CVPR)"},{"key":"2222_CR2","doi-asserted-by":"crossref","unstructured":"Wang, W. et al. Salient object detection in the deep learning era: an in-depth survey.\u00a0IEEE Trans. Pattern Anal. Mach. Intell. (2021): n. pag","DOI":"10.1109\/TPAMI.2021.3051099"},{"key":"2222_CR3","doi-asserted-by":"crossref","unstructured":"Li, Y., et al. The secrets of salient object segmentation. 2014 IEEE Conference on Computer Vision and Pattern Recognition, pp. 280\u2013287 (2014)","DOI":"10.1109\/CVPR.2014.43"},{"key":"2222_CR4","doi-asserted-by":"publisher","first-page":"107","DOI":"10.1016\/S0042-6989(01)00250-4","volume":"42","author":"DJ Parkhurst","year":"2002","unstructured":"Parkhurst, D.J., et al.: Modelling the role of salience in the allocation of overt visual attention. Vision. Res. 42, 107\u2013123 (2002)","journal-title":"Vision. Res."},{"key":"2222_CR5","doi-asserted-by":"publisher","first-page":"10427","DOI":"10.1007\/s11042-016-3430-2","volume":"76","author":"J Niu","year":"2016","unstructured":"Niu, J., et al.: Exploiting contrast cues for salient region detection. Multimed. Tools Appl. 76, 10427\u201310441 (2016)","journal-title":"Multimed. Tools Appl."},{"key":"2222_CR6","doi-asserted-by":"crossref","unstructured":"Wang, Q., Yuan, Y., Yan, P.: Visual saliency by selective contrast. IEEE Trans. Circ. Syst. Video Technol. 23.7, 1150\u20131155 (2012)","DOI":"10.1109\/TCSVT.2012.2226528"},{"key":"2222_CR7","doi-asserted-by":"crossref","unstructured":"Wei, Y. et al. Geodesic saliency using background priors. ECCV (2012)","DOI":"10.1007\/978-3-642-33712-3_3"},{"key":"2222_CR8","doi-asserted-by":"crossref","unstructured":"Zhu, W. et al. Saliency Optimization from Robust Background Detection. 2014 IEEE Conference on Computer Vision and Pattern Recognition (2014): 2814\u20132821","DOI":"10.1109\/CVPR.2014.360"},{"key":"2222_CR9","doi-asserted-by":"publisher","first-page":"56170","DOI":"10.1109\/ACCESS.2018.2873022","volume":"6","author":"Y Niu","year":"2018","unstructured":"Niu, Y., et al.: Salient Object Segmentation Based on Superpixel and Background Connectivity Prior. IEEE Access 6, 56170\u201356183 (2018)","journal-title":"IEEE Access"},{"key":"2222_CR10","doi-asserted-by":"publisher","first-page":"1392","DOI":"10.1109\/LGRS.2018.2838151","volume":"15","author":"W Ni","year":"2018","unstructured":"Ni, W., et al.: Background context-aware-based sar image saliency detection. IEEE Geosci. Remote Sens. Lett. 15, 1392\u20131396 (2018)","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"2222_CR11","doi-asserted-by":"publisher","first-page":"77","DOI":"10.1016\/S0004-3702(02)00399-5","volume":"146","author":"Y Sun","year":"2003","unstructured":"Sun, Y., Robert, B.F.: Object-based visual attention for computer vision. Artif. Intell 146, 77\u2013123 (2003)","journal-title":"Artif. Intell"},{"key":"2222_CR12","first-page":"18.1","volume":"8.14","author":"W Einh\u00e4user","year":"2008","unstructured":"Einh\u00e4user, W., et al.: Objects predict fixations better than early saliency. J. Vis. 8 8.14, 18.1-18.26 (2008)","journal-title":"J. Vis. 8"},{"key":"2222_CR13","doi-asserted-by":"publisher","first-page":"188","DOI":"10.1016\/j.neucom.2018.09.081","volume":"323","author":"Y Ji","year":"2019","unstructured":"Ji, Y., et al.: Graph model-based salient object detection using objectness and multiple saliency cues. Neurocomputing 323, 188\u2013202 (2019)","journal-title":"Neurocomputing"},{"issue":"2018","key":"2222_CR14","doi-asserted-by":"publisher","first-page":"3159","DOI":"10.1109\/TCYB.2017.2761361","volume":"48","author":"F Guo","year":"2018","unstructured":"Guo, F., et al.: Video saliency detection using object proposals. IEEE Trans Cybern 48(2018), 3159\u20133170 (2018)","journal-title":"IEEE Trans Cybern"},{"issue":"2012","key":"2222_CR15","doi-asserted-by":"publisher","first-page":"2189","DOI":"10.1109\/TPAMI.2012.28","volume":"34","author":"B Alexe","year":"2012","unstructured":"Alexe, B., et al.: Measuring the objectness of image windows. IEEE Trans. Pattern Anal. Mach. Intell. 34(2012), 2189\u20132202 (2012)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2222_CR16","unstructured":"Simonyan, K., Andrew Z. Very deep convolutional networks for large-scale image recognition. CoRR abs\/1409.1556 n. pag. (2015)"},{"key":"2222_CR17","doi-asserted-by":"crossref","unstructured":"Wu, Z. et al. Cascaded Partial Decoder for Fast and Accurate Salient Object Detection. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2019): 3902\u20133911","DOI":"10.1109\/CVPR.2019.00403"},{"key":"2222_CR18","doi-asserted-by":"crossref","unstructured":"Wang, T. et al. A stagewise refinement model for detecting salient objects in images. 2017 IEEE International Conference on Computer Vision (ICCV), pp 4039\u20134048, (2017)","DOI":"10.1109\/ICCV.2017.433"},{"key":"2222_CR19","doi-asserted-by":"publisher","first-page":"815","DOI":"10.1109\/TPAMI.2018.2815688","volume":"41","author":"Q Hou","year":"2019","unstructured":"Hou, Q., et al.: Deeply supervised salient object detection with short connections. IEEE Trans. Pattern Anal. Mach. Intell. 41, 815\u2013828 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2222_CR20","doi-asserted-by":"crossref","unstructured":"Zhang, L. et al. A Bi-Directional message passing model for salient object detection. 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 1741\u20131750 (2018)","DOI":"10.1109\/CVPR.2018.00187"},{"key":"2222_CR21","doi-asserted-by":"crossref","unstructured":"Deng, Z. et al. R3Net: Recurrent Residual Refinement Network for Saliency Detection. IJCAI (2018)","DOI":"10.24963\/ijcai.2018\/95"},{"key":"2222_CR22","doi-asserted-by":"crossref","unstructured":"Feng, M et al. Attentive Feedback Network for Boundary-Aware Salient Object Detection. 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2019): 1623\u20131632","DOI":"10.1109\/CVPR.2019.00172"},{"issue":"4","key":"2222_CR23","doi-asserted-by":"publisher","first-page":"501","DOI":"10.1037\/0096-3445.113.4.501","volume":"113","author":"J Duncan","year":"1984","unstructured":"Duncan, J.: Selective attention and the organization of visual information. J. Exp. Psychol. General 113(4), 501\u2013517 (1984)","journal-title":"J. Exp. Psychol. General"},{"key":"2222_CR24","doi-asserted-by":"publisher","first-page":"161","DOI":"10.1037\/0096-3445.123.2.161","volume":"123","author":"R Egly","year":"1994","unstructured":"Egly, R., Driver, J., Rafal, R.: Shifting visual attention between objects and locations: Evidence from normal and parietal-lesion patients. J. Exp. Psychol. Gen. 123, 161\u2013177 (1994)","journal-title":"J. Exp. Psychol. Gen."},{"key":"2222_CR25","doi-asserted-by":"publisher","first-page":"446","DOI":"10.3389\/fnhum.2016.00446","volume":"10","author":"F Marini","year":"2016","unstructured":"Marini, F., Carlo, A.M.: Gestalt perceptual organization of visual stimuli captures attention automatically: electrophysiological evidence. Front. Hum. Neurosci. 10, 446 (2016)","journal-title":"Front. Hum. Neurosci."},{"key":"2222_CR26","doi-asserted-by":"crossref","unstructured":"Hess R.F., May K.A., Dumoulin S.O. Oxford Handbook of Perceptual Organization. Oxford University Press; Oxford, UK: 2013. Contour integration: Psychophysical, neurophysiological and computational perspectives","DOI":"10.1093\/oxfordhb\/9780199686858.013.013"},{"key":"2222_CR27","doi-asserted-by":"crossref","unstructured":"Koffka, K. Principles of Gestalt Psychology. (2012)","DOI":"10.4324\/9781315009292"},{"key":"2222_CR28","doi-asserted-by":"crossref","unstructured":"Ronneberger, O. et al. U-Net: convolutional networks for biomedical image segmentation. MICCAI\u00a0(2015)","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"2222_CR29","doi-asserted-by":"publisher","first-page":"194","DOI":"10.1038\/35058500","volume":"2","author":"L Itti","year":"2001","unstructured":"Itti, L., Koch, C.: Computational modelling of visual attention. Nat. Rev. Neurosci. 2, 194\u2013203 (2001)","journal-title":"Nat. Rev. Neurosci."},{"key":"2222_CR30","doi-asserted-by":"publisher","first-page":"802","DOI":"10.1109\/TPAMI.2006.86","volume":"28","author":"OL Meur","year":"2006","unstructured":"Meur, O.L., et al.: A coherent computational approach to model bottom-up visual attention. IEEE Trans. Pattern Anal. Mach. Intell. 28, 802\u2013817 (2006)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2222_CR31","first-page":"433","volume":"2011","author":"N Murray","year":"2011","unstructured":"Murray, N., et al.: Saliency estimation using a non-parametric low-level vision model. CVPR 2011, 433\u2013440 (2011)","journal-title":"CVPR"},{"key":"2222_CR32","doi-asserted-by":"crossref","unstructured":"Xu, X., Wang, J. Extended non-local feature for visual saliency detection in low contrast images. ECCV Workshops (2018)","DOI":"10.1007\/978-3-030-11018-5_46"},{"key":"2222_CR33","unstructured":"Bruce, N. D. B., Tsotsos, J. K. Saliency based on information maximization. NIPS (2005)"},{"key":"2222_CR34","doi-asserted-by":"crossref","unstructured":"Li Y et al. (2009) Visual saliency based on conditional entropy Asian Conference on Computer Vision. Springer, Berlin","DOI":"10.1007\/978-3-642-12307-8_23"},{"issue":"12","key":"2222_CR35","doi-asserted-by":"publisher","first-page":"15.1","DOI":"10.1167\/9.12.15","volume":"9","author":"HJ Seo","year":"2009","unstructured":"Seo, H.J., Peyman, M.: Static and space-time visual saliency detection by self-resemblance. J. Vis 9(12), 15.1-15.27 (2009)","journal-title":"J. Vis"},{"key":"2222_CR36","doi-asserted-by":"publisher","first-page":"195","DOI":"10.1109\/TPAMI.2013.158","volume":"36","author":"S Lu","year":"2014","unstructured":"Lu, S., et al.: Robust and efficient saliency modelling from image co-occurrence histograms. IEEE Trans. Pattern Anal. Mach. Intell. 36, 195\u2013201 (2014)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2222_CR37","doi-asserted-by":"crossref","unstructured":"Liu, T. et al. (2007) Learning to detect a salient object. CVPR (2007)","DOI":"10.1109\/CVPR.2007.383047"},{"key":"2222_CR38","doi-asserted-by":"crossref","unstructured":"Harel, J. et al. Graph-Based Visual Saliency. NIPS (2006)","DOI":"10.7551\/mitpress\/7503.003.0073"},{"key":"2222_CR39","doi-asserted-by":"crossref","unstructured":"Achanta, R. et al. Frequency-tuned salient region detection. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 1597\u20131604 (2009)","DOI":"10.1109\/CVPR.2009.5206596"},{"key":"2222_CR40","first-page":"473","volume":"2011","author":"L Duan","year":"2011","unstructured":"Duan, L., et al.: Visual saliency detection by spatially weighted dissimilarity. CVPR 2011, 473\u2013480 (2011)","journal-title":"CVPR"},{"key":"2222_CR41","doi-asserted-by":"crossref","unstructured":"Perazzi, F. et al. (2012) Saliency filters: Contrast based filtering for salient region detection. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition, pp. 733\u2013740 (2012)","DOI":"10.1109\/CVPR.2012.6247743"},{"key":"2222_CR42","doi-asserted-by":"crossref","unstructured":"Jiang, H. et al. Salient object detection: a discriminative regional feature integration approach. 2013 IEEE Conference on Computer Vision and Pattern Recognition (2013): 2083\u20132090","DOI":"10.1109\/CVPR.2013.271"},{"key":"2222_CR43","doi-asserted-by":"publisher","first-page":"1191","DOI":"10.1007\/s00371-014-1005-7","volume":"31","author":"Y Shi","year":"2014","unstructured":"Shi, Y., et al.: Region contrast and supervised locality-preserving projection-based saliency detection. Vis. Comput. 31, 1191\u20131205 (2014)","journal-title":"Vis. Comput."},{"issue":"14","key":"2222_CR44","doi-asserted-by":"publisher","first-page":"997","DOI":"10.1049\/el.2014.0903","volume":"50","author":"Q Zhou","year":"2014","unstructured":"Zhou, Q.: Object-based attention: saliency detection using contrast via background prototypes. Electron. Lett. 50(14), 997\u2013999 (2014)","journal-title":"Electron. Lett."},{"key":"2222_CR45","first-page":"409","volume":"2011","author":"M-M Cheng","year":"2011","unstructured":"Cheng, M.-M., et al.: Global contrast based salient region detection. CVPR 2011, 409\u2013416 (2011)","journal-title":"CVPR"},{"key":"2222_CR46","first-page":"1","volume":"2016","author":"Q Zhou","year":"2016","unstructured":"Zhou, Q., et al.: Salient region detection by fusing foreground and background cues extracted from single image. Math. Prob. Eng. 2016, 1 (2016)","journal-title":"Math. Prob. Eng."},{"key":"2222_CR47","doi-asserted-by":"publisher","first-page":"330","DOI":"10.1007\/s11263-015-0822-0","volume":"115","author":"S He","year":"2015","unstructured":"He, S., et al.: SuperCNN: A superpixelwise convolutional neural network for salient object detection. Int. J. Comput. Vis. 115, 330\u2013344 (2015)","journal-title":"Int. J. Comput. Vis."},{"key":"2222_CR48","doi-asserted-by":"crossref","unstructured":"Wang, L. et al. Deep networks for saliency detection via local estimation and global search. 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) pp 3183\u20133192 (2015)","DOI":"10.1109\/CVPR.2015.7298938"},{"key":"2222_CR49","doi-asserted-by":"crossref","unstructured":"Zhao, R. et al. Saliency detection by multi-context deep learning. 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2015): 1265\u20131274","DOI":"10.1109\/CVPR.2015.7298731"},{"key":"2222_CR50","doi-asserted-by":"crossref","unstructured":"Liu, N., Han, J. DHSNet: Deep hierarchical saliency network for salient object detection. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016): 678\u2013686","DOI":"10.1109\/CVPR.2016.80"},{"key":"2222_CR51","doi-asserted-by":"crossref","unstructured":"Zhang, P et al. Amulet: aggregating multi-level convolutional features for salient object detection. 2017 IEEE International Conference on Computer Vision (ICCV)\u00a0(2017): 202\u2013211","DOI":"10.1109\/ICCV.2017.31"},{"key":"2222_CR52","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/s11263-017-1004-z","volume":"125","author":"S Xie","year":"2015","unstructured":"Xie, S., Zhuowen, Tu.: Holistically-nested edge detection. Int. J. Comput. Vision 125, 3\u201318 (2015)","journal-title":"Int. J. Comput. Vision"},{"key":"2222_CR53","doi-asserted-by":"crossref","unstructured":"Liu, N. et al. PiCANet: learning pixel-wise contextual attention for saliency detection. 2018 IEEE\/CVF conference on computer vision and pattern recognition, pp 3089\u20133098 (2018)","DOI":"10.1109\/CVPR.2018.00326"},{"key":"2222_CR54","doi-asserted-by":"crossref","unstructured":"Wu, Z et al. Stacked cross refinement network for edge-aware salient object detection. 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), pp 7263\u20137272 (2019)","DOI":"10.1109\/ICCV.2019.00736"},{"key":"2222_CR55","unstructured":"Wei, J. et al. \u201cF3Net: Fusion, Feedback and Focus for Salient Object Detection.\u201d\u00a0ArXiv\u00a0abs\/1911.11445 (2019): n. pag"},{"key":"2222_CR56","doi-asserted-by":"publisher","first-page":"107484","DOI":"10.1016\/j.patcog.2020.107484","volume":"107","author":"Q Zhang","year":"2020","unstructured":"Zhang, Q., et al.: Attention and boundary guided salient object detection. Pattern Recognit 107, 107484 (2020)","journal-title":"Pattern Recognit"},{"key":"2222_CR57","unstructured":"Ramachandran, P. et al. Stand-alone self-attention in vision models. NeurIPS (2019)."},{"key":"2222_CR58","doi-asserted-by":"publisher","first-page":"2011","DOI":"10.1109\/TPAMI.2019.2913372","volume":"42","author":"J Hu","year":"2020","unstructured":"Hu, J., et al.: Squeeze-and-Excitation Networks. IEEE Trans. Pattern Anal. Mach. Intell. 42, 2011\u20132023 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2222_CR59","doi-asserted-by":"crossref","unstructured":"Wang, X. et al. Non-local Neural Networks. 2018 IEEE\/CVF conference on computer vision and pattern recognition (2018): 7794\u20137803","DOI":"10.1109\/CVPR.2018.00813"},{"key":"2222_CR60","doi-asserted-by":"crossref","unstructured":"Bello, I. et al. Attention augmented convolutional networks. 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), pp 3285\u20133294, (2019).","DOI":"10.1109\/ICCV.2019.00338"},{"key":"2222_CR61","doi-asserted-by":"crossref","unstructured":"Woo, S. et al. CBAM: Convolutional Block Attention Module. ECCV (2018)","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"2222_CR62","doi-asserted-by":"publisher","first-page":"2825","DOI":"10.1109\/TIP.2019.2891104","volume":"28","author":"H Chen","year":"2019","unstructured":"Chen, H., Li, Y.: Three-stream attention-aware network for RGB-D salient object detection. IEEE Trans. Image Process. 28, 2825\u20132835 (2019)","journal-title":"IEEE Trans. Image Process."},{"key":"2222_CR63","doi-asserted-by":"publisher","first-page":"88","DOI":"10.1109\/TCYB.2020.2969255","volume":"51","author":"C Li","year":"2021","unstructured":"Li, C., et al.: ASIF-Net: attention steered interweave fusion network for RGB-D salient object detection. IEEE Trans. Cybern. 51, 88\u2013100 (2021)","journal-title":"IEEE Trans. Cybern."},{"key":"2222_CR64","doi-asserted-by":"publisher","first-page":"21","DOI":"10.1016\/j.knosys.2018.09.033","volume":"164","author":"H Yang","year":"2019","unstructured":"Yang, H., et al.: ContourGAN: Image contour detection with generative adversarial network. Knowl. Based Syst. 164, 21\u201328 (2019)","journal-title":"Knowl. Based Syst."},{"key":"2222_CR65","doi-asserted-by":"crossref","unstructured":"Zhao, J. et al. EGNet: Edge Guidance Network for Salient Object Detection. 2019 IEEE\/CVF International Conference on Computer Vision (ICCV) (2019), pp 8778\u20138787","DOI":"10.1109\/ICCV.2019.00887"},{"key":"2222_CR66","doi-asserted-by":"crossref","unstructured":"Wang, T et al. Detect globally, refine locally: a novel approach to saliency detection. 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (2018): 3127\u20133135","DOI":"10.1109\/CVPR.2018.00330"},{"key":"2222_CR67","doi-asserted-by":"crossref","unstructured":"Qin, X. et al. BASNet: boundary-aware salient object detection. 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 7471\u20137481, (2019)","DOI":"10.1109\/CVPR.2019.00766"},{"key":"2222_CR68","doi-asserted-by":"publisher","first-page":"107404","DOI":"10.1016\/j.patcog.2020.107404","volume":"106","author":"X Qin","year":"2020","unstructured":"Qin, X., et al.: U2-Net: Going deeper with nested u-structure for salient object detection. Pattern Recognit. 106, 107404 (2020)","journal-title":"Pattern Recognit."}],"updated-by":[{"DOI":"10.1007\/s00371-022-02680-2","type":"correction","label":"Correction","source":"publisher","updated":{"date-parts":[[2022,10,8]],"date-time":"2022-10-08T00:00:00Z","timestamp":1665187200000}}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-021-02222-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-021-02222-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-021-02222-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,4]],"date-time":"2024-09-04T02:34:31Z","timestamp":1725417271000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-021-02222-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,7,14]]},"references-count":68,"journal-issue":{"issue":"11","published-print":{"date-parts":[[2022,11]]}},"alternative-id":["2222"],"URL":"https:\/\/doi.org\/10.1007\/s00371-021-02222-2","relation":{"correction":[{"id-type":"doi","id":"10.1007\/s00371-022-02680-2","asserted-by":"object"}]},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,7,14]]},"assertion":[{"value":"21 June 2021","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 July 2021","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 October 2022","order":3,"name":"change_date","label":"Change Date","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"Correction","order":4,"name":"change_type","label":"Change Type","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"A Correction to this paper has been published:","order":5,"name":"change_details","label":"Change Details","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"https:\/\/doi.org\/10.1007\/s00371-022-02680-2","URL":"https:\/\/doi.org\/10.1007\/s00371-022-02680-2","order":6,"name":"change_details","label":"Change Details","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"We declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}