{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,29]],"date-time":"2026-04-29T20:06:12Z","timestamp":1777493172279,"version":"3.51.4"},"reference-count":46,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2024,3,29]],"date-time":"2024-03-29T00:00:00Z","timestamp":1711670400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,3,29]],"date-time":"2024-03-29T00:00:00Z","timestamp":1711670400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"Natural Science Foundation of China","doi-asserted-by":"crossref","award":["52172381"],"award-info":[{"award-number":["52172381"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Chongqing Postgraduate Joint Training Base","award":["JDLHPYJD2018003"],"award-info":[{"award-number":["JDLHPYJD2018003"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Real-Time Image Proc"],"published-print":{"date-parts":[[2024,6]]},"DOI":"10.1007\/s11554-024-01436-6","type":"journal-article","created":{"date-parts":[[2024,3,29]],"date-time":"2024-03-29T12:02:41Z","timestamp":1711713761000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":522,"title":["Slim-neck by GSConv: a lightweight-design for real-time detector architectures"],"prefix":"10.1007","volume":"21","author":[{"given":"Hulin","family":"Li","sequence":"first","affiliation":[]},{"given":"Jun","family":"Li","sequence":"additional","affiliation":[]},{"given":"Hanbing","family":"Wei","sequence":"additional","affiliation":[]},{"given":"Zheng","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Zhenfei","family":"Zhan","sequence":"additional","affiliation":[]},{"given":"Qiliang","family":"Ren","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,3,29]]},"reference":[{"key":"1436_CR1","doi-asserted-by":"publisher","unstructured":"Girshick, R., Donahue, J., Darrell, T., Malik, J.: Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA 23\u201328 June 2014, pp. 580\u2013587. https:\/\/doi.org\/10.1109\/CVPR.2014.81","DOI":"10.1109\/CVPR.2014.81"},{"key":"1436_CR2","doi-asserted-by":"publisher","unstructured":"Girshick, R.: Fast R-CNN. In: Proceedings of IEEE International Conference on Computer Vision (ICCV), Santiago, Chile 07\u201313 December 2015, pp. 1440\u20131448. https:\/\/doi.org\/10.1109\/ICCV.2015.169","DOI":"10.1109\/ICCV.2015.169"},{"issue":"6","key":"1436_CR3","doi-asserted-by":"publisher","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","volume":"39","author":"S Ren","year":"2017","unstructured":"Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. IEEE Trans Pattern Anal. Mach. Intel. 39(6), 1137\u20131149 (2017). https:\/\/doi.org\/10.1109\/TPAMI.2016.2577031","journal-title":"IEEE Trans Pattern Anal. Mach. Intel."},{"key":"1436_CR4","doi-asserted-by":"publisher","unstructured":"Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA 27\u201330 June 2016, pp. 779\u2013788. https:\/\/doi.org\/10.1109\/CVPR.2016.91","DOI":"10.1109\/CVPR.2016.91"},{"key":"1436_CR5","doi-asserted-by":"publisher","unstructured":"Redmon, J., Farhadi, A.: YOLO9000: better, faster, stronger. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA 21\u201326 July 2017; pp. 6517\u20136525, arXiv:1612.08242. [Online]. Available: https:\/\/arxiv.org\/abs\/1612.-08242v1. https:\/\/doi.org\/10.1109\/CVPR.2017.690","DOI":"10.1109\/CVPR.2017.690"},{"key":"1436_CR6","unstructured":"Redmon, J., Farhadi, A.: YOLOv3: an incremental improvement. arXiv eprints (2018). arXiv:1804.02767. [Online]. https:\/\/arxiv.org\/abs\/1804.02767"},{"key":"1436_CR7","unstructured":"Bochkovskiy, A., Wang, C.Y., Liao, H-Y. M.: Yolov4: optimal speed and accuracy of object detection. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020, arXiv:2004.10934. [Online]. https:\/\/arxiv.org\/abs\/2004.10934"},{"key":"1436_CR8","doi-asserted-by":"publisher","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C.: Reed, S., Fu, C.Y., Berg, A.C.: SSD: single shot multibox detector. In: Proceedings of European Conference on Computer Vision (ECCV), Sep. 2016, pp. 21\u201337. https:\/\/doi.org\/10.1007\/978-3-319-46448-0_2","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"1436_CR9","doi-asserted-by":"publisher","unstructured":"Fu, C.Y., Liu, W., Ranga, A., Tyagi, A., Berg, A.C.: DssD: deconvolutional single shot detector. arXiv eprints 2017, arXiv:1701.06659. [Online]. Available: https:\/\/arxiv.org\/abs\/1701.06659. https:\/\/doi.org\/10.48550\/arXiv.1701.06659","DOI":"10.48550\/arXiv.1701.06659"},{"key":"1436_CR10","doi-asserted-by":"publisher","unstructured":"Chollet, F.: Xception: deep learning with depthwise separable convolutions. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA 21\u201326 July 2017, pp. 1800\u20131807. [Online]. Available: https:\/\/arxiv.org\/abs\/1610.02357v1. https:\/\/doi.org\/10.1109\/CVPR.2017.195","DOI":"10.1109\/CVPR.2017.195"},{"key":"1436_CR11","unstructured":"Howard, A.G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., Andreetto, N., Hartwig, A.:. Mobilenets: efficient convolutional neural networks for mobile vision applications. arXiv eprints 2017, arXiv:1704.04861. [Online]. Available: https:\/\/arxiv.org\/abs-\/1704.04861"},{"key":"1436_CR12","doi-asserted-by":"publisher","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.: Mobilenetv2: inverted residuals and linear bottlenecks. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018, arXiv:1801.04381. [Online]. Available: https:\/\/arxiv.org\/abs\/1801.-04381v4. https:\/\/doi.org\/10.1109\/CVPR.2018.00474","DOI":"10.1109\/CVPR.2018.00474"},{"key":"1436_CR13","doi-asserted-by":"publisher","unstructured":"Howard, A., Sandler, M., Chu, G., Chen, L., Chen, B., Tan, M., Wang, W., Zhu, Y., Pang, R., Vasudevan, V., Le, Q.V., Adam, H.: Searching for MobileNetV3. In: Proceedings of IEEE International Conference on Computer Vision (ICCV), Apr. 2019, arXiv:1704.04861. [Online]. Available: https:\/\/arxiv.org\/abs\/1704.04861. https:\/\/doi.org\/10.1109\/ICCV.2019.00140","DOI":"10.1109\/ICCV.2019.00140"},{"key":"1436_CR14","doi-asserted-by":"publisher","unstructured":"Zhang, X., Zhou, X., Lin, M., Sun, J.: ShuffleNet: An extremely efficient convolutional neural network for mobile devices. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Jul. 2017, arXiv:1707.01083. [Online]. Available: https:\/\/arxiv.org\/abs\/1707.01083v1. https:\/\/doi.org\/10.1109\/CVPR.2018.00716","DOI":"10.1109\/CVPR.2018.00716"},{"key":"1436_CR15","doi-asserted-by":"publisher","unstructured":"Ma, N., Zhang, X., Zheng, H., Sun, J.: ShuffleNet V2: practical guidelines for efficient CNN architecture design. In: Proceedings of European Conference on Computer Vision (ECCV), 2018, arXiv:1807.11164. [Online]. Available: https:\/\/arxiv.org\/abs\/1807.11164v1. https:\/\/doi.org\/10.1007\/978-3-030-01264-9_8","DOI":"10.1007\/978-3-030-01264-9_8"},{"key":"1436_CR16","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-022-01657-x","author":"\u00c9 Zablocki","year":"2022","unstructured":"Zablocki, \u00c9., Ben-Younes, H., P\u00e9rez, P., et al.: Explainability of deep vision-based autonomous driving systems: review and challenges. Int. J. Comput. Vis. (2022). https:\/\/doi.org\/10.1007\/s11263-022-01657-x","journal-title":"Int. J. Comput. Vis."},{"key":"1436_CR17","doi-asserted-by":"publisher","unstructured":"K. Han; Y. Wang; Q. Tian; J. Guo; C. Xu; C. Xu. GhostNet: More features from cheapoperations. in Proc. IEEE Conf. Comput. Vis. Pattern Recog. (CVPR), Mar. 2020, arXiv:1911.11907. [Online]. Available: https:\/\/arxiv.org\/abs\/1911.11907 DOI: https:\/\/doi.org\/10.1109\/CVPR42600.2020.00165","DOI":"10.1109\/CVPR42600.2020.00165"},{"key":"1436_CR18","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1145\/3065386","volume":"25","author":"A Krizhevsky","year":"2012","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. Assoc. Comput. Mach. 25, 84\u201390 (2012). https:\/\/doi.org\/10.1145\/3065386","journal-title":"Assoc. Comput. Mach."},{"key":"1436_CR19","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. CoRR 2015; arXiv:1409.1556. [Online]. Available: https:\/\/arxiv.org\/abs\/1409.1556"},{"key":"1436_CR20","doi-asserted-by":"publisher","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 770\u2013778. https:\/\/doi.org\/10.1109\/CVPR.2016.90","DOI":"10.1109\/CVPR.2016.90"},{"key":"1436_CR21","doi-asserted-by":"crossref","unstructured":"Niu, W., Ma, X., Lin, S., Wang, S., Qian, X. Lin, X., Wang, Y. Ren, B.: PatDNN: achieving real-time DNN execution on mobile devices with pattern-based weight pruning. In: Proceedings of the Twenty-Fifth International Conference on Architectural Support for Programming Languages and Operating Systems. 2020, pp. 907\u2013922","DOI":"10.1145\/3373376.3378534"},{"key":"1436_CR22","doi-asserted-by":"publisher","unstructured":"Lin, T.-Y., Doll\u00e1r, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA 21\u201326 July 2017; pp. 936\u2013944. https:\/\/doi.org\/10.1109\/CVPR.2017.106","DOI":"10.1109\/CVPR.2017.106"},{"key":"1436_CR23","doi-asserted-by":"publisher","unstructured":"Wang, C.-Y., Bochkovskiy, A., Liao, H.-Y. M.: Scaled-yolov4: scaling cross stage partial network. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20\u201325, 2021; pp. 13024\u201313033. https:\/\/doi.org\/10.1109\/CVPR46437.2021.01283","DOI":"10.1109\/CVPR46437.2021.01283"},{"key":"1436_CR24","doi-asserted-by":"publisher","unstructured":"Tian, Z., Shen, C., Chen, H., He, T.: FCOS: fully convolutional one-stage object detection. In: International Conference on Computer Vision. (ICCV), Seoul, Korea (South) 27 October 2019\u201302 November 2019; pp. 9626\u20139635. https:\/\/doi.org\/10.1109\/ICCV.2019.00972","DOI":"10.1109\/ICCV.2019.00972"},{"key":"1436_CR25","doi-asserted-by":"publisher","unstructured":"Zhu, C., He, Y., Savvides, M.: Feature selective anchor-free module for single-shot object detection. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA 15\u201320 June 2019; pp. 840\u2013849. https:\/\/doi.org\/10.1109\/CVPR.2019.00093","DOI":"10.1109\/CVPR.2019.00093"},{"issue":"9","key":"1436_CR26","doi-asserted-by":"publisher","first-page":"1904","DOI":"10.1109\/TPAMI.2015.2389824","volume":"37","author":"K He","year":"2015","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 37(9), 1904\u20131916 (2015). https:\/\/doi.org\/10.1109\/TPAMI.2015.2389824","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1436_CR27","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2019.2913372","author":"J Hu","year":"2020","unstructured":"Hu, J., Shen, L., Albanie, S., Sun, G., Wu, E.: Squeeze-and-excitation networks. IEEE Trans. Pattern Anal. Mach. Intell. (2020). https:\/\/doi.org\/10.1109\/TPAMI.2019.2913372","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1436_CR28","doi-asserted-by":"publisher","unstructured":"Woo, S., Park, J., Lee, J., Kweon, I.S.: CBAM: convolutional block attention module. In: Proceedings of European Conference on Computer Vision (ECCV), Jul. 2018, arXiv:1807.06521. [Online]. Available: https:\/\/arxiv.org\/abs\/1807.06521v1. https:\/\/doi.org\/10.1007\/978-3-030-01234-2_1","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"1436_CR29","doi-asserted-by":"publisher","unstructured":"Hou, Q., Zhou, D., Feng, J.: Coordinate attention for efficient mobile network design. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Jun. 2021, arXiv:2103.02907. [Online]. Available: https:\/\/arxiv.org\/abs\/2103.02907. https:\/\/doi.org\/10.1109\/CVPR46437.2021.01350","DOI":"10.1109\/CVPR46437.2021.01350"},{"key":"1436_CR30","doi-asserted-by":"publisher","unstructured":"Huang, G., Liu, Z., Van Der Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA 21\u201326 2017, pp. 2261\u20132269. https:\/\/doi.org\/10.1109\/CVPR.2017.243","DOI":"10.1109\/CVPR.2017.243"},{"key":"1436_CR31","doi-asserted-by":"publisher","unstructured":"Lee, Y., Hwang, J.-w., Lee, S., Bae, Y., Park, J.: An energy and GPU-computation efficient backbone network for real-time object detection. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA 16\u201317 June 2019, pp. 752\u2013760. https:\/\/doi.org\/10.1109\/CVPRW.2019.00103","DOI":"10.1109\/CVPRW.2019.00103"},{"key":"1436_CR32","doi-asserted-by":"publisher","unstructured":"Wang, C.-Y., Mark Liao, H.-Y., Wu, Y.-H., Chen, P.-Y., Hsieh, J.-W., Yeh, I.-H.: CSPNet: a new backbone that can enhance learning capability of CNN. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA 14\u201319 June 2020; pp. 1571\u20131580. https:\/\/doi.org\/10.1109\/CVPRW50498.2020.00203","DOI":"10.1109\/CVPRW50498.2020.00203"},{"key":"1436_CR33","doi-asserted-by":"publisher","unstructured":"Yu, J., Jiang, Y., Wang, Z., Cao, Z., Huang, T.: UnitBox: An advanced object detection network. Association for Computing Machinery, New York, NY, USA Oct. 2016; pp. 516\u2013520. https:\/\/doi.org\/10.1145\/2964284.2967274","DOI":"10.1145\/2964284.2967274"},{"key":"1436_CR34","doi-asserted-by":"publisher","unstructured":"Rezatofighi, H., Tsoi, N., Gwak, J., Sadeghian, A., Reid, I., Savarese, S.: Generalized intersection over union: a metric and a loss for bounding box regression. In Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Jun. 2019; pp. 658\u2013666. https:\/\/doi.org\/10.1109\/CVPR.2019.00075","DOI":"10.1109\/CVPR.2019.00075"},{"issue":"7","key":"1436_CR35","doi-asserted-by":"publisher","first-page":"12993","DOI":"10.1609\/aaai.v34i07.6999","volume":"34","author":"Z Zheng","year":"2020","unstructured":"Zheng, Z., Wang, P., Liu, W., Li, J., Ye, R., Ren, D.: Distance-IoU Loss: faster and better learning for bounding box regression. Proc. AAAI Conf. Artif. Intell. (AAAI) 34(7), 12993\u201313000 (2020). https:\/\/doi.org\/10.1609\/aaai.v34i07.6999","journal-title":"Proc. AAAI Conf. Artif. Intell. (AAAI)"},{"key":"1436_CR36","doi-asserted-by":"publisher","DOI":"10.1109\/TCYB.2021.3095305","author":"Z Zheng","year":"2021","unstructured":"Zheng, Z., Wang, P., Ren, D., Liu, W., Ye, R., Hu, Q., Zuo, W.: Enhancing geometric factors in model learning and inference for object detection and instance segmentation. IEEE Trans. Cybern. (2021). https:\/\/doi.org\/10.1109\/TCYB.2021.3095305","journal-title":"IEEE Trans. Cybern."},{"key":"1436_CR37","doi-asserted-by":"publisher","unstructured":"Zhang, Y., Ren, W., Zhang, Z., Jia, Z., Wang, L., Tan, T.: Focal and efficient IoU loss for accurate bounding box regression. arXiv eprints 2021, arXiv:2101.08158 2021. [Online]. Available: https:\/\/arxiv.org\/abs-\/2101.08158. https:\/\/doi.org\/10.1016\/j.neucom.2022.07.042","DOI":"10.1016\/j.neucom.2022.07.042"},{"key":"1436_CR38","doi-asserted-by":"publisher","unstructured":"Ramachandran, P., Zoph, B., Le, Q.V.: Searching for activation functions. arXiv eprints 2017, arXiv:1710.05941. [Online]. https:\/\/doi.org\/10.48550\/arXiv.1710.05941","DOI":"10.48550\/arXiv.1710.05941"},{"key":"1436_CR39","unstructured":"Misra. Mish, D.: A self-regularized non-monotonic activation function. arXiv eprints 2020, arXiv:1908.08681. [Online]. https:\/\/arxiv.org\/abs\/1908.08681"},{"key":"1436_CR40","unstructured":"Glorot, X., Bordes, A., Bengio, Y.: Deep sparse rectifier neural networks. In: Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, 15, pp. 315\u2013323 (2011)"},{"key":"1436_CR41","unstructured":"Glenn, J.: Yolov5, 2022. https:\/\/github.com\/ultralytics\/yolov5"},{"key":"1436_CR42","doi-asserted-by":"crossref","unstructured":"Wang, C.Y., Bochkovskiy, A., Liao, H.Y.M.: YOLOv7: trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2023, pp. 7464\u20137475","DOI":"10.1109\/CVPR52729.2023.00721"},{"issue":"2","key":"1436_CR43","doi-asserted-by":"publisher","first-page":"380","DOI":"10.1109\/TMM.2019.2929005","volume":"22","author":"S Zhang","year":"2020","unstructured":"Zhang, S., Xie, Y., Wan, J., Xia, H., Li, S.Z., Guo, G.: WiderPerson: a diverse dataset for dense pedestrian detection in the wild. IEEE Trans. Multimedia 22(2), 380\u2013393 (2020). https:\/\/doi.org\/10.1109\/TMM.2019.2929005","journal-title":"IEEE Trans. Multimedia"},{"key":"1436_CR44","doi-asserted-by":"publisher","unstructured":"Everingham, M., Ali Eslami, S.M., Gool, L.V., Williams, C.K.I., Winn, J., Zisserman, A.: The PASCAL visual object classes challenge: a retrospective. Int. J. Comput. Vis. 111, 98\u2013136. (2015). https:\/\/doi.org\/10.1007\/s11263-014-0733-5","DOI":"10.1007\/s11263-014-0733-5"},{"key":"1436_CR45","doi-asserted-by":"publisher","unstructured":"Han, J., Liang, X., Xu, H., Chen, K., Hong, L., Ye, C., Zhang, W., Li, Z., Liang, X., Xu, C.: Soda10m: towards large-scale object detection benchmark for autonomous driving. arXiv eprints 2021, arXiv: 2106.11118. https:\/\/doi.org\/10.48550\/arXiv.2106.11118","DOI":"10.48550\/arXiv.2106.11118"},{"key":"1436_CR46","doi-asserted-by":"publisher","unstructured":"Xia, G.S., Bai, X., Ding, J., Zhu, Z., Belongie, S., Luo, J., Datcu, M., Pelillo, P., Zhang, L.: DOTA: a large-scale dataset for object detection in aerial images. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition 2018, Salt Lake City, UT, USA. https:\/\/doi.org\/10.1109\/CVPR.2018.00418","DOI":"10.1109\/CVPR.2018.00418"}],"container-title":["Journal of Real-Time Image Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-024-01436-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11554-024-01436-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-024-01436-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,6,6]],"date-time":"2024-06-06T08:31:03Z","timestamp":1717662663000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11554-024-01436-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,3,29]]},"references-count":46,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2024,6]]}},"alternative-id":["1436"],"URL":"https:\/\/doi.org\/10.1007\/s11554-024-01436-6","relation":{},"ISSN":["1861-8200","1861-8219"],"issn-type":[{"value":"1861-8200","type":"print"},{"value":"1861-8219","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,3,29]]},"assertion":[{"value":"13 October 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 February 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 March 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interests"}}],"article-number":"62"}}