{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,15]],"date-time":"2026-04-15T05:41:26Z","timestamp":1776231686376,"version":"3.50.1"},"reference-count":59,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2024,6,14]],"date-time":"2024-06-14T00:00:00Z","timestamp":1718323200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,6,14]],"date-time":"2024-06-14T00:00:00Z","timestamp":1718323200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2024,8]]},"DOI":"10.1007\/s00530-024-01367-z","type":"journal-article","created":{"date-parts":[[2024,7,2]],"date-time":"2024-07-02T10:03:08Z","timestamp":1719914588000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["LMFE-RDD: a road damage detector with a lightweight multi-feature extraction network"],"prefix":"10.1007","volume":"30","author":[{"given":"Qihan","family":"He","sequence":"first","affiliation":[]},{"given":"Zhongxu","family":"Li","sequence":"additional","affiliation":[]},{"given":"Wenyuan","family":"Yang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,6,14]]},"reference":[{"issue":"3","key":"1367_CR1","doi-asserted-by":"publisher","first-page":"257","DOI":"10.1109\/JPROC.2023.3238524","volume":"111","author":"Z Zou","year":"2023","unstructured":"Zou, Z., Chen, K., Shi, Z., Guo, Y., Ye, J.: Object detection in 20 years: a survey. Proceed. IEEE 111(3), 257\u2013276 (2023)","journal-title":"Proceed. IEEE"},{"issue":"12","key":"1367_CR2","doi-asserted-by":"publisher","first-page":"10812","DOI":"10.1109\/TNNLS.2022.3171553","volume":"34","author":"C Lin","year":"2023","unstructured":"Lin, C., Tian, D., Duan, X., Zhou, J., Zhao, D., Cao, D.: 3d-dfm: Anchor-free multimodal 3-d object detection with dynamic fusion module for autonomous driving. IEEE Trans. Neural Netw. Learn. Syst. 34(12), 10812\u201310822 (2023)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"1367_CR3","unstructured":"Surmann, H., Leinweber, A., Senkowski, G., Meine, J., Slomma, D.: Uavs and neural networks for search and rescue missions. arXiv preprint arXiv:2310.05512 (2023)"},{"issue":"1","key":"1367_CR4","doi-asserted-by":"publisher","first-page":"27","DOI":"10.1007\/s00530-023-01228-1","volume":"30","author":"J Yuan","year":"2024","unstructured":"Yuan, J., Hu, Y., Sun, Y., Wang, B., Yin, B.: A plug-and-play image enhancement model for end-to-end object detection in low-light condition. Multim. Syst. 30(1), 27 (2024)","journal-title":"Multim. Syst."},{"issue":"5","key":"1367_CR5","doi-asserted-by":"publisher","first-page":"2825","DOI":"10.1007\/s00530-023-01134-6","volume":"29","author":"G Wang","year":"2023","unstructured":"Wang, G., Li, J., Wu, Z., Xu, J., Shen, J., Yang, W.: Efficientface: an efficient deep network with feature enhancement for accurate face detection. Multim. Syst. 29(5), 2825\u20132839 (2023)","journal-title":"Multim. Syst."},{"issue":"1","key":"1367_CR6","doi-asserted-by":"publisher","first-page":"51","DOI":"10.1007\/s00530-023-01250-3","volume":"30","author":"H Bi","year":"2024","unstructured":"Bi, H., Tong, Y., Zhang, J., Zhang, C., Tong, J., Jin, W.: Depth alignment interaction network for camouflaged object detection. Multim. Syst. 30(1), 51 (2024)","journal-title":"Multim. Syst."},{"key":"1367_CR7","first-page":"91","volume":"28","author":"S Ren","year":"2015","unstructured":"Ren, S., He, K., Girshick, R., Sun, J.: Faster r-cnn: towards real-time object detection with region proposal networks. Adv. Neural Inform. Process. Syst. 28, 91\u201399 (2015)","journal-title":"Adv. Neural Inform. Process. Syst."},{"key":"1367_CR8","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P., Girshick, R.: Mask r-cnn. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2961\u20132969 (2017)","DOI":"10.1109\/ICCV.2017.322"},{"key":"1367_CR9","doi-asserted-by":"crossref","unstructured":"Cai, Z., Vasconcelos, N.: Cascade r-cnn: delving into high quality object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6154\u20136162 (2018)","DOI":"10.1109\/CVPR.2018.00644"},{"key":"1367_CR10","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"1367_CR11","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"1367_CR12","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4700\u20134708 (2017)","DOI":"10.1109\/CVPR.2017.243"},{"key":"1367_CR13","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 779\u2013788 (2016)","DOI":"10.1109\/CVPR.2016.91"},{"key":"1367_CR14","doi-asserted-by":"crossref","unstructured":"Redmon, J., Farhadi, A.: Yolo9000: better, faster, stronger. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7263\u20137271 (2017)","DOI":"10.1109\/CVPR.2017.690"},{"key":"1367_CR15","unstructured":"Redmon, J., Farhadi, A.: Yolov3: an incremental improvement. arXiv preprint arXiv:1804.02767 (2018)"},{"key":"1367_CR16","unstructured":"Bochkovskiy, A., Wang, C.Y., Liao, H.Y.M.: Yolov4: optimal speed and accuracy of object detection. arXiv preprint arXiv:2004.10934 (2020)"},{"key":"1367_CR17","doi-asserted-by":"publisher","unstructured":"Jocher, G.: YOLOv5 by Ultralytics (2020) https:\/\/doi.org\/10.5281\/zenodo.3908559","DOI":"10.5281\/zenodo.3908559"},{"key":"1367_CR18","unstructured":"Li, C., Li, L., Jiang, H., Weng, K., Geng, Y., Li, L., Ke, Z., Li, Q., Cheng, M., Nie, W., et al.: Yolov6: a single-stage object detection framework for industrial applications. arXiv preprint arXiv:2209.02976 (2022)"},{"key":"1367_CR19","doi-asserted-by":"crossref","unstructured":"Wang, C.Y., Bochkovskiy, A., Liao, H.Y.M.: Yolov7: trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7464\u20137475 (2023)","DOI":"10.1109\/CVPR52729.2023.00721"},{"key":"1367_CR20","unstructured":"Jocher, G., Chaurasia, A., Qiu, J.: Ultralytics YOLO (2023)"},{"key":"1367_CR21","doi-asserted-by":"crossref","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S.E., Fu, C.Y., Berg, A.C.: Ssd: single shot multibox detector. In: European Conference on Computer Vision, pp. 21\u201337 (2016)","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"1367_CR22","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Goyal, P., Girshick, R., He, K., Dollar, P.: Focal loss for dense object detection. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2980\u20132988 (2017)","DOI":"10.1109\/ICCV.2017.324"},{"issue":"4","key":"1367_CR23","first-page":"1068","volume":"19","author":"Y Shi","year":"2018","unstructured":"Shi, Y., Cui, Z., Xu, J., Zhao, Y., Jia, J.: Roadnet+: a novel approach to real-time road condition detection with deep neural networks. IEEE Trans. Intell. Transp. Syst. 19(4), 1068\u20131079 (2018)","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"issue":"11","key":"1367_CR24","first-page":"2925","volume":"18","author":"J Wang","year":"2017","unstructured":"Wang, J., Xia, Y., Jiao, L.: Crackdetect: real-time crack detection on roads using optimization-based active contour model. IEEE Trans. Intell. Transp. Syst. 18(11), 2925\u20132935 (2017)","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"issue":"3","key":"1367_CR25","first-page":"1286","volume":"21","author":"Z Wang","year":"2020","unstructured":"Wang, Z., Wang, L., Yang, X., Liu, S., Yang, J.: Real-time road surface crack detection using temporal feature pyramid networks. IEEE Trans. Intell. Transp. Syst. 21(3), 1286\u20131296 (2020)","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"1367_CR26","unstructured":"Tong, Z., Chen, Y., Xu, Z., Yu, R.: Wise-iou: bounding box regression loss with dynamic focusing mechanism. arXiv preprint arXiv:2301.10051 (2023)"},{"key":"1367_CR27","doi-asserted-by":"crossref","unstructured":"Arya, D., Maeda, H., Ghosh, S.K., Toshniwal, D., Sekimoto, Y.: Rdd2022: a multi-national image dataset for automatic road damage detection. arXiv preprint arXiv:2209.08538 (2022)","DOI":"10.1016\/j.dib.2021.107133"},{"key":"1367_CR28","doi-asserted-by":"crossref","unstructured":"Wang, C.Y., Bochkovskiy, A., Liao, H.Y.M.: Scaled-yolov4: scaling cross stage partial network. arXiv preprint arXiv:2011.08036 (2020)","DOI":"10.1109\/CVPR46437.2021.01283"},{"key":"1367_CR29","doi-asserted-by":"crossref","unstructured":"Liu, S., Qi, L., Qin, H., Shi, J., Jia, J.: Path aggregation network for instance segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8951\u20138959 (2018)","DOI":"10.1109\/CVPR.2018.00913"},{"key":"1367_CR30","unstructured":"Lin, T.Y., Goyal, P., Girshick, R., He, K., Doll\u00e1r, P.: Distribution-balanced loss for multi-label classification in long-tailed datasets. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10162\u201310171 (2020)"},{"key":"1367_CR31","unstructured":"Iandola, F.N., Han, S., Moskewicz, M.W., Ashraf, K., Dally, W.J., Keutzer, K.: Squeezenet: alexnet-level accuracy with 50x fewer parameters and $$<$$ 0.5mb model size. arXiv preprint arXiv:1602.07360 (2016)"},{"key":"1367_CR32","doi-asserted-by":"crossref","unstructured":"Gholami, A., Kwon, K., Wu, B., Tai, Y., Yue, X., Jin, P., Zhao, S., Keutzer, K.: Squeezenext: hardware-aware neural network design. arXiv preprint arXiv:1803.10615 (2018)","DOI":"10.1109\/CVPRW.2018.00215"},{"key":"1367_CR33","doi-asserted-by":"crossref","unstructured":"Wu, B., Zhou, X., Zhao, S., Yue, X., Keutzer, K.: Squeezesegv2: improved model structure and unsupervised domain adaptation for road-object segmentation from a lidar point cloud. In: International Conference on Robotics and Automation, pp. 4376\u20134382 (2019)","DOI":"10.1109\/ICRA.2019.8793495"},{"key":"1367_CR34","unstructured":"Howard, A.G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., Andreetto, M., Adam, H.: Mobilenets: efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861 (2017)"},{"key":"1367_CR35","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.C.: Mobilenetv2: inverted residuals and linear bottlenecks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4510\u20134520 (2018)","DOI":"10.1109\/CVPR.2018.00474"},{"key":"1367_CR36","doi-asserted-by":"crossref","unstructured":"Howard, A., Sandler, M., Chu, G., Chen, L.C., Chen, B., Tan, M., Wang, W., Zhu, Y., Pang, R., Vasudevan, V., : Searching for mobilenetv3. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1314\u20131324 (2019)","DOI":"10.1109\/ICCV.2019.00140"},{"key":"1367_CR37","doi-asserted-by":"crossref","unstructured":"Zhang, X., Zhou, X., Lin, M., Sun, J.: Shufflenet: an extremely efficient convolutional neural network for mobile devices. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6848\u20136856 (2018)","DOI":"10.1109\/CVPR.2018.00716"},{"key":"1367_CR38","unstructured":"Zhang, X., Zhou, X., Lin, M., Sun, J.: Shufflenetv2: practical guidelines for efficient cnn architecture design. In: European Conference on Computer Vision, pp. 122\u2013138 (2020)"},{"key":"1367_CR39","doi-asserted-by":"crossref","unstructured":"Han, K., Wang, Y., Tian, Q., Guo, J., Xu, C., Xu, C.: Ghostnet: more features from cheap operations. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1580\u20131589 (2020)","DOI":"10.1109\/CVPR42600.2020.00165"},{"key":"1367_CR40","unstructured":"Han, K., Wang, Y., Xu, C., Xu, C., Zhang, W., Tian, Q., Xu, C., Xu, C.: Ghostnetv2: more features and flexible decoupling. arXiv preprint arXiv:2104.06978 (2021)"},{"key":"1367_CR41","doi-asserted-by":"crossref","unstructured":"Rezatofighi, H., Tsoi, N., Gwak, J., Sadeghian, A., Reid, I., Savarese, S.: Generalized intersection over union: a metric and a loss for bounding box regression. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 658\u2013666 (2019)","DOI":"10.1109\/CVPR.2019.00075"},{"issue":"07","key":"1367_CR42","first-page":"12993","volume":"34","author":"Z Zheng","year":"2020","unstructured":"Zheng, Z., Wang, P., Liu, W., Li, J., Ye, R., Ren, D.: Distance-iou loss: faster and better learning for bounding box regression. Proceed. Conf. Artif. Intell. 34(07), 12993\u201313000 (2020)","journal-title":"Proceed. Conf. Artif. Intell."},{"key":"1367_CR43","unstructured":"Wu, Y., Kirillov, A., Massa, F., Lo, W.Y., Girshick, R.: Iou loss for 2d\/3d object detection. In: International Conference on 3D Vision, pp. 85\u201394 (2020)"},{"issue":"3","key":"1367_CR44","doi-asserted-by":"publisher","first-page":"3091","DOI":"10.1109\/TITS.2022.3221067","volume":"24","author":"C Lin","year":"2022","unstructured":"Lin, C., Tian, D., Duan, X., Zhou, J., Zhao, D., Cao, D.: Da-rdd: Toward domain adaptive road damage detection across different countries. IEEE Trans. Intell. Transp. Syst. 24(3), 3091\u20133103 (2022)","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"issue":"2254","key":"1367_CR45","doi-asserted-by":"publisher","first-page":"20220172","DOI":"10.1098\/rsta.2022.0172","volume":"381","author":"C Lin","year":"2023","unstructured":"Lin, C., Tian, D., Duan, X., Zhou, J.: Transcrack: revisiting fine-grained road crack detection with a transformer design. Phil. Trans. R. Soc. A 381(2254), 20220172 (2023)","journal-title":"Phil. Trans. R. Soc. A"},{"issue":"12","key":"1367_CR46","first-page":"10812","volume":"34","author":"J Gao","year":"2023","unstructured":"Gao, J., Chen, M., Xu, C.: Vectorized evidential learning for weakly-supervised temporal action localization. IEEE Trans. Pattern Anal. Mach. Intell. 34(12), 10812\u201310822 (2023)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1367_CR47","doi-asserted-by":"publisher","first-page":"5410","DOI":"10.1109\/TMM.2023.3333206","volume":"26","author":"Y Hu","year":"2023","unstructured":"Hu, Y., Gao, J., Dong, J., Fan, B., Liu, H.: Exploring rich semantics for open-set action recognition. IEEE Trans. Multim. 26, 5410\u20135421 (2023)","journal-title":"IEEE Trans. Multim."},{"issue":"3","key":"1367_CR48","doi-asserted-by":"publisher","first-page":"1646","DOI":"10.1109\/TCSVT.2021.3075470","volume":"32","author":"J Gao","year":"2021","unstructured":"Gao, J., Xu, C.: Learning video moment retrieval without a single annotated video. IEEE Trans. Circuits Syst. Video Technol. 32(3), 1646\u20131657 (2021)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"10","key":"1367_CR49","doi-asserted-by":"publisher","first-page":"3476","DOI":"10.1109\/TPAMI.2020.2985708","volume":"43","author":"J Gao","year":"2020","unstructured":"Gao, J., Zhang, T., Xu, C.: Learning to model relationships for zero-shot video classification. IEEE Trans. Pattern Anal. Mach. Intell. 43(10), 3476\u20133491 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1367_CR50","doi-asserted-by":"publisher","DOI":"10.1016\/j.autcon.2021.103935","volume":"132","author":"D Arya","year":"2021","unstructured":"Arya, D., Maeda, H., Ghosh, S.K., Toshniwal, D., Mraz, A., Kashiyama, T., Sekimoto, Y.: Deep learning-based road damage detection and classification for multiple countries. Autom. Constr. 132, 103935 (2021)","journal-title":"Autom. Constr."},{"key":"1367_CR51","doi-asserted-by":"publisher","first-page":"107133","DOI":"10.1016\/j.dib.2021.107133","volume":"36","author":"D Arya","year":"2021","unstructured":"Arya, D., Maeda, H., Ghosh, S.K., Toshniwal, D., Sekimoto, Y.: Rdd 2020: an annotated image dataset for automatic road damage detection using deep learning. Data Brief 36, 107133\u2013107144 (2021)","journal-title":"Data Brief"},{"key":"1367_CR52","doi-asserted-by":"crossref","unstructured":"Arya, D., Maeda, H., Ghosh, S.K., Toshniwal, D., Omata, H., Kashiyama, T., Sekimoto, Y.: Global road damage detection: State-of-the-art solutions. In: 2020 IEEE International Conference on Big Data, pp. 5533\u20135539 (2020)","DOI":"10.1109\/BigData50022.2020.9377790"},{"key":"1367_CR53","unstructured":"Nanting: Pavement disease product dataset. https:\/\/aistudio.baidu.com\/datasetdetail\/140177\/0 (2022)"},{"key":"1367_CR54","doi-asserted-by":"crossref","unstructured":"Liu, Z., Mao, H., Wu, C.Y., Feichtenhofer, C., Darrell, T., Xie, S.: A convnet for the 2020s. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11976\u201311986 (2022)","DOI":"10.1109\/CVPR52688.2022.01167"},{"key":"1367_CR55","doi-asserted-by":"crossref","unstructured":"Woo, S., Debnath, S., Hu, R., Chen, X., Liu, Z., Kweon, I.S., Xie, S.: Convnext v2: Co-designing and scaling convnets with masked autoencoders. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 16133\u201316142 (2023)","DOI":"10.1109\/CVPR52729.2023.01548"},{"key":"1367_CR56","unstructured":"Tan, M., Le, Q.: Efficientnet: Rethinking model scaling for convolutional neural networks. In: International Conference on Machine Learning, pp. 6105\u20136114 (2019)"},{"key":"1367_CR57","unstructured":"Mehta, S., Rastegari, M.: Mobilevit: Light-weight, general-purpose, and mobile-friendly vision transformer. arXiv preprint arXiv:2110.02178 (2021)"},{"key":"1367_CR58","unstructured":"Wadekar, S.N., Chaurasia, A.: Mobilevitv3: Mobile-friendly vision transformer with simple and effective fusion of local, global and input features (2022)"},{"key":"1367_CR59","unstructured":"Chen, H., Wang, Y., Guo, J., Tao, D.: Vanillanet: The power of minimalism in deep learning. arXiv preprint arXiv:2305.12972 (2023)"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01367-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-024-01367-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01367-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,8,22]],"date-time":"2024-08-22T08:22:35Z","timestamp":1724314955000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-024-01367-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,14]]},"references-count":59,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2024,8]]}},"alternative-id":["1367"],"URL":"https:\/\/doi.org\/10.1007\/s00530-024-01367-z","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,6,14]]},"assertion":[{"value":"6 February 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 May 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 June 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"No potential Conflict of interest was reported by the authors.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"176"}}