{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,11]],"date-time":"2025-09-11T20:47:13Z","timestamp":1757623633068,"version":"3.44.0"},"reference-count":45,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2025,5,24]],"date-time":"2025-05-24T00:00:00Z","timestamp":1748044800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2025,5,24]],"date-time":"2025-05-24T00:00:00Z","timestamp":1748044800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100000046","name":"National Research Council Canada","doi-asserted-by":"crossref","id":[{"id":"10.13039\/501100000046","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int. J. ITS Res."],"published-print":{"date-parts":[[2025,8]]},"abstract":"<jats:title>Abstract<\/jats:title>\n          <jats:p>Vulnerable road users (VRUs), such as pedestrians and bicyclists, face a higher risk of severe injuries and fatalities in road collisions, with intersections being particularly hazardous. Enhancing VRU safety at intersections is therefore critical for a safer transportation system. This study introduces a proof-of-concept system capable of detecting VRUs at intersections leveraging image data from vision sensors mounted on roadside infrastructure (e.g., traffic poles). The approach includes the development of a unique VRU detection dataset, comprising labeled images of various VRU types \u2013 adults, children, and bicyclists \u2013 captured under a range of illumination and weather conditions at real-world public intersections. This dataset addresses a notable gap in VRU detection research, as few datasets offer such environmental diversity from a roadside infrastructure perspective. The dataset was leveraged to train state-of-the-art deep learning models optimized for VRU detection. The models were evaluated using data from both public intersections and a controlled test facility, with particular focus on performance under challenging conditions such as snow and low nighttime visibility. Real-time performance benchmarking of the models was assessed, highlighting their effectiveness in dynamic environments. The results demonstrated that the best model achieved a mean average precision (mAP) of 82% in VRU detection while processing full-HD (1920 <jats:inline-formula>\n              <jats:alternatives>\n                <jats:tex-math>$$\\times $$<\/jats:tex-math>\n                <mml:math xmlns:mml=\"http:\/\/www.w3.org\/1998\/Math\/MathML\">\n                  <mml:mo>\u00d7<\/mml:mo>\n                <\/mml:math>\n              <\/jats:alternatives>\n            <\/jats:inline-formula> 1080) frames in real time at 75 ms. Additionally, major challenges in VRU detection at intersections were identified, and recommendations for future research directions were provided.<\/jats:p>","DOI":"10.1007\/s13177-025-00507-7","type":"journal-article","created":{"date-parts":[[2025,5,24]],"date-time":"2025-05-24T02:25:27Z","timestamp":1748053527000},"page":"1179-1196","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Intelligent Infrastructure for Enhancing Vulnerable Road User Safety using Machine Vision Technologies"],"prefix":"10.1007","volume":"23","author":[{"given":"Md Atiqur","family":"Rahman","sequence":"first","affiliation":[]},{"given":"Abdelhamid","family":"Mammeri","sequence":"additional","affiliation":[]},{"given":"Samy","family":"Metari","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,24]]},"reference":[{"key":"507_CR1","unstructured":"Bishop, C.M.: Pattern Recognition and Machine Learning. Information Science and Statistics. Springer, New York, NY (2006)"},{"key":"507_CR2","doi-asserted-by":"publisher","unstructured":"Bochkovskiy, A., Wang, C., Liao, H.M.: Yolov4: Optimal speed and accuracy of object detection. CoRR arXiv:2004.10934 (2020). https:\/\/doi.org\/10.48550\/arXiv.2004.10934","DOI":"10.48550\/arXiv.2004.10934"},{"issue":"8","key":"507_CR3","doi-asserted-by":"publisher","first-page":"1844","DOI":"10.1109\/TPAMI.2018.2858824","volume":"41","author":"M Braun","year":"2019","unstructured":"Braun, M., Krebs, S., Flohr, F., Gavrila, D.M.: Eurocity persons: A novel benchmark for person detection in traffic scenes. IEEE Trans. Pattern Anal. Mach. Intell. 41(8), 1844\u20131861 (2019). https:\/\/doi.org\/10.1109\/TPAMI.2018.2858824","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"5","key":"507_CR4","doi-asserted-by":"publisher","first-page":"1483","DOI":"10.1109\/TPAMI.2019.2956516","volume":"43","author":"Z Cai","year":"2021","unstructured":"Cai, Z., Vasconcelos, N.: Cascade r-cnn: High quality object detection and instance segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 43(5), 1483\u20131498 (2021). https:\/\/doi.org\/10.1109\/TPAMI.2019.2956516","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"507_CR5","doi-asserted-by":"publisher","unstructured":"Cao, J., Pang, Y., Xie, J., Khan, F., Shao, L.: From handcrafted to deep features for pedestrian detection: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 44(9), 4913\u20134934 . https:\/\/doi.org\/10.1109\/TPAMI.2021.3076733","DOI":"10.1109\/TPAMI.2021.3076733"},{"key":"507_CR6","doi-asserted-by":"publisher","unstructured":"Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-end object detection with transformers. In: Computer Vision \u2013 ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part I, p. 213\u2013229. Springer-Verlag, Berlin, Heidelberg (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_13","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"507_CR7","doi-asserted-by":"crossref","unstructured":"Chen, N., Xie, J., Nie, J., Cao, J., Shao, Z., Pang, Y.: Attentive alignment network for multispectral pedestrian detection. In: Proceedings of the 31st ACM International Conference on Multimedia, p. 3787\u20133795. Association for Computing Machinery, New York, NY, USA (2023)","DOI":"10.1145\/3581783.3613444"},{"key":"507_CR8","unstructured":"Chen, Z., Mishra, A., Rasmuson, M., Shridhar, M., Abbeel, P., Pathak, D.: You only look at one sequence: Rethinking transformer in vision through object detection. In: Advances in Neural Information Processing Systems (NeurIPS), vol.\u00a034 (2021)"},{"key":"507_CR9","unstructured":"Dai, J., Li, Y., He, K., Sun, J.: R-fcn: object detection via region-based fully convolutional networks. In: Proceedings of the 30th International Conference on Neural Information Processing Systems, NIPS\u201916, p. 379\u2013387. Curran Associates Inc., Red Hook, NY, USA (2016)"},{"key":"507_CR10","doi-asserted-by":"crossref","unstructured":"Dalal, N., Triggs, B.: Histograms of oriented gradients for human detection. Computer Vision and Pattern Recognition, 2005. CVPR 2005. IEEE Computer Society Conference on 1, 886\u2013893 (2005)","DOI":"10.1109\/CVPR.2005.177"},{"key":"507_CR11","unstructured":"DarkLabel: https:\/\/github.com\/darkpgmr\/DarkLabel. Retrieved from Github:"},{"issue":"4","key":"507_CR12","doi-asserted-by":"publisher","first-page":"743","DOI":"10.1109\/TPAMI.2011.155","volume":"34","author":"P Doll\u00e1r","year":"2012","unstructured":"Doll\u00e1r, P., Girshick, R., Belongie, S., Perona, P.: Pedestrian detection: An evaluation of the state of the art. IEEE Trans. Pattern Anal. Mach. Intell. 34(4), 743\u2013761 (2012). https:\/\/doi.org\/10.1109\/TPAMI.2011.155","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"507_CR13","doi-asserted-by":"publisher","unstructured":"Dollar, P., Wojek, C., Schiele, B., Perona, P.: Caltech pedestrians. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2009). https:\/\/doi.org\/10.1109\/CVPR.2009.5206631","DOI":"10.1109\/CVPR.2009.5206631"},{"issue":"8","key":"507_CR14","doi-asserted-by":"publisher","first-page":"1532","DOI":"10.1109\/TPAMI.2014.2300479","volume":"36","author":"P Doll\u00e1r","year":"2014","unstructured":"Doll\u00e1r, P., Appel, R., Belongie, S., Perona, P.: Fast feature pyramids for object detection. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI) 36(8), 1532\u20131545 (2014)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI)"},{"key":"507_CR15","doi-asserted-by":"crossref","unstructured":"Doll\u00e1r, P., Tu, Z., Perona, P., Belongie, S.J.: Integral channel features. In: BMVC, pp. 1\u201311. British Machine Vision Association (2009)","DOI":"10.5244\/C.23.91"},{"key":"507_CR16","doi-asserted-by":"publisher","unstructured":"Felzenszwalb, P.F., Girshick, R.B., McAllester, D.: Cascade object detection with deformable part models. In: 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 2241\u20132248 (2010). https:\/\/doi.org\/10.1109\/CVPR.2010.5539906","DOI":"10.1109\/CVPR.2010.5539906"},{"key":"507_CR17","doi-asserted-by":"publisher","unstructured":"Girshick, R.: Fast r-cnn. In: 2015 IEEE International Conference on Computer Vision (ICCV), pp. 1440\u20131448 (2015). https:\/\/doi.org\/10.1109\/ICCV.2015.169","DOI":"10.1109\/ICCV.2015.169"},{"issue":"1","key":"507_CR18","doi-asserted-by":"publisher","first-page":"142","DOI":"10.1109\/TPAMI.2015.2437384","volume":"38","author":"R Girshick","year":"2016","unstructured":"Girshick, R., Donahue, J., Darrell, T., Malik, J.: Region-based convolutional networks for accurate object detection and segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 38(1), 142\u2013158 (2016). https:\/\/doi.org\/10.1109\/TPAMI.2015.2437384","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"507_CR19","doi-asserted-by":"publisher","unstructured":"Hao, Y., Pei, H., Lyu, Y., Yuan, Z., Rizzo, J., Wang, Y., Fang, Y.: Understanding the impact of image quality and distance of objects to object detection performance. In: 2023 IEEE\/RSJ International Conference on Intelligent Robots and Systems, IROS 2023, IEEE International Conference on Intelligent Robots and Systems, pp. 11436\u201311442. Institute of Electrical and Electronics Engineers Inc. (2023). https:\/\/doi.org\/10.1109\/IROS55552.2023.10342139","DOI":"10.1109\/IROS55552.2023.10342139"},{"key":"507_CR20","doi-asserted-by":"publisher","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P., Girshick, R.: Mask r-cnn. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 2980\u20132988 (2017).https:\/\/doi.org\/10.1109\/ICCV.2017.322","DOI":"10.1109\/ICCV.2017.322"},{"key":"507_CR21","doi-asserted-by":"publisher","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770\u2013778 (2016). https:\/\/doi.org\/10.1109\/CVPR.2016.90","DOI":"10.1109\/CVPR.2016.90"},{"key":"507_CR22","doi-asserted-by":"publisher","unstructured":"Huang, J., Rathod, V., Sun, C., Zhu, M., Korattikara, A., Fathi, A., Fischer, I., Wojna, Z., Song, Y., Guadarrama, S., Murphy, K.: Speed\/accuracy trade-offs for modern convolutional object detectors. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, July 21-26, 2017, pp. 3296\u20133297. IEEE Computer Society (2017). https:\/\/doi.org\/10.1109\/CVPR.2017.351","DOI":"10.1109\/CVPR.2017.351"},{"key":"507_CR23","doi-asserted-by":"crossref","unstructured":"Jia, X., Zhu, C., Li, M., Tang, W., Zhou, W.: Llvip: A visible-infrared paired dataset for low-light vision. In: IEEE International Conference on Computer Vision, pp. 3496\u20133504 (2021)","DOI":"10.1109\/ICCVW54120.2021.00389"},{"key":"507_CR24","doi-asserted-by":"publisher","unstructured":"Jocher, G., Chaurasia, A., Stoken, A., Borovec, J., NanoCode012, K., Y., o.: ultralytics\/yolov5: v6.2 - yolov5 classification models, apple m1, reproducibility, clearml and deci.ai integrations. https:\/\/doi.org\/10.5281\/zenodo.7002879","DOI":"10.5281\/zenodo.7002879"},{"key":"507_CR25","unstructured":"Li, X., Flohr, F., Yang, Y., Xiong, H., Braun, M., Pan, S., Li, K., Gavrila, D.: A new benchmark for vision-based cyclist detection. In: 2016 IEEE Intelligent Vehicles Symposium (IV), pp. 1028\u20131033 (2016)"},{"key":"507_CR26","doi-asserted-by":"publisher","unstructured":"Liang, J.A., Ding, J.J.: Swin transformer for pedestrian and occluded pedestrian detection. In: 2024 IEEE International Symposium on Circuits and Systems (ISCAS), pp. 1\u20135 (2024). https:\/\/doi.org\/10.1109\/ISCAS58744.2024.10558302","DOI":"10.1109\/ISCAS58744.2024.10558302"},{"key":"507_CR27","doi-asserted-by":"publisher","unstructured":"Lin, T.Y., Dollar, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2017). https:\/\/doi.org\/10.1109\/CVPR.2017.106","DOI":"10.1109\/CVPR.2017.106"},{"key":"507_CR28","doi-asserted-by":"publisher","unstructured":"Lin, T.Y., Goyal, P., Girshick, R., He, K., Doll\u00e1r, P.: Focal loss for dense object detection. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 2999\u20133007 (2017). https:\/\/doi.org\/10.1109\/ICCV.2017.324","DOI":"10.1109\/ICCV.2017.324"},{"key":"507_CR29","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Maire, M., Belongie, S., Bourdev, L., Girshick, R., Hays, J., Perona, P., Ramanan, D., Doll\u00e1r, P., Zitnick, C.L.: Microsoft coco: Common objects in context. In: European Conference on Computer Vision (ECCV), pp. 740\u2013755. Springer (2014)","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"507_CR30","doi-asserted-by":"crossref","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., Berg, A.: Ssd: Single shot multibox detector. In: European conference on computer vision. Lecture Notes in Computer Science, vol. 9905. Springer (2016)","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"507_CR31","doi-asserted-by":"publisher","first-page":"207","DOI":"10.1109\/TIP.2020.3034487","volume":"30","author":"Y Pang","year":"2021","unstructured":"Pang, Y., Cao, J., Li, Y., Xie, J., Sun, H., Gong, J.: Tju-dhd: A diverse high-resolution dataset for object detection. IEEE Trans Image Process 30, 207\u2013219 (2021). https:\/\/doi.org\/10.1109\/TIP.2020.3034487","journal-title":"IEEE Trans Image Process"},{"key":"507_CR32","doi-asserted-by":"publisher","unstructured":"Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You Only Look Once: Unified, Real-Time Object Detection . In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 779\u2013788. IEEE Computer Society (2016). https:\/\/doi.org\/10.1109\/CVPR.2016.91","DOI":"10.1109\/CVPR.2016.91"},{"key":"507_CR33","doi-asserted-by":"publisher","unstructured":"Redmon, J., Farhadi, A.: YOLO9000: better, faster, stronger. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, July 21-26, 2017, pp. 6517\u20136525. IEEE Computer Society (2017). https:\/\/doi.org\/10.1109\/CVPR.2017.690","DOI":"10.1109\/CVPR.2017.690"},{"key":"507_CR34","unstructured":"Redmon, J., Farhadi, A.: Yolov3: An incremental improvement. CoRR arXiv:1804.02767 (2018)"},{"key":"507_CR35","unstructured":"Ren, S., He, K., Girshick, R., Sun, J.: Faster r-cnn: Towards real-time object detection with region proposal networks. In: Cortes, C., Lawrence, N., Lee, D., Sugiyama, M., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol.\u00a028. Curran Associates, Inc. (2015)"},{"key":"507_CR36","unstructured":"Road Network Operations & Intelligent Transport Systems\u2013PIARC (World Road Association): Safety of vulnerable road users (2023). https:\/\/rno-its.piarc.org\/en\/network-operations-its-road-safety\/vulnerable-road-users. Accessed 15 Oct 2024"},{"key":"507_CR37","doi-asserted-by":"crossref","unstructured":"Song, X., Chen, B., Li, P., He, J.Y., Wang, B., Geng, Y., Xie, X., Zhang, H.: Optimal proposal learning for deployable end-to-end pedestrian detection. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3250\u20133260 (2023)","DOI":"10.1109\/CVPR52729.2023.00317"},{"key":"507_CR38","unstructured":"Statistics Canada: Pedestrian fatalities in canada, 2018 to 2020 (2023). https:\/\/www150.statcan.gc.ca\/n1\/pub\/11-627-m\/11-627-m2023059-eng.htm. Accessed 9 Oct 2024"},{"key":"507_CR39","unstructured":"Touvron, H., Cord, M., Douze, M., Massa, F., Sablayrolles, A., Jegou, H.: Training data-efficient image transformers & distillation through attention. In: 38th International Conference on Machine Learning, pp. 10347\u201310357 (2021)"},{"key":"507_CR40","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L., Polosukhin, I.: Attention is all you need. In: Advances in Neural Information Processing Systems (NeurIPS), vol.\u00a030 (2017)"},{"key":"507_CR41","doi-asserted-by":"publisher","unstructured":"Viola, P., Jones, M.: Robust real-time face detection. In: Proceedings Eighth IEEE International Conference on Computer Vision. ICCV 2001, vol.\u00a02, pp. 747\u2013747 (2001). https:\/\/doi.org\/10.1109\/ICCV.2001.937709","DOI":"10.1109\/ICCV.2001.937709"},{"key":"507_CR42","doi-asserted-by":"crossref","unstructured":"Wolf, T., Debut, L., Sanh, V., Chaumond, J., Delangue, C., Moi, A., Cistac, P., Rault, T., Louf, R., Funtowicz, M., Davison, J., Shleifer, S., von Platen, P., Ma, C., Jernite, Y., Plu, J., Xu, C., Scao, T.L., Gugger, S., Drame, M., Lhoest, Q., Rush, A.M.: Transformers: State-of-the-art natural language processing (2020). https:\/\/github.com\/huggingface\/transformers","DOI":"10.18653\/v1\/2020.emnlp-demos.6"},{"key":"507_CR43","unstructured":"World Health Organization: Global status report on road safety 2023 (2023). https:\/\/iris.who.int\/bitstream\/handle\/10665\/375016\/9789240086517-eng.pdf?sequence=1. Accessed 9 Oct 2024"},{"key":"507_CR44","doi-asserted-by":"crossref","unstructured":"Zhang, S., Benenson, R., Schiele, B.: Citypersons: A diverse dataset for pedestrian detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4457\u20134465 (2017)","DOI":"10.1109\/CVPR.2017.474"},{"issue":"11","key":"507_CR45","doi-asserted-by":"publisher","first-page":"7380","DOI":"10.1109\/TPAMI.2021.3119563","volume":"44","author":"P Zhu","year":"2022","unstructured":"Zhu, P., Wen, L., Du, D., Bian, X., Fan, H., Hu, Q., Ling, H.: Detection and tracking meet drones challenge. IEEE Trans. Pattern Anal. Mach. Intell. 44(11), 7380\u20137399 (2022). https:\/\/doi.org\/10.1109\/TPAMI.2021.3119563","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."}],"container-title":["International Journal of Intelligent Transportation Systems Research"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13177-025-00507-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s13177-025-00507-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13177-025-00507-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,9]],"date-time":"2025-09-09T07:52:42Z","timestamp":1757404362000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s13177-025-00507-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,24]]},"references-count":45,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2025,8]]}},"alternative-id":["507"],"URL":"https:\/\/doi.org\/10.1007\/s13177-025-00507-7","relation":{},"ISSN":["1348-8503","1868-8659"],"issn-type":[{"type":"print","value":"1348-8503"},{"type":"electronic","value":"1868-8659"}],"subject":[],"published":{"date-parts":[[2025,5,24]]},"assertion":[{"value":"4 December 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 April 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 May 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"24 May 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Not applicable.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval and consent to participate"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}},{"value":"The authors declare that they have no competing interests.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}]}}