{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,3]],"date-time":"2026-02-03T20:18:43Z","timestamp":1770149923738,"version":"3.49.0"},"reference-count":53,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2022,2,5]],"date-time":"2022-02-05T00:00:00Z","timestamp":1644019200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,2,5]],"date-time":"2022-02-05T00:00:00Z","timestamp":1644019200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61572162"],"award-info":[{"award-number":["61572162"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61572551"],"award-info":[{"award-number":["61572551"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61802095"],"award-info":[{"award-number":["61802095"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61702144"],"award-info":[{"award-number":["61702144"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Zhejiang Provincial Key Science and Technology Project Foundation","award":["2018C01012"],"award-info":[{"award-number":["2018C01012"]}]},{"DOI":"10.13039\/501100004731","name":"Natural Science Foundation of Zhejiang Province","doi-asserted-by":"publisher","award":["LQ17F020003"],"award-info":[{"award-number":["LQ17F020003"]}],"id":[{"id":"10.13039\/501100004731","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2023,3]]},"DOI":"10.1007\/s00371-021-02397-8","type":"journal-article","created":{"date-parts":[[2022,2,5]],"date-time":"2022-02-05T20:02:37Z","timestamp":1644091357000},"page":"1193-1204","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Action detection with two-stream enhanced detector"],"prefix":"10.1007","volume":"39","author":[{"given":"Min","family":"Zhang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6070-8524","authenticated-orcid":false,"given":"Haiyang","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Zhongjin","family":"Li","sequence":"additional","affiliation":[]},{"given":"Jie","family":"Chen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,2,5]]},"reference":[{"key":"2397_CR1","doi-asserted-by":"publisher","first-page":"546","DOI":"10.1109\/TIP.2020.3037472","volume":"30","author":"M Mandal","year":"2021","unstructured":"Mandal, M., Dhar, V., Mishra, A., Vipparthi, S.K., Abdel-Mottaleb, M.: 3DCD: scene independent end-to-end spatiotemporal feature learning framework for change detection in unseen videos. IEEE Trans. Image Process. 30, 546\u2013558 (2021)","journal-title":"IEEE Trans. Image Process."},{"key":"2397_CR2","doi-asserted-by":"publisher","first-page":"846","DOI":"10.1109\/TMM.2020.2990070","volume":"23","author":"J Deng","year":"2021","unstructured":"Deng, J., Pan, Y., Yao, T., Zhou, W., Li, H., Mei, T.: Single shot video object detector. IEEE Trans. Multimed. 23, 846\u2013858 (2021)","journal-title":"IEEE Trans. Multimed."},{"issue":"3","key":"2397_CR3","doi-asserted-by":"publisher","first-page":"567","DOI":"10.1007\/s00371-020-01824-6","volume":"37","author":"E Dong","year":"2021","unstructured":"Dong, E., Deng, M., Wang, Z.: A robust tracking algorithm with on online detector and high-confidence updating strategy. Vis. Comput. 37(3), 567\u2013585 (2021)","journal-title":"Vis. Comput."},{"key":"2397_CR4","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2019.105820","volume":"86","author":"C Dai","year":"2020","unstructured":"Dai, C., Liu, X., Lai, J.: Human action recognition using two-stream attention based LSTM networks. Appl. Soft Comput. 86, 105820 (2020)","journal-title":"Appl. Soft Comput."},{"issue":"1","key":"2397_CR5","doi-asserted-by":"publisher","first-page":"393","DOI":"10.1109\/TII.2019.2938527","volume":"16","author":"R Nawaratne","year":"2019","unstructured":"Nawaratne, R., Alahakoon, D., De Silva, D., Yu, X.: Spatiotemporal anomaly detection using deep learning for real-time video surveillance. IEEE Trans. Ind. Inf. 16(1), 393\u2013402 (2019)","journal-title":"IEEE Trans. Ind. Inf."},{"issue":"10","key":"2397_CR6","doi-asserted-by":"publisher","first-page":"2537","DOI":"10.1109\/TIFS.2019.2900907","volume":"14","author":"JT Zhou","year":"2019","unstructured":"Zhou, J.T., Du, J., Zhu, H., Peng, X., Liu, Y., Goh, R.S.M.: Anomalynet: an anomaly detection network for video surveillance. IEEE Trans. Inf. Forensics Secur. 14(10), 2537\u20132550 (2019)","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"2397_CR7","doi-asserted-by":"crossref","unstructured":"Wang, D., Devin, C., Cai, QZ., Yu, F., Darrell, T.: Deep object-centric policies for autonomous driving. In: International Conference on Robotics and Automation (ICRA), pp. 8853\u20138859 (2019)","DOI":"10.1109\/ICRA.2019.8794224"},{"key":"2397_CR8","doi-asserted-by":"crossref","unstructured":"Gu, R., Wang, G., Hwang, J.N.: Efficient multi-person hierarchical 3D pose estimation for autonomous driving. In: IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), pp. 163\u2013168 (2019)","DOI":"10.1109\/MIPR.2019.00036"},{"issue":"2","key":"2397_CR9","doi-asserted-by":"publisher","first-page":"371","DOI":"10.1007\/s00371-020-01805-9","volume":"37","author":"K Gong","year":"2021","unstructured":"Gong, K., Cao, Z., Xiao, Y., Fang, Z.: Abrupt-motion-aware lightweight visual tracking for unmanned aerial vehicles. Vis. Comput. 37(2), 371\u2013383 (2021)","journal-title":"Vis. Comput."},{"key":"2397_CR10","doi-asserted-by":"crossref","unstructured":"Chen, S., Jiang, Y.G.: Motion guided spatial attention for video captioning. In: Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), pp. 8191\u20138198 (2019)","DOI":"10.1609\/aaai.v33i01.33018191"},{"issue":"1","key":"2397_CR11","doi-asserted-by":"publisher","first-page":"20","DOI":"10.1049\/ccs.2018.0005","volume":"1","author":"C Li","year":"2019","unstructured":"Li, C., Yang, C., Giannetti, C.: Segmentation and generalisation for writing skills transfer from humans to robots. Cogn. Comput. Syst. 1(1), 20\u201325 (2019)","journal-title":"Cogn. Comput. Syst."},{"key":"2397_CR12","doi-asserted-by":"crossref","unstructured":"Zhou, Y., Sun, X., Zha, Z.J., Zeng, W.: MiCT: mixed 3d\/2d convolutional tube for human action recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 449\u2013458 (2018)","DOI":"10.1109\/CVPR.2018.00054"},{"issue":"1","key":"2397_CR13","doi-asserted-by":"publisher","first-page":"133","DOI":"10.1007\/s00371-019-01787-3","volume":"37","author":"L Wei","year":"2021","unstructured":"Wei, L., Cui, W., Hu, Z., Sun, H., Hou, S.: A single-shot multi-level feature reused neural network for object detection. Vis. Comput. 37(1), 133\u2013142 (2021)","journal-title":"Vis. Comput."},{"key":"2397_CR14","doi-asserted-by":"crossref","unstructured":"Li, Y., Lin, W., Wang, T et al.: Finding action tubes with a sparse-to-dense framework. In: Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), pp. 11466\u201311473 (2020)","DOI":"10.1609\/aaai.v34i07.6811"},{"key":"2397_CR15","doi-asserted-by":"crossref","unstructured":"Peng, X., Schmid, C.: Multi-region two-stream R-CNN for action detection. In: European Conference on Computer Vision (ECCV), pp. 744\u2013759 (2016)","DOI":"10.1007\/978-3-319-46493-0_45"},{"key":"2397_CR16","doi-asserted-by":"crossref","unstructured":"Saha, S., Singh, G., Sapienza, M., Torr, P.H., Cuzzolin, F.: Deep learning for detecting multiple space-time action tubes in videos. arXiv:1608.01529 (2016)","DOI":"10.5244\/C.30.58"},{"key":"2397_CR17","doi-asserted-by":"crossref","unstructured":"Xu, H., Das, A., Saenko, K.: R-c3d: region convolutional 3d network for temporal activity detection. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 5783\u20135792 (2017)","DOI":"10.1109\/ICCV.2017.617"},{"key":"2397_CR18","doi-asserted-by":"crossref","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., Berg, A.C.: SSD: single shot multibox detector. In: European Conference on Computer Vision (ECCV), pp. 21\u201337 (2016)","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"2397_CR19","doi-asserted-by":"crossref","unstructured":"Liu, Z., Xiang, Q., Tang, J., Wang, Y., Zhao, P.: Robust salient object detection for RGB images. Vis. Comput. 36(9), 1823\u20131835 (2020)","DOI":"10.1007\/s00371-019-01778-4"},{"key":"2397_CR20","doi-asserted-by":"crossref","unstructured":"Zhao, X., Zhang, L., Pang, Y., Lu, H., Zhang, L.: A single stream network for robust and real-time RGB-d salient object detection. In: European Conference on Computer Vision (ECCV), pp. 646\u2013662 (2020)","DOI":"10.1007\/978-3-030-58542-6_39"},{"issue":"2","key":"2397_CR21","doi-asserted-by":"publisher","first-page":"154","DOI":"10.1007\/s11263-013-0620-5","volume":"104","author":"JR Uijlings","year":"2013","unstructured":"Uijlings, J.R., Van De Sande, K.E., Gevers, T., Smeulders, A.W.: Selective search for object recognition. Int. J. Comput. Vis. 104(2), 154\u2013171 (2013)","journal-title":"Int. J. Comput. Vis."},{"key":"2397_CR22","doi-asserted-by":"crossref","unstructured":"Wang, X., Yang, M., Zhu, S., Lin, Y.: Regionlets for generic object detection. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 17\u201324 (2013)","DOI":"10.1109\/ICCV.2013.10"},{"key":"2397_CR23","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1016\/j.neucom.2020.01.085","volume":"396","author":"X Wu","year":"2020","unstructured":"Wu, X., Sahoo, D., Hoi, S.C.: Recent advances in deep learning for object detection. Neurocomputing 396, 39\u201364 (2020)","journal-title":"Neurocomputing"},{"key":"2397_CR24","doi-asserted-by":"crossref","unstructured":"Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-end object detection with transformers. In: European Conference on Computer Vision (ECCV), pp. 213\u2013229 (2020)","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"2397_CR25","doi-asserted-by":"crossref","unstructured":"Tan, M., Pang, R., Le, Q.V.: Efficientdet: scalable and efficient object detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10781\u201310790 (2020)","DOI":"10.1109\/CVPR42600.2020.01079"},{"key":"2397_CR26","doi-asserted-by":"crossref","unstructured":"Girshick, R., Donahue, J., Darrell, T., Malik, J.: Rich feature hierarchies for accurate object detection and semantic segmentation. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 580\u2013587 (2014)","DOI":"10.1109\/CVPR.2014.81"},{"key":"2397_CR27","doi-asserted-by":"crossref","unstructured":"Girshick, R., Donahue, J., Darrell, T., Malik, J.: Fast R-CNN. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 1440\u20131448 (2015)","DOI":"10.1109\/ICCV.2015.169"},{"key":"2397_CR28","unstructured":"Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. In: Advances in Neural Information Processing Systems (NIPS), pp. 91\u201399 (2015)"},{"key":"2397_CR29","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 779\u2013788 (2016)","DOI":"10.1109\/CVPR.2016.91"},{"key":"2397_CR30","doi-asserted-by":"crossref","unstructured":"Escorcia, V., Heilbron, F.C., Niebles, J.C., Ghanem, B.: DAPs: deep action proposals for action understanding. In: European Conference on Computer Vision (ECCV), pp. 768\u2013784 (2016)","DOI":"10.1007\/978-3-319-46487-9_47"},{"issue":"11","key":"2397_CR31","doi-asserted-by":"publisher","first-page":"2990","DOI":"10.1109\/TMM.2020.2965434","volume":"22","author":"J Li","year":"2020","unstructured":"Li, J., Liu, X., Zhang, W., Zhang, M., Song, J., Sebe, N.: Spatio-temporal attention networks for action recognition and detection. IEEE Trans. Multimed. 22(11), 2990\u20133001 (2020)","journal-title":"IEEE Trans. Multimed."},{"issue":"6","key":"2397_CR32","doi-asserted-by":"publisher","first-page":"1261","DOI":"10.1007\/s00371-019-01733-3","volume":"36","author":"J Cai","year":"2020","unstructured":"Cai, J., Hu, J.: 3D RANs: 3D residual attention networks for action recognition. Vis. Comput. 36(6), 1261\u20131270 (2020)","journal-title":"Vis. Comput."},{"key":"2397_CR33","doi-asserted-by":"crossref","unstructured":"Gkioxari, G., Malik, J.: Finding action tubes. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 759\u2013768 (2015)","DOI":"10.1109\/CVPR.2015.7298676"},{"key":"2397_CR34","doi-asserted-by":"crossref","unstructured":"Weinzaepfel, P., Harchaoui, Z., Schmid, C.: Learning to track for spatio-temporal action localization. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 3164\u20133172 (2015)","DOI":"10.1109\/ICCV.2015.362"},{"key":"2397_CR35","doi-asserted-by":"crossref","unstructured":"Li, D., Qiu, Z., Dai, Q., Yao, T., Mei, T.: Recurrent tubelet proposal and recognition networks for action detection. In: European Conference on Computer Vision (ECCV), pp. 303\u2013318 (2018)","DOI":"10.1007\/978-3-030-01231-1_19"},{"key":"2397_CR36","doi-asserted-by":"crossref","unstructured":"Kalogeiton, V., Weinzaepfel, P., Ferrari, V., Schmid, C.: Action tubelet detector for spatio-temporal action localization. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 4405\u20134413 (2017)","DOI":"10.1109\/ICCV.2017.472"},{"issue":"4","key":"2397_CR37","doi-asserted-by":"publisher","first-page":"831","DOI":"10.1007\/s00371-020-01833-5","volume":"37","author":"MY Abbass","year":"2021","unstructured":"Abbass, M.Y., Kwon, K.C., Kim, N., et al.: Efficient object tracking using hierarchical convolutional features model and correlation filters. Vis. Comput. 37(4), 831\u2013842 (2021)","journal-title":"Vis. Comput."},{"key":"2397_CR38","doi-asserted-by":"crossref","unstructured":"Yang, C., Xu, Y., Shi, J., Dai, B., Zhou, B.: Temporal pyramid network for action recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 591\u2013600 (2020)","DOI":"10.1109\/CVPR42600.2020.00067"},{"key":"2397_CR39","first-page":"13","volume-title":"Elliptic Partial Differential Equations of Second Order","author":"D Gilbarg","year":"2015","unstructured":"Gilbarg, D., Trudinger, N.S.: Elliptic Partial Differential Equations of Second Order, pp. 13\u201370. Springer, Berlin (2015)"},{"key":"2397_CR40","doi-asserted-by":"crossref","unstructured":"Singh, G., Saha, S., Sapienza, M., Torr, P.H., Cuzzolin, F.: Online real-time multiple spatiotemporal action localisation and prediction. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 3637\u20133646 (2017)","DOI":"10.1109\/ICCV.2017.393"},{"key":"2397_CR41","doi-asserted-by":"crossref","unstructured":"Rodriguez, M.D., Ahmed, J., Shah, M.: Action mach a spatio-temporal maximum average correlation height filter for action recognition. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1\u20138 (2008)","DOI":"10.1109\/CVPR.2008.4587727"},{"key":"2397_CR42","unstructured":"Soomro, K., Zamir, A.R., Shah, M.: UCF101: a dataset of 101 human actions classes from videos in the wild. arXiv:1212.0402 (2012)"},{"key":"2397_CR43","doi-asserted-by":"crossref","unstructured":"Jhuang, H., Gall, J., Zuffi, S., Schmid, C., Black, M.J.: Towards understanding action recognition. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 3192\u20133199 (2013)","DOI":"10.1109\/ICCV.2013.396"},{"key":"2397_CR44","doi-asserted-by":"crossref","unstructured":"Lan, T., Wang, Y., Mori, G.: Discriminative figure-centric models for joint action localization and recognition. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 2003\u20132010 (2011)","DOI":"10.1109\/ICCV.2011.6126472"},{"key":"2397_CR45","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556 (2014)"},{"key":"2397_CR46","doi-asserted-by":"crossref","unstructured":"Hou, R., Chen, C., Shah, M.: Tube convolutional neural network (T-CNN) for action detection in videos. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 5822\u20135831 (2017)","DOI":"10.1109\/ICCV.2017.620"},{"key":"2397_CR47","unstructured":"Li, W., Yuan, Z., Guo, D., Huang, L., Fang, X., Wang, C.: Deformable tube network for action detection in videos. arXiv:1907.01847 (2019)"},{"key":"2397_CR48","doi-asserted-by":"crossref","unstructured":"Pramono, R.R.A., Chen, Y.T., Fang, W.H.: Hierarchical self-attention network for action localization in videos. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 61\u201370 (2019)","DOI":"10.1109\/ICCV.2019.00015"},{"key":"2397_CR49","doi-asserted-by":"crossref","unstructured":"Wu, Y., Wang, H., Wang, S., Li, Q.: Enhanced action tubelet detector for spatio-temporal video action detection. In: IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 2388\u20132392 (2020)","DOI":"10.1109\/ICASSP40776.2020.9054394"},{"key":"2397_CR50","unstructured":"Zhou, X., Wang, D., Kr\u00e4henb\u00fchl, P.: Objects as points. arXiv:1904.07850 (2019)"},{"key":"2397_CR51","unstructured":"Bochkovskiy, A., Wang, C.Y., Liao, H.Y.M.: Yolov4: optimal speed and accuracy of object detection. arXiv:2004.10934 (2020)"},{"key":"2397_CR52","doi-asserted-by":"crossref","unstructured":"Li, Y., Lin, W., Wang, T., See, J., Qian, R., Xu, N., Xu, S.: Finding action tubes with a sparse-to-dense framework. In: Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), pp. 11466\u201311473 (2020)","DOI":"10.1609\/aaai.v34i07.6811"},{"key":"2397_CR53","doi-asserted-by":"crossref","unstructured":"Zhao, J., Snoek, C.G.: Dance with flow: two-in-one stream action detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 9935\u20139944 (2019)","DOI":"10.1109\/CVPR.2019.01017"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-021-02397-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-021-02397-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-021-02397-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,17]],"date-time":"2024-09-17T23:24:16Z","timestamp":1726615456000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-021-02397-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,2,5]]},"references-count":53,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2023,3]]}},"alternative-id":["2397"],"URL":"https:\/\/doi.org\/10.1007\/s00371-021-02397-8","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,2,5]]},"assertion":[{"value":"30 December 2021","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 February 2022","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}