{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,10]],"date-time":"2025-06-10T16:32:46Z","timestamp":1749573166543},"reference-count":43,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2023,10,11]],"date-time":"2023-10-11T00:00:00Z","timestamp":1696982400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,10,11]],"date-time":"2023-10-11T00:00:00Z","timestamp":1696982400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Ministry of Science and Technology of thePeople\u2019s Republic of China","award":["2018AAA0103004"],"award-info":[{"award-number":["2018AAA0103004"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2023,12]]},"DOI":"10.1007\/s00530-023-01185-9","type":"journal-article","created":{"date-parts":[[2023,10,11]],"date-time":"2023-10-11T14:03:12Z","timestamp":1697032992000},"page":"3639-3653","update-policy":"http:\/\/dx.doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Workpiece tracking based on improved SiamFC++ and virtual dataset"],"prefix":"10.1007","volume":"29","author":[{"given":"Kaisi","family":"Yang","sequence":"first","affiliation":[]},{"given":"Lianyu","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Chenglin","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,10,11]]},"reference":[{"key":"1185_CR1","doi-asserted-by":"publisher","DOI":"10.1145\/3478513.3480509","author":"P Chandran","year":"2021","unstructured":"Chandran, P., Winberg, S., Zoss, G., Riviere, J., Gross, M., Gotardo, P., Bradley, D.: Rendering with style: combining traditional and neural approaches for high-quality face rendering. ACM Trans. Graph. (2021). https:\/\/doi.org\/10.1145\/3478513.3480509","journal-title":"ACM Trans. Graph."},{"issue":"4","key":"1185_CR2","doi-asserted-by":"publisher","first-page":"535","DOI":"10.1007\/s41095-022-0280-x","volume":"8","author":"J Zhu","year":"2022","unstructured":"Zhu, J., Zhao, S., Xu, Y., Meng, X., Wang, L., Yan, L.-Q.: Recent advances in glinty appearance rendering. Comput. Vis. Media 8(4), 535\u2013552 (2022). https:\/\/doi.org\/10.1007\/s41095-022-0280-x","journal-title":"Comput. Vis. Media"},{"issue":"19","key":"1185_CR3","doi-asserted-by":"publisher","first-page":"1","DOI":"10.3390\/app11199065","volume":"11","author":"M Choi","year":"2021","unstructured":"Choi, M., Park, J.-H., Zhang, Q., Hong, B.-S., Kim, C.-H.: Deep representation of a normal map for screen-space fluid rendering. Appl. Sci. Basel 11(19), 1 (2021). https:\/\/doi.org\/10.3390\/app11199065","journal-title":"Appl. Sci. Basel"},{"issue":"8","key":"1185_CR4","doi-asserted-by":"publisher","first-page":"3505","DOI":"10.1109\/TVCG.2021.3074607","volume":"27","author":"C Neuhauser","year":"2021","unstructured":"Neuhauser, C., Wang, J., Westermann, R.: Interactive focus plus context rendering for hexahedral mesh inspection. IEEE Trans. Vis. Comput. Graph. 27(8), 3505\u20133518 (2021). https:\/\/doi.org\/10.1109\/TVCG.2021.3074607","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"1185_CR5","doi-asserted-by":"publisher","unstructured":"Chen, X., Yan, B., Zhu, J., Wang, D., Yang, X., Lu, H.: Transformer tracking. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8122\u20138131 (2021). https:\/\/doi.org\/10.1109\/CVPR46437.2021.00803","DOI":"10.1109\/CVPR46437.2021.00803"},{"key":"1185_CR6","doi-asserted-by":"publisher","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Kai, L., Li, F.-F.: Imagenet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255 (2009). https:\/\/doi.org\/10.1109\/CVPR.2009.5206848","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"1185_CR7","doi-asserted-by":"publisher","unstructured":"Xiao, H., Rasul, K., Vollgraf, R.: Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms (2017). https:\/\/doi.org\/10.48550\/arXiv.1708.07747","DOI":"10.48550\/arXiv.1708.07747"},{"key":"1185_CR8","doi-asserted-by":"publisher","unstructured":"Recht, B., Roelofs, R., Schmidt, L., Shankar, V.: Do CIFAR-10 Classifiers Generalize to CIFAR-10? (2018). https:\/\/doi.org\/10.48550\/arXiv.1806.00451","DOI":"10.48550\/arXiv.1806.00451"},{"key":"1185_CR9","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3179507","author":"Y Liao","year":"2022","unstructured":"Liao, Y., Xie, J., Geiger, A.: Kitti-360: a novel dataset and benchmarks for urban scene understanding in 2d and 3d. IEEE Trans. Pattern Anal. Mach. Intell. (2022). https:\/\/doi.org\/10.1109\/TPAMI.2022.3179507","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1185_CR10","doi-asserted-by":"publisher","unstructured":"Fan, H., Bai, H., Lin, L., Yang, F., Chu, P., Deng, G., Yu, S., Harshit, Huang, M., Liu, J., Xu, Y., Liao, C., Yuan, L., Ling, H.: Lasot: a high-quality large-scale single object tracking benchmark. Int. J. Comput. Vis. 129(2), 439\u2013461 (2021). https:\/\/doi.org\/10.1007\/s11263-020-01387-y","DOI":"10.1007\/s11263-020-01387-y"},{"issue":"5","key":"1185_CR11","doi-asserted-by":"publisher","first-page":"1562","DOI":"10.1109\/TPAMI.2019.2957464","volume":"43","author":"L Huang","year":"2021","unstructured":"Huang, L., Zhao, X., Huang, K.: Got-10k: a large high-diversity benchmark for generic object tracking in the wild. IEEE Trans. Pattern Anal. Mach. Intell. 43(5), 1562\u20131577 (2021). https:\/\/doi.org\/10.1109\/TPAMI.2019.2957464","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"2","key":"1185_CR12","doi-asserted-by":"publisher","first-page":"171","DOI":"10.1109\/tiv.2018.2886678","volume":"4","author":"Y Kang","year":"2019","unstructured":"Kang, Y., Yin, H., Berger, C.: Test your self-driving algorithm: an overview of publicly available driving datasets and virtual testing environments. IEEE Trans. Intell. Veh. 4(2), 171\u2013185 (2019). https:\/\/doi.org\/10.1109\/tiv.2018.2886678","journal-title":"IEEE Trans. Intell. Veh."},{"issue":"22","key":"1185_CR13","doi-asserted-by":"publisher","first-page":"11883","DOI":"10.1007\/s00500-018-03744-z","volume":"23","author":"DC Li","year":"2019","unstructured":"Li, D.C., Lin, L.S., Chen, C.C., Yu, W.H.: Using virtual samples to improve learning performance for small datasets with multimodal distributions. Soft. Comput. 23(22), 11883\u201311900 (2019). https:\/\/doi.org\/10.1007\/s00500-018-03744-z","journal-title":"Soft. Comput."},{"issue":"2","key":"1185_CR14","doi-asserted-by":"publisher","first-page":"539","DOI":"10.1109\/jas.2017.7510841","volume":"5","author":"YL Tian","year":"2018","unstructured":"Tian, Y.L., Li, X., Wang, K.F., Wang, F.Y.: Training and testing object detectors with virtual images. IEEE-CAA J. Autom. Sin. 5(2), 539\u2013546 (2018). https:\/\/doi.org\/10.1109\/jas.2017.7510841","journal-title":"IEEE-CAA J. Autom. Sin."},{"issue":"16","key":"1185_CR15","doi-asserted-by":"publisher","first-page":"13485","DOI":"10.1007\/s00521-022-07179-4","volume":"34","author":"ZF Xue","year":"2022","unstructured":"Xue, Z.F., Chen, L., Liu, Z.T., Liu, Y., Mao, W.J.: Virfd: a virtual-realistic fused dataset for rock size analysis in tbm construction. Neural Comput. Appl. 34(16), 13485\u201313498 (2022). https:\/\/doi.org\/10.1007\/s00521-022-07179-4","journal-title":"Neural Comput. Appl."},{"issue":"1","key":"1185_CR16","doi-asserted-by":"publisher","first-page":"101","DOI":"10.1631\/fitee.2000318","volume":"23","author":"YZ Zhou","year":"2022","unstructured":"Zhou, Y.Z., Feng, T., Shuai, S.H., Li, X.D., Sun, L.Y., Duh, H.B.L.: Edvam: a 3d eye-tracking dataset for visual attention modeling in a virtual museum. Front. Inf. Technol. Electron. Eng. 23(1), 101\u2013112 (2022). https:\/\/doi.org\/10.1631\/fitee.2000318","journal-title":"Front. Inf. Technol. Electron. Eng."},{"issue":"6","key":"1185_CR17","doi-asserted-by":"publisher","first-page":"2072","DOI":"10.1109\/tits.2018.2857566","volume":"20","author":"X Li","year":"2019","unstructured":"Li, X., Wang, K.F., Tian, Y.L., Yan, L., Deng, F., Wang, F.Y.: The paralleleye dataset: a large collection of virtual images for traffic vision research. IEEE Trans. Intell. Transp. Syst. 20(6), 2072\u20132084 (2019). https:\/\/doi.org\/10.1109\/tits.2018.2857566","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"issue":"4","key":"1185_CR18","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1371\/journal.pone.0231266","volume":"15","author":"MM Oliver","year":"2020","unstructured":"Oliver, M.M., Alcover, E.A.: Uibvfed: virtual facial expression dataset. PLoS ONE 15(4), 1\u201310 (2020). https:\/\/doi.org\/10.1371\/journal.pone.0231266","journal-title":"PLoS ONE"},{"key":"1185_CR19","doi-asserted-by":"publisher","unstructured":"Montulet, R., Briassouli, A.: Densely annotated photorealistic virtual dataset generation for abnormal event detection. In: Proceedings of the International Conference on Pattern Recognition, pp. 5\u201319. Springer, Berlin, Heidelberg (2021). https:\/\/doi.org\/10.1007\/978-3-030-68799-1_1","DOI":"10.1007\/978-3-030-68799-1_1"},{"key":"1185_CR20","doi-asserted-by":"publisher","unstructured":"Jeon, H.-G., Im, S., Lee, B.-U., Choi, D.-G., Hebert, M., Kweon, I.S.: Disc: a large-scale virtual dataset for simulating disaster scenarios. In: 2019 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 187\u2013194 (2019). https:\/\/doi.org\/10.1109\/IROS40897.2019.8967839","DOI":"10.1109\/IROS40897.2019.8967839"},{"key":"1185_CR21","doi-asserted-by":"publisher","unstructured":"Shen, Q.Y., Huang, T.G., Ding, P.X., He, J.: Training real-time panoramic object detectors with virtual dataset. In: ICASSP 2021\u20142021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1520\u20131524 (2021). https:\/\/doi.org\/10.1109\/ICASSP39728.2021.9414503","DOI":"10.1109\/ICASSP39728.2021.9414503"},{"key":"1185_CR22","doi-asserted-by":"publisher","unstructured":"Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Doll\u00e1r, P., Zitnick, C.L.: Microsoft coco: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) Computer Vision-ECCV 2014, pp. 740\u2013755 (2014). Springer. https:\/\/doi.org\/10.1007\/978-3-319-10602-1_48","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"1185_CR23","doi-asserted-by":"publisher","unstructured":"Xu, Y., Wang, Z., Li, Z., Yuan, Y., Yu, G.: Siamfc++: Towards robust and accurate visual tracking with target estimation guidelines. In: Proceedings of the AAAI Conference on Artificial Intelligence 34, pp. 12549\u201312556 (2020) https:\/\/doi.org\/10.1609\/aaai.v34i07.6944","DOI":"10.1609\/aaai.v34i07.6944"},{"key":"1185_CR24","doi-asserted-by":"publisher","unstructured":"Li, B., Wu, W., Wang, Q., Zhang, F., Xing, J., Yan, J.: Siamrpn++: Evolution of siamese visual tracking with very deep networks. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4277\u20134286 (2019). https:\/\/doi.org\/10.1109\/CVPR.2019.00441","DOI":"10.1109\/CVPR.2019.00441"},{"key":"1185_CR25","doi-asserted-by":"publisher","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S.: An image is worth 16x16 words: transformers for image recognition at scale. In: International Conference on Learning Representations, pp. 1\u201322 (2010). https:\/\/doi.org\/10.48550\/arXiv.2010.11929","DOI":"10.48550\/arXiv.2010.11929"},{"key":"1185_CR26","doi-asserted-by":"publisher","unstructured":"He, K., Chen, X., Xie, S., Li, Y., Doll\u00e1r, P., Girshick, R.: Masked Autoencoders Are Scalable Vision Learners (2021). https:\/\/doi.org\/10.48550\/arXiv.2111.06377","DOI":"10.48550\/arXiv.2111.06377"},{"key":"1185_CR27","doi-asserted-by":"publisher","first-page":"624","DOI":"10.1016\/j.jmapro.2022.06.046","volume":"81","author":"F Kaji","year":"2022","unstructured":"Kaji, F., Nguyen-Huu, H., Budhwani, A., Narayanan, J.A., Zimny, M., Toyserkani, E.: A deep-learning-based in-situ surface anomaly detection methodology for laser directed energy deposition via powder feeding. J. Manuf. Process. 81, 624\u2013637 (2022). https:\/\/doi.org\/10.1016\/j.jmapro.2022.06.046","journal-title":"J. Manuf. Process."},{"key":"1185_CR28","doi-asserted-by":"publisher","unstructured":"Bertinetto, L., Valmadre, J., Henriques, J.F., Vedaldi, A., Torr, P.H.S.: Fully-convolutional siamese networks for object tracking. In: Hua, G., J\u00e9gou, H. (eds.) Computer Vision\u2014ECCV 2016 Workshops, pp. 850\u2013865. Springer (2016). https:\/\/doi.org\/10.1007\/978-3-319-48881-3_56","DOI":"10.1007\/978-3-319-48881-3_56"},{"key":"1185_CR29","doi-asserted-by":"publisher","unstructured":"Bo, L., Yan, J., Wei, W., Zheng, Z., Hu, X.: High performance visual tracking with siamese region proposal network. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8971\u20138980 (2018). https:\/\/doi.org\/10.1109\/CVPR.2018.00935","DOI":"10.1109\/CVPR.2018.00935"},{"key":"1185_CR30","doi-asserted-by":"crossref","unstructured":"Liu, Z., Mao, H., Wu, C.-Y., Feichtenhofer, C., Darrell, T., Xie, S.: A convnet for the 2020s. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 11976\u201311986 (2022)","DOI":"10.1109\/CVPR52688.2022.01167"},{"key":"1185_CR31","doi-asserted-by":"publisher","unstructured":"Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., Lin, S., Guo, B.: Swin transformer: Hierarchical vision transformer using shifted windows. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 9992\u201310002 (2021). https:\/\/doi.org\/10.1109\/ICCV48922.2021.00986","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"1185_CR32","doi-asserted-by":"publisher","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770\u2013778 (2016). https:\/\/doi.org\/10.1109\/CVPR.2016.90","DOI":"10.1109\/CVPR.2016.90"},{"key":"1185_CR33","doi-asserted-by":"publisher","unstructured":"Loshchilov, I., Hutter, F.: Decoupled Weight Decay Regularization (2017). https:\/\/doi.org\/10.48550\/arXiv.1711.05101","DOI":"10.48550\/arXiv.1711.05101"},{"key":"1185_CR34","doi-asserted-by":"publisher","unstructured":"Huang, G., Liu, Z., Maaten, L.V.D., Weinberger, K.Q.: Densely connected convolutional networks. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2261\u20132269 (2017). https:\/\/doi.org\/10.1109\/CVPR.2017.243","DOI":"10.1109\/CVPR.2017.243"},{"key":"1185_CR35","doi-asserted-by":"publisher","unstructured":"Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z.: Rethinking the inception architecture for computer vision. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2818\u20132826 (2016). https:\/\/doi.org\/10.1109\/CVPR.2016.308","DOI":"10.1109\/CVPR.2016.308"},{"key":"1185_CR36","doi-asserted-by":"publisher","unstructured":"Howard, A.G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., Andreetto, M., Adam, H.: MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications (2017). https:\/\/doi.org\/10.48550\/arXiv.1704.04861","DOI":"10.48550\/arXiv.1704.04861"},{"issue":"6","key":"1185_CR37","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1145\/3065386","volume":"60","author":"A Krizhevsky","year":"2017","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. Commun. ACM 60(6), 84\u201390 (2017). https:\/\/doi.org\/10.1145\/3065386","journal-title":"Commun. ACM"},{"key":"1185_CR38","doi-asserted-by":"publisher","unstructured":"Radosavovic, I., Johnson, J., Xie, S., Lo, W.Y., Dollar, P.: On network design spaces for visual recognition. In: 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 1882\u20131890 (2019). https:\/\/doi.org\/10.1109\/ICCV.2019.00197","DOI":"10.1109\/ICCV.2019.00197"},{"key":"1185_CR39","doi-asserted-by":"publisher","unstructured":"Radosavovic, I., Kosaraju, R.P., Girshick, R., He, K., Doll\u00e1r, P.: Designing network design spaces. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10425\u201310433 (2020). https:\/\/doi.org\/10.1109\/CVPR42600.2020.01044","DOI":"10.1109\/CVPR42600.2020.01044"},{"key":"1185_CR40","doi-asserted-by":"publisher","unstructured":"Tian, Z., Shen, C., Chen, H., He, T.: Fcos: fully convolutional one-stage object detection. In: 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 9626\u20139635 (2019). https:\/\/doi.org\/10.1109\/ICCV.2019.00972","DOI":"10.1109\/ICCV.2019.00972"},{"key":"1185_CR41","doi-asserted-by":"publisher","unstructured":"Zheng, Z., Wang, P., Liu, W., Li, J., Ye, R., Ren, D.: Distance-iou loss: faster and better learning for bounding box regression. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. (07), pp. 12993\u201313000 (2020). https:\/\/doi.org\/10.1609\/aaai.v34i07.6999","DOI":"10.1609\/aaai.v34i07.6999"},{"key":"1185_CR42","doi-asserted-by":"publisher","unstructured":"Redmon, J., Farhadi, A.: YOLOv3: An Incremental Improvement (2018). https:\/\/doi.org\/10.48550\/arXiv.1804.02767","DOI":"10.48550\/arXiv.1804.02767"},{"issue":"11","key":"1185_CR43","doi-asserted-by":"publisher","first-page":"2137","DOI":"10.1109\/TPAMI.2016.2516982","volume":"38","author":"M Kristan","year":"2016","unstructured":"Kristan, M., Matas, J., Leonardis, A., Voj\u00ed\u0159, T., Pflugfelder, R., Fern\u00e1ndez, G., Nebehay, G., Porikli, F., \u010cehovin, L.: A novel performance evaluation methodology for single-target trackers. IEEE Trans. Pattern Anal. Mach. Intell. 38(11), 2137\u20132155 (2016). https:\/\/doi.org\/10.1109\/TPAMI.2016.2516982","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-023-01185-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-023-01185-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-023-01185-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,11,16]],"date-time":"2023-11-16T11:19:15Z","timestamp":1700133555000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-023-01185-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,11]]},"references-count":43,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2023,12]]}},"alternative-id":["1185"],"URL":"https:\/\/doi.org\/10.1007\/s00530-023-01185-9","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,10,11]]},"assertion":[{"value":"4 April 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 September 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 October 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}