{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,26]],"date-time":"2026-03-26T09:53:06Z","timestamp":1774518786509,"version":"3.50.1"},"reference-count":58,"publisher":"Springer Science and Business Media LLC","issue":"6-7","license":[{"start":{"date-parts":[[2024,5,18]],"date-time":"2024-05-18T00:00:00Z","timestamp":1715990400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,5,18]],"date-time":"2024-05-18T00:00:00Z","timestamp":1715990400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2024,8]]},"DOI":"10.1007\/s11760-024-03245-7","type":"journal-article","created":{"date-parts":[[2024,5,18]],"date-time":"2024-05-18T18:01:32Z","timestamp":1716055292000},"page":"5433-5448","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":11,"title":["A novel framework using 3D-CNN and BiLSTM model with dynamic learning rate scheduler for visual speech recognition"],"prefix":"10.1007","volume":"18","author":[{"given":"Vishnu","family":"Chandrabanshi","sequence":"first","affiliation":[]},{"given":"S.","family":"Domnic","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,5,18]]},"reference":[{"key":"3245_CR1","doi-asserted-by":"crossref","unstructured":"Wand, M., Schmidhuber, J., Vu, N.T.: Investigations on end-to-end audiovisual fusion, In: 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 3041\u20133045, IEEE, (2018)","DOI":"10.1109\/ICASSP.2018.8461900"},{"issue":"12","key":"3245_CR2","doi-asserted-by":"publisher","first-page":"8717","DOI":"10.1109\/TPAMI.2018.2889052","volume":"44","author":"T Afouras","year":"2018","unstructured":"Afouras, T., Chung, J.S., Senior, A., Vinyals, O., Zisserman, A.: Deep audio-visual speech recognition. IEEE Trans. Pattern Anal. Mach. Intell. 44(12), 8717\u20138727 (2018)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"3245_CR3","unstructured":"Assael,Y.M., Shillingford, B., Whiteson, S., De Freitas, N.: Lipnet: End-to-end sentence-level lipreading, arXiv preprint arXiv:1611.01599, (2016)"},{"key":"3245_CR4","doi-asserted-by":"crossref","unstructured":"Xu, K., Li, D., Cassimatis, N., Wang, X.: Lcanet: End-to-end lipreading with cascaded attention-ctc, in 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), 548\u2013555, IEEE, (2018)","DOI":"10.1109\/FG.2018.00088"},{"key":"3245_CR5","doi-asserted-by":"crossref","unstructured":"Yousaf, K., Mehmood, Z., Saba, T. , Rehman, A., Rashid, M., Altaf, M., Shuguang, Z.: A novel technique for speech recognition and visualization based mobile application to support two-way communication between deaf-mute and normal peoples. Wirel. Commun. Mob. Comput., 2018, (2018)","DOI":"10.1155\/2018\/1013234"},{"key":"3245_CR6","unstructured":"Thanda, A., Venkatesan, S.M.: Multi-task learning of deep neural networks for audio visual automatic speech recognition, arXiv preprint arXiv:1701.02477, (2017)"},{"key":"3245_CR7","doi-asserted-by":"crossref","unstructured":"Kumar, L.A., Renuka, D.K., Rose, S.L., Shunmugapriya, M.: Attention based multi modal learning for audio visual speech recognition. In: 2022 4th International Conference on Artificial Intelligence and Speech Technology (AIST), 1\u20134, IEEE, (2022)","DOI":"10.1109\/AIST55798.2022.10065019"},{"issue":"6","key":"3245_CR8","doi-asserted-by":"publisher","first-page":"3092","DOI":"10.1109\/TIP.2012.2186310","volume":"21","author":"Y-F Liu","year":"2012","unstructured":"Liu, Y.-F., Lin, C.-Y., Guo, J.-M.: Impact of the lips for biometrics. IEEE Trans. Image Process. 21(6), 3092\u20133101 (2012)","journal-title":"IEEE Trans. Image Process."},{"key":"3245_CR9","doi-asserted-by":"crossref","unstructured":"Liu, M., Wang, L., Lee, K.A., Zhang, H., Zeng, C., Dang, J.: Deeplip: A benchmark for deep learning-based audio-visual lip biometrics. In: 2021 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU), 122\u2013129, IEEE, (2021)","DOI":"10.1109\/ASRU51503.2021.9688240"},{"key":"3245_CR10","doi-asserted-by":"crossref","unstructured":"Sankar, S., Beautemps, D., Hueber, T.: Multistream neural architectures for cued speech recognition using a pre-trained visual feature extractor and constrained ctc decoding. In ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 8477\u20138481, IEEE, (2022)","DOI":"10.1109\/ICASSP43922.2022.9746976"},{"key":"3245_CR11","doi-asserted-by":"crossref","unstructured":"Wang, X., Han, Z., Wang, J., Guo, M.: Speech recognition system based on visual feature for the hearing impaired. In 2008 Fourth International Conference on Natural Computation, 2, 543\u2013546, IEEE, (2008)","DOI":"10.1109\/ICNC.2008.550"},{"key":"3245_CR12","unstructured":"Hassanat, A.B.: Visual passwords using automatic lip reading, arXiv preprint arXiv:1409.0924, (2014)"},{"key":"3245_CR13","doi-asserted-by":"publisher","first-page":"1788","DOI":"10.1109\/TASLP.2020.3000593","volume":"28","author":"M Sadeghi","year":"2020","unstructured":"Sadeghi, M., Leglaive, S., Alameda-Pineda, X., Girin, L., Horaud, R.: Audio-visual speech enhancement using conditional variational auto-encoders. IEEE\/ACM Trans. Audio, Speech, Lang. Process. 28, 1788\u20131800 (2020)","journal-title":"IEEE\/ACM Trans. Audio, Speech, Lang. Process."},{"key":"3245_CR14","unstructured":"Viola, P., Jones, M.: Rapid object detection using a boosted cascade of simple features. In Proceedings of the 2001 IEEE computer society conference on computer vision and pattern recognition. CVPR 2001, 1, I\u2013I, Ieee, (2001)"},{"issue":"11","key":"3245_CR15","doi-asserted-by":"publisher","first-page":"2673","DOI":"10.1109\/78.650093","volume":"45","author":"M Schuster","year":"1997","unstructured":"Schuster, M., Paliwal, K.K.: Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 45(11), 2673\u20132681 (1997)","journal-title":"IEEE Trans. Signal Process."},{"key":"3245_CR16","doi-asserted-by":"crossref","unstructured":"Graves, A., Fern\u00e1ndez, S., Gomez, F., Schmidhuber, J.: Connectionist temporal classification: labelling unsegmented sequence data with recurrent neural networks. In: Proceedings of the 23rd international conference on Machine learning, 369\u2013376, (2006)","DOI":"10.1145\/1143844.1143891"},{"key":"3245_CR17","doi-asserted-by":"crossref","unstructured":"Cho, K., Van Merri\u00ebnboer, B., Gulcehre, C., Bahdanau, D., Bougares, F., Schwenk, H., Bengio, Y.: Learning phrase representations using rnn encoder-decoder for statistical machine translation, arXiv preprint arXiv:1406.1078, (2014)","DOI":"10.3115\/v1\/D14-1179"},{"issue":"8","key":"3245_CR18","doi-asserted-by":"publisher","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","volume":"9","author":"S Hochreiter","year":"1997","unstructured":"Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735\u20131780 (1997)","journal-title":"Neural Comput."},{"key":"3245_CR19","doi-asserted-by":"crossref","unstructured":"Chung, J.S., Senior, A., Vinyals, O., Zisserman, A.: Lip reading sentences in the wild. In 2017 IEEE conference on computer vision and pattern recognition (CVPR), 3444\u20133453, IEEE, (2017)","DOI":"10.1109\/CVPR.2017.367"},{"key":"3245_CR20","doi-asserted-by":"crossref","unstructured":"Yang, S., Zhang, Y., Feng, D., Yang, M., Wang, C., Xiao, J., Long, K., Shan, S., Chen, X.: Lrw-1000: A naturally-distributed large-scale benchmark for lip reading in the wild. In: 2019 14th IEEE international conference on automatic face & gesture recognition (FG 2019), 1\u20138, IEEE, (2019)","DOI":"10.1109\/FG.2019.8756582"},{"issue":"2","key":"3245_CR21","doi-asserted-by":"publisher","first-page":"198","DOI":"10.1109\/34.982900","volume":"24","author":"I Matthews","year":"2002","unstructured":"Matthews, I., Cootes, T.F., Bangham, J.A., Cox, S., Harvey, R.: Extraction of visual features for lipreading. IEEE Trans. Pattern Anal. Mach. Intell. 24(2), 198\u2013213 (2002)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"3245_CR22","doi-asserted-by":"crossref","unstructured":"Liu, H., Zhang, X., Wu, P.: Regression based landmark estimation and multi-feature fusion for visual speech recognition. In: 2015 IEEE International Conference on Image Processing (ICIP), 808\u2013812, IEEE, (2015)","DOI":"10.1109\/ICIP.2015.7350911"},{"key":"3245_CR23","unstructured":"Afouras, T., Chung, J.S., Zisserman, A.: Lrs3-ted: a large-scale dataset for visual speech recognition, arXiv preprint arXiv:1809.00496, (2018)"},{"key":"3245_CR24","doi-asserted-by":"crossref","unstructured":"Rekik, A., Ben-Hamadou, A., Mahdi, W.: A new visual speech recognition approach for rgb-d cameras. In: Image Analysis and Recognition: 11th International Conference, ICIAR 2014, Vilamoura, Portugal, October 22-24, 2014, Proceedings, Part II 11, 21\u201328, Springer, (2014)","DOI":"10.1007\/978-3-319-11755-3_3"},{"key":"3245_CR25","doi-asserted-by":"crossref","unstructured":"Nemani, P., Krishna, G.S., Ramisetty, N., Sai, B.D.S., Kumar, S.: Deep learning based holistic speaker independent visual speech recognition, IEEE Transactions on Artificial Intelligence, (2022)","DOI":"10.1109\/TAI.2022.3220190"},{"key":"3245_CR26","doi-asserted-by":"crossref","unstructured":"Petridis, S., Shen, J., Cetin, D., Pantic, M.: Visual-only recognition of normal, whispered and silent speech. In: 2018 ieee international conference on acoustics, speech and signal processing (icassp), 6219\u20136223, IEEE, (2018)","DOI":"10.1109\/ICASSP.2018.8461596"},{"key":"3245_CR27","doi-asserted-by":"crossref","unstructured":"Anina, I., Zhou, Z., Zhao, G., Pietik\u00e4inen, M.: Ouluvs2: A multi-view audiovisual database for non-rigid mouth motion analysis. In 2015 11th IEEE international conference and workshops on automatic face and gesture recognition (FG), 1, 1\u20135, IEEE, (2015)","DOI":"10.1109\/FG.2015.7163155"},{"issue":"5","key":"3245_CR28","doi-asserted-by":"publisher","first-page":"2421","DOI":"10.1121\/1.2229005","volume":"120","author":"M Cooke","year":"2006","unstructured":"Cooke, M., Barker, J., Cunningham, S., Shao, X.: An audio-visual corpus for speech perception and automatic speech recognition. The J. Acoust. Soc. Am. 120(5), 2421\u20132424 (2006)","journal-title":"The J. Acoust. Soc. Am."},{"key":"3245_CR29","doi-asserted-by":"crossref","unstructured":"Petridis, S., Pantic, M.: Deep complementary bottleneck features for visual speech recognition. In: 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2304\u20132308, IEEE, (2016)","DOI":"10.1109\/ICASSP.2016.7472088"},{"key":"3245_CR30","unstructured":"Garg, A., Noyola, J., Bagadia, S.: Lip reading using cnn and lstm, Technical report, Stanford University, CS231 n project report, (2016)"},{"key":"3245_CR31","unstructured":"Wang, C.: Multi-grained spatio-temporal modeling for lip-reading, arXiv preprint arXiv:1908.11618, (2019)"},{"key":"3245_CR32","doi-asserted-by":"crossref","unstructured":"Stafylakis, T., Tzimiropoulos, G.: Deep word embeddings for visual speech recognition. In: 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 4974\u20134978, IEEE, (2018)","DOI":"10.1109\/ICASSP.2018.8461347"},{"key":"3245_CR33","doi-asserted-by":"crossref","unstructured":"Liu, J., Ren, Y., Zhao, Z., Zhang, C., Huai, B., Yuan, J.: Fastlr: Non-autoregressive lipreading model with integrate-and-fire. In: Proceedings of the 28th ACM International Conference on Multimedia, 4328\u20134336, (2020)","DOI":"10.1145\/3394171.3413740"},{"key":"3245_CR34","doi-asserted-by":"crossref","unstructured":"Debnath, S., Roy, P., Namasudra, S., Crespo, R.G.: Audio-visual automatic speech recognition towards education for disabilities. J. Autism Dev. Disord., 1\u201314, (2022)","DOI":"10.1007\/s10803-022-05654-4"},{"key":"3245_CR35","doi-asserted-by":"publisher","first-page":"1432","DOI":"10.1016\/j.procs.2022.01.181","volume":"199","author":"H Huang","year":"2022","unstructured":"Huang, H., Song, C., Ting, J., Tian, T., Hong, C., Di, Z., Gao, D.: A novel machine lip reading model. Proced. Comput. Sci. 199, 1432\u20131437 (2022)","journal-title":"Proced. Comput. Sci."},{"key":"3245_CR36","doi-asserted-by":"crossref","unstructured":"Ma, P., Martinez, ., Petridis, S., Pantic, M.: Towards practical lipreading with distilled and efficient models. In: ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 7608\u20137612, IEEE, (2021)","DOI":"10.1109\/ICASSP39728.2021.9415063"},{"issue":"1","key":"3245_CR37","doi-asserted-by":"publisher","first-page":"113","DOI":"10.1049\/ipr2.12337","volume":"16","author":"L He","year":"2022","unstructured":"He, L., Ding, B., Wang, H., Zhang, T.: An optimal 3d convolutional neural network based lipreading method. IET Image Process. 16(1), 113\u2013122 (2022)","journal-title":"IET Image Process."},{"key":"3245_CR38","doi-asserted-by":"crossref","unstructured":"Rahmani, M.H., Almasganj, F.: Lip-reading via a dnn-hmm hybrid system using combination of the image-based and model-based features. In: 2017 3rd International Conference on Pattern Recognition and Image Analysis (IPRIA), 195\u2013199, IEEE, (2017)","DOI":"10.1109\/PRIA.2017.7983045"},{"key":"3245_CR39","doi-asserted-by":"crossref","unstructured":"Zhao, X., Yang, S., Shan, S., Chen, X.: Mutual information maximization for effective lip reading. In: 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020), 420\u2013427, IEEE, (2020)","DOI":"10.1109\/FG47880.2020.00133"},{"key":"3245_CR40","doi-asserted-by":"publisher","first-page":"55354","DOI":"10.1109\/ACCESS.2020.2982359","volume":"8","author":"M Ezz","year":"2020","unstructured":"Ezz, M., Mostafa, A.M., Nasr, A.A.: A silent password recognition framework based on lip analysis. IEEE Access 8, 55354\u201355371 (2020)","journal-title":"IEEE Access"},{"key":"3245_CR41","doi-asserted-by":"publisher","first-page":"215516","DOI":"10.1109\/ACCESS.2020.3040906","volume":"8","author":"S Fenghour","year":"2020","unstructured":"Fenghour, S., Chen, D., Guo, K., Xiao, P.: Lip reading sentences using deep learning with only visual cues. IEEE Access 8, 215516\u2013215530 (2020)","journal-title":"IEEE Access"},{"key":"3245_CR42","doi-asserted-by":"crossref","unstructured":"Martinez, B., Ma, P., Petridis, S., Pantic, M.: Lipreading using temporal convolutional networks. In: ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 6319\u20136323, IEEE, (2020)","DOI":"10.1109\/ICASSP40776.2020.9053841"},{"key":"3245_CR43","doi-asserted-by":"crossref","unstructured":"Wand, M., Koutn\u00edk, J., Schmidhuber, J.: Lipreading with long short-term memory. In: 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 6115\u20136119, IEEE, (2016)","DOI":"10.1109\/ICASSP.2016.7472852"},{"key":"3245_CR44","doi-asserted-by":"crossref","unstructured":"Wand, M., Schmidhuber, J.: Improving speaker-independent lipreading with domain-adversarial training, arXiv preprint arXiv:1708.01565, (2017)","DOI":"10.21437\/Interspeech.2017-421"},{"key":"3245_CR45","unstructured":"Rastogi,A. Agarwal,R., Gupta,V., Dhar,J., Bhattacharya,M.: Lrneunet: An attention based deep architecture for lipreading from multitudinous sized videos, in 2019 International Conference on Computing, Power and Communication Technologies (GUCON), 1001\u20131007, IEEE, 2019"},{"key":"3245_CR46","doi-asserted-by":"crossref","unstructured":"Luo, M., Yang, S., Shan, S., Chen, X.: Pseudo-convolutional policy gradient for sequence-to-sequence lip-reading. In: 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020), 273\u2013280, IEEE, (2020)","DOI":"10.1109\/FG47880.2020.00010"},{"issue":"2","key":"3245_CR47","first-page":"1531","volume":"68","author":"AM Sarhan","year":"2021","unstructured":"Sarhan, A.M., Elshennawy, N.M., Ibrahim, D.M.: Hlr-net: a hybrid lip-reading model based on deep convolutional neural networks. Comput. Mater. & Contin. 68(2), 1531\u20131549 (2021)","journal-title":"Comput. Mater. & Contin."},{"key":"3245_CR48","doi-asserted-by":"crossref","unstructured":"Vayadande, K., Adsare, T., Agrawal, N., Dharmik, T., Patil, A., Zod, S.: Lipreadnet: A deep learning approach to lip reading. In: 2023 International Conference on Applied Intelligence and Sustainable Computing (ICAISC), 1\u20136, IEEE, (2023)","DOI":"10.1109\/ICAISC58445.2023.10200426"},{"key":"3245_CR49","doi-asserted-by":"crossref","unstructured":"NadeemHashmi, S., Gupta, H., Mittal, D., Kumar, K., Nanda, A., Gupta, S.: A lip reading model using cnn with batch normalization. In: 2018 eleventh international conference on contemporary computing (IC3), 1\u20136, IEEE, (2018)","DOI":"10.1109\/IC3.2018.8530509"},{"issue":"1s","key":"3245_CR50","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3524620","volume":"19","author":"F Xue","year":"2023","unstructured":"Xue, F., Yang, T., Liu, K., Hong, Z., Cao, M., Guo, D., Hong, R.: Lcsnet: End-to-end lipreading with channel-aware feature selection. ACM Trans. Multimed. Comput., Commun. Appl. 19(1s), 1\u201321 (2023)","journal-title":"ACM Trans. Multimed. Comput., Commun. Appl."},{"key":"3245_CR51","doi-asserted-by":"crossref","unstructured":"Almajai, I., Cox, S., Harvey, R., Lan, Y.: Improved speaker independent lip reading using speaker adaptive training and deep neural networks. In: 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2722\u20132726, IEEE, (2016)","DOI":"10.1109\/ICASSP.2016.7472172"},{"key":"3245_CR52","unstructured":"Cooke, M., Barker, J., Cunningham, S., Shao, X.: The grid audiovisual sentence corpus. https:\/\/spandh.dcs.shef.ac.uk\/gridcorpus\/, (2006)"},{"key":"3245_CR53","unstructured":"Nair, V., Hinton, G.E.: Rectified linear units improve restricted boltzmann machines. In: Proceedings of the 27th international conference on machine learning (ICML-10), 807\u2013814, (2010)"},{"key":"3245_CR54","unstructured":"Agarap, A.F.: Deep learning using rectified linear units (relu), arXiv preprint arXiv:1803.08375, (2018)"},{"key":"3245_CR55","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., Huang, Z., Karpathy, A., Khosla, A., Bernstein, M., et al.: Imagenet large scale visual recognition challenge. Int. J. Comput. Vision 115, 211\u2013252 (2015)","journal-title":"Int. J. Comput. Vision"},{"key":"3245_CR56","unstructured":"Maas, A.L., Hannun, A.Y., Ng, A.Y., et al.: Rectifier nonlinearities improve neural network acoustic models. In: Proc. icml, 30, 3, Atlanta, GA, (2013)"},{"key":"3245_CR57","doi-asserted-by":"crossref","unstructured":"Khalil, K., Dey, B., Kumar, A., Bayoumi, M.: A reversible-logic based architecture for long short-term memory (lstm) network. In: 2021 IEEE International Symposium on Circuits and Systems (ISCAS), 1\u20135, IEEE, (2021)","DOI":"10.1109\/ISCAS51556.2021.9401395"},{"key":"3245_CR58","unstructured":"Hinton, G.E., Srivastava, N., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.R.: Improving neural networks by preventing co-adaptation of feature detectors, arXiv preprint arXiv:1207.0580, (2012)"}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-024-03245-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-024-03245-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-024-03245-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,7,15]],"date-time":"2024-07-15T08:26:53Z","timestamp":1721032013000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-024-03245-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,18]]},"references-count":58,"journal-issue":{"issue":"6-7","published-print":{"date-parts":[[2024,8]]}},"alternative-id":["3245"],"URL":"https:\/\/doi.org\/10.1007\/s11760-024-03245-7","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"value":"1863-1703","type":"print"},{"value":"1863-1711","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,5,18]]},"assertion":[{"value":"7 March 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 April 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 April 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 May 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}]}}