{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,30]],"date-time":"2025-09-30T10:35:10Z","timestamp":1759228510068,"version":"3.38.0"},"reference-count":58,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2024,2,20]],"date-time":"2024-02-20T00:00:00Z","timestamp":1708387200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,2,20]],"date-time":"2024-02-20T00:00:00Z","timestamp":1708387200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Intell Manuf"],"published-print":{"date-parts":[[2025,3]]},"DOI":"10.1007\/s10845-023-02318-7","type":"journal-article","created":{"date-parts":[[2024,2,20]],"date-time":"2024-02-20T12:02:32Z","timestamp":1708430552000},"page":"1603-1621","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Enhancing spatiotemporal predictive learning: an approach with nested attention module"],"prefix":"10.1007","volume":"36","author":[{"given":"Shaoping","family":"Wang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1728-7891","authenticated-orcid":false,"given":"Ren","family":"Han","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,2,20]]},"reference":[{"key":"2318_CR1","doi-asserted-by":"publisher","first-page":"1615","DOI":"10.1007\/s00382-013-1845-2","volume":"41","author":"AG Barnston","year":"2013","unstructured":"Barnston, A. G., & Tippett, M. K. (2013). Predictions of Nino3. 4 SST in CFSv1 and CFSv2: A diagnostic comparison. Climate Dynamics, 41, 1615\u20131633.","journal-title":"Climate Dynamics"},{"key":"2318_CR2","doi-asserted-by":"publisher","first-page":"266","DOI":"10.1016\/j.renene.2023.02.130","volume":"207","author":"C Brester","year":"2023","unstructured":"Brester, C., Kallio-Myers, V., Lindfors, A. V., Kolehmainen, M., & Niska, H. (2023). Evaluating neural network models in site-specific solar PV forecasting using numerical weather prediction data and weather observations. Renewable Energy, 207, 266\u2013274.","journal-title":"Renewable Energy"},{"key":"2318_CR3","doi-asserted-by":"crossref","unstructured":"Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., & Zagoruyko, S. (2020). End-to-end object detection with transformers. In European conference on computer vision (pp. 213\u2013229). Springer.","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"2318_CR4","first-page":"26950","volume":"34","author":"Z Chang","year":"2021","unstructured":"Chang, Z., Zhang, X., Wang, S., Ma, S., Ye, Y., Xinguang, X., & Gao, W. (2021). Mau: A motion-aware unit for video prediction and beyond. Advances in Neural Information Processing Systems, 34, 26950\u201326962.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"2318_CR5","doi-asserted-by":"crossref","unstructured":"Chen, M., Peng, H., Fu, J., & Ling, H. (2021). Autoformer: Searching transformers for visual recognition. In Proceedings of the IEEE\/CVF international conference on computer vision (pp. 12270\u201312280).","DOI":"10.1109\/ICCV48922.2021.01205"},{"key":"2318_CR6","doi-asserted-by":"crossref","unstructured":"Cheng, K., Zhang, Y., He, X., Chen, W., Cheng, J., & Lu, H. (2020). Skeleton-based action recognition with shift graph convolutional network. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 183\u2013192).","DOI":"10.1109\/CVPR42600.2020.00026"},{"key":"2318_CR7","unstructured":"Child, R. (2020). Very deep vaes generalize autoregressive models and can outperform them on images. arXiv preprint arXiv:2011.10650"},{"issue":"3","key":"2318_CR8","doi-asserted-by":"publisher","first-page":"4202","DOI":"10.1109\/LRA.2020.2992184","volume":"5","author":"H Chiu","year":"2020","unstructured":"Chiu, H., Adeli, E., & Niebles, J. C. (2020). Segmenting the future. IEEE Robotics and Automation Letters, 5(3), 4202\u20134209.","journal-title":"IEEE Robotics and Automation Letters"},{"key":"2318_CR9","first-page":"3965","volume":"34","author":"Z Dai","year":"2021","unstructured":"Dai, Z., Liu, H., Le, Q. V., & Tan, M. (2021). Coatnet: Marrying convolution and attention for all data sizes. Advances in Neural Information Processing Systems, 34, 3965\u20133977.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"2318_CR10","doi-asserted-by":"crossref","unstructured":"Ding, X., Zhang, X., Han, J., & Ding, G. (2022). Scaling up your kernels to 31x31: Revisiting large kernel design in cnns. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 11963\u201311975).","DOI":"10.1109\/CVPR52688.2022.01166"},{"key":"2318_CR11","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., et al. (2020). An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929"},{"issue":"3","key":"2318_CR12","doi-asserted-by":"publisher","first-page":"1347","DOI":"10.1109\/TIP.2017.2778563","volume":"27","author":"W Du","year":"2017","unstructured":"Du, W., Wang, Y., & Qiao, Y. (2017). Recurrent spatial-temporal attention network for action recognition in videos. IEEE Transactions on Image Processing, 27(3), 1347\u20131360.","journal-title":"IEEE Transactions on Image Processing"},{"key":"2318_CR13","doi-asserted-by":"crossref","unstructured":"Feichtenhofer, C., Fan, H., Xiong, B., Girshick, R., & He, K. (2021). A large-scale study on unsupervised spatiotemporal representation learning. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (pp. 3299\u20133309).","DOI":"10.1109\/CVPR46437.2021.00331"},{"key":"2318_CR14","doi-asserted-by":"crossref","unstructured":"Fu, J., Liu, J., Tian, H., Li, Y., Bao, Y., Fang, Z., & Lu, H. (2019). Dual attention network for scene segmentation. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 3146\u20133154).","DOI":"10.1109\/CVPR.2019.00326"},{"key":"2318_CR15","doi-asserted-by":"crossref","unstructured":"Gao, H., Xu, H., Cai, Q.-Z., Wang, R., Yu, F., & Darrell, T. (2019). Disentangling propagation and generation for video prediction. In Proceedings of the IEEE\/CVF international conference on computer vision (pp. 9006\u20139015).","DOI":"10.1109\/ICCV.2019.00910"},{"key":"2318_CR16","doi-asserted-by":"crossref","unstructured":"Gao, Z., Tan, C., Wu, L., & Li, S. Z. (2022). Simvp: Simpler yet better video prediction. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 3170\u20133180).","DOI":"10.1109\/CVPR52688.2022.00317"},{"issue":"11","key":"2318_CR17","doi-asserted-by":"publisher","first-page":"1231","DOI":"10.1177\/0278364913491297","volume":"32","author":"A Geiger","year":"2013","unstructured":"Geiger, A., Lenz, P., Stiller, C., & Urtasun, R. (2013). Vision meets robotics: The kitti dataset. The International Journal of Robotics Research, 32(11), 1231\u20131237.","journal-title":"The International Journal of Robotics Research"},{"key":"2318_CR18","unstructured":"Guen, V. L., & Thome, N. (2020). Disentangling physical dynamics from unknown factors for unsupervised video prediction. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 11474\u201311484)."},{"key":"2318_CR19","doi-asserted-by":"crossref","unstructured":"Guo, M.-H., Lu, C.-Z., Liu, Z.-N., Cheng, M.-M., & Hu, S.-M. (2023). Visual attention network. Computational Visual Media, 1\u201320.","DOI":"10.1007\/s41095-023-0364-2"},{"key":"2318_CR20","doi-asserted-by":"crossref","unstructured":"Hamdi, A., Shaban, K., Erradi, A., Mohamed, A., Rumi, S. K., & Salim, F. D. (2022). Spatiotemporal data mining: a survey on challenges and open problems. Artificial Intelligence Review, 1\u201348.","DOI":"10.1007\/s10462-021-09994-y"},{"issue":"8","key":"2318_CR21","doi-asserted-by":"publisher","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","volume":"9","author":"S Hochreiter","year":"1997","unstructured":"Hochreiter, S., & Schmidhuber, J. (1997). Long short-term memory. Neural Computation, 9(8), 1735\u20131780.","journal-title":"Neural Computation"},{"key":"2318_CR22","doi-asserted-by":"publisher","unstructured":"Ionescu, C., Li, F., & Sminchisescu, C. (2011). Latent structured models for human pose estimation (pp. 2220\u20132227). Presented at the Proceedings\/IEEE international conference on computer vision. IEEE international conference on computer vision. https:\/\/doi.org\/10.1109\/ICCV.2011.6126500","DOI":"10.1109\/ICCV.2011.6126500"},{"issue":"7","key":"2318_CR23","doi-asserted-by":"publisher","first-page":"1325","DOI":"10.1109\/TPAMI.2013.248","volume":"36","author":"C Ionescu","year":"2013","unstructured":"Ionescu, C., Papava, D., Olaru, V., & Sminchisescu, C. (2013). Human3. 6m: Large scale datasets and predictive methods for 3d human sensing in natural environments. IEEE Transactions on Pattern Analysis and Machine Intelligence, 36(7), 1325\u20131339.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"2318_CR24","doi-asserted-by":"crossref","unstructured":"Jenni, S., Meishvili, G., & Favaro, P. (2020). Video representation learning by recognizing temporal transformations. In European conference on computer vision (pp. 425\u2013442). Springer.","DOI":"10.1007\/978-3-030-58604-1_26"},{"key":"2318_CR25","doi-asserted-by":"crossref","unstructured":"Kong, Z., Dong, P., Ma, X., Meng, X., Niu, W., Sun, M., et al. (2022). Spvit: Enabling faster vision transformers via latency-aware soft token pruning. In European conference on computer vision (pp. 620\u2013640). Springer.","DOI":"10.1007\/978-3-031-20083-0_37"},{"key":"2318_CR26","doi-asserted-by":"crossref","unstructured":"Lee, S., Kim, H. G., Choi, D. H., Kim, H.-I., & Ro, Y. M. (2021). Video prediction recalling long-term motion context via memory alignment learning. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 3054\u20133063).","DOI":"10.1109\/CVPR46437.2021.00307"},{"key":"2318_CR27","unstructured":"Li, K., Wang, Y., Gao, P., Song, G., Liu, Y., Li, H., & Qiao, Y. (2022). Uniformer: Unified transformer for efficient spatiotemporal representation learning. arXiv preprint arXiv:2201.04676"},{"key":"2318_CR28","doi-asserted-by":"crossref","unstructured":"Liang, C., Wang, W., Zhou, T., Miao, J., Luo, Y., & Yang, Y. (2023). Local-global context aware transformer for language-guided video segmentation. IEEE Transactions on Pattern Analysis and Machine Intelligence.","DOI":"10.1109\/TPAMI.2023.3262578"},{"key":"2318_CR30","doi-asserted-by":"crossref","unstructured":"Lin, Y., Sun, H., Liu, N., Bian, Y., Cen, J., & Zhou, H. (2022). A lightweight multi-scale context network for salient object detection in optical remote sensing images. In 2022 26th international conference on pattern recognition (ICPR) (pp. 238\u2013244). IEEE.","DOI":"10.1109\/ICPR56361.2022.9956350"},{"key":"2318_CR29","doi-asserted-by":"crossref","unstructured":"Lin, Z., Li, M., Zheng, Z., Cheng, Y., & Yuan, C. (2020). Self-attention convlstm for spatiotemporal prediction. In Proceedings of the AAAI conference on artificial intelligence (Vol. 34, pp. 11531\u201311538).","DOI":"10.1609\/aaai.v34i07.6819"},{"key":"2318_CR32","doi-asserted-by":"crossref","unstructured":"Liu, Z., Ning, J., Cao, Y., Wei, Y., Zhang, Z., Lin, S., & Hu, H. (2022). Video swin transformer. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 3202\u20133211).","DOI":"10.1109\/CVPR52688.2022.00320"},{"key":"2318_CR31","doi-asserted-by":"crossref","unstructured":"Liu, Z., Yeh, R. A., Tang, X., Liu, Y., & Agarwala, A. (2017). Video frame synthesis using deep voxel flow. In Proceedings of the IEEE international conference on computer vision (pp. 4463\u20134471).","DOI":"10.1109\/ICCV.2017.478"},{"key":"2318_CR33","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2021.107900","volume":"239","author":"C Luo","year":"2022","unstructured":"Luo, C., Zhao, X., Sun, Y., Li, X., & Ye, Y. (2022). Predrann: The spatiotemporal attention convolution recurrent neural network for precipitation nowcasting. Knowledge-Based Systems, 239, 107900.","journal-title":"Knowledge-Based Systems"},{"key":"2318_CR34","doi-asserted-by":"crossref","unstructured":"Muhammad, K., Hussain, T., Ullah, H., Del Ser, J., Rezaei, M., Kumar, N., et al. (2022). Vision-based semantic segmentation in scene understanding for autonomous driving: Recent achievements, challenges, and outlooks. IEEE Transactions on Intelligent Transportation Systems.","DOI":"10.1109\/TITS.2022.3207665"},{"key":"2318_CR35","unstructured":"Patraucean, V., Handa, A., & Cipolla, R. (2015). Spatio-temporal video autoencoder with differentiable memory. arXiv preprint arXiv:1511.06309"},{"key":"2318_CR36","doi-asserted-by":"crossref","unstructured":"Qin, H., Gong, R., Liu, X., Shen, M., Wei, Z., Yu, F., & Song, J. (2020). Forward and backward information retention for accurate binary neural networks. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 2250\u20132259).","DOI":"10.1109\/CVPR42600.2020.00232"},{"issue":"13","key":"2318_CR37","doi-asserted-by":"publisher","first-page":"1609","DOI":"10.1175\/1520-0442(2002)015<1609:AIISAS>2.0.CO;2","volume":"15","author":"RW Reynolds","year":"2002","unstructured":"Reynolds, R. W., Rayner, N. A., Smith, T. M., Stokes, D. C., & Wang, W. (2002). An improved in situ and satellite SST analysis for climate. Journal of Climate, 15(13), 1609\u20131625.","journal-title":"Journal of Climate"},{"key":"2318_CR38","unstructured":"Shi, X., Chen, Z., Wang, H., Yeung, D.-Y., Wong, W.-K., & Woo, W. (2015). Convolutional LSTM network: A machine learning approach for precipitation nowcasting. Advances in Neural Information Processing Systems, 28."},{"key":"2318_CR39","unstructured":"Shouno, O. (2020). Photo-realistic video prediction on natural videos of largely changing frames. arXiv preprint arXiv:2003.08635"},{"key":"2318_CR40","unstructured":"Srivastava, N., Mansimov, E., & Salakhudinov, R. (2015). Unsupervised learning of video representations using lstms. In International conference on machine learning (pp. 843\u2013852). PMLR."},{"key":"2318_CR41","unstructured":"Wang, Y., Long, M., Wang, J., Gao, Z., & Yu, P. S. (2017). Predrnn: Recurrent neural networks for predictive learning using spatiotemporal lstms. Advances in Neural Information Processing Systems, 30."},{"key":"2318_CR46","doi-asserted-by":"publisher","first-page":"118","DOI":"10.1016\/j.cviu.2018.04.007","volume":"171","author":"P Wang","year":"2018","unstructured":"Wang, P., Li, W., Ogunbona, P., Wan, J., & Escalera, S. (2018a). RGB-D-based human motion recognition with deep learning: A survey. Computer Vision and Image Understanding, 171, 118\u2013139.","journal-title":"Computer Vision and Image Understanding"},{"key":"2318_CR44","doi-asserted-by":"crossref","unstructured":"Wang, X., Girshick, R., Gupta, A., & He, K. (2018b). Non-local neural networks. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 7794\u20137803).","DOI":"10.1109\/CVPR.2018.00813"},{"key":"2318_CR43","unstructured":"Wang, Y., Gao, Z., Long, M., Wang, J., & Philip, S. Y. (2018c). Predrnn++: Towards a resolution of the deep-in-time dilemma in spatiotemporal predictive learning. In International Conference on Machine Learning (pp. 5123\u20135132). PMLR."},{"key":"2318_CR42","unstructured":"Wang, Y., Jiang, L., Yang, M.-H., Li, L.-J., Long, M., & Fei-Fei, L. (2018d). Eidetic 3D LSTM: A model for video prediction and beyond. In International conference on learning representations."},{"issue":"2","key":"2318_CR47","doi-asserted-by":"publisher","first-page":"2208","DOI":"10.1109\/TPAMI.2022.3165153","volume":"45","author":"Y Wang","year":"2022","unstructured":"Wang, Y., Wu, H., Zhang, J., Gao, Z., Wang, J., Philip, S. Y., & Long, M. (2022). Predrnn: A recurrent neural network for spatiotemporal predictive learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(2), 2208\u20132225.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"2318_CR45","doi-asserted-by":"crossref","unstructured":"Wang, Y., Zhang, J., Zhu, H., Long, M., Wang, J., & Yu, P. S. (2019). Memory in memory: A predictive neural network for learning higher-order non-stationarity from spatiotemporal dynamics. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 9154\u20139162).","DOI":"10.1109\/CVPR.2019.00937"},{"key":"2318_CR48","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.-Y., & Kweon, I. S. (2018). Cbam: Convolutional block attention module. In Proceedings of the European conference on computer vision (ECCV) (pp. 3\u201319).","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"2318_CR49","doi-asserted-by":"crossref","unstructured":"Wu, H., Yao, Z., Wang, J., & Long, M. (2021). MotionRNN: A flexible model for video prediction with spacetime-varying motions. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 15435\u201315444).","DOI":"10.1109\/CVPR46437.2021.01518"},{"key":"2318_CR50","first-page":"10890","volume":"34","author":"L Wu","year":"2021","unstructured":"Wu, L., Li, J., Wang, Y., Meng, Q., Qin, T., Chen, W., et al. (2021b). R-drop: Regularized dropout for neural networks. Advances in Neural Information Processing Systems, 34, 10890\u201310905.","journal-title":"Advances in Neural Information Processing Systems"},{"issue":"11","key":"2318_CR51","doi-asserted-by":"publisher","first-page":"12191","DOI":"10.1007\/s10489-021-03030-w","volume":"52","author":"J Xie","year":"2022","unstructured":"Xie, J., Gu, L., Li, Z., & Lyu, L. (2022). HRANet: Hierarchical region-aware network for crowd counting. Applied Intelligence, 52(11), 12191\u201312205.","journal-title":"Applied Intelligence"},{"key":"2318_CR53","doi-asserted-by":"crossref","unstructured":"Xu, H., Jiang, C., Liang, X., & Li, Z. (2019). Spatial-aware graph relation network for large-scale object detection. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (pp. 9298\u20139307).","DOI":"10.1109\/CVPR.2019.00952"},{"key":"2318_CR52","doi-asserted-by":"crossref","unstructured":"Xu, Z., Wang, Y., Long, M., Wang, J., & KLiss, M. (2018). PredCNN: Predictive learning with cascade convolutions. In IJCAI (pp. 2940\u20132947).","DOI":"10.24963\/ijcai.2018\/408"},{"key":"2318_CR54","unstructured":"Yu, W., Lu, Y., Easterbrook, S., & Fidler, S. (2020). Efficient and information-preserving future frame prediction and beyond."},{"key":"2318_CR56","doi-asserted-by":"crossref","unstructured":"Zhang, X., Chen, C., Meng, Z., Yang, Z., Jiang, H., & Cui, X. (2022a). CoAtGIN: Marrying convolution and attention for graph-based molecule property prediction. In 2022 IEEE international conference on bioinformatics and biomedicine (BIBM) (pp. 374\u2013379). IEEE.","DOI":"10.1109\/BIBM55620.2022.9995324"},{"key":"2318_CR55","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Zhang, H., Wu, G., & Li, J. (2022b). Spatio-temporal self-supervision enhanced transformer networks for action recognition. In 2022 IEEE International Conference on Multimedia and Expo (ICME) (pp. 1\u20136). IEEE.","DOI":"10.1109\/ICME52920.2022.9859741"},{"key":"2318_CR57","unstructured":"Zheng, L., Wang, C., & Kong, L. (2022). Linear complexity randomized self-attention mechanism. In International conference on machine learning (pp. 27011\u201327041). PMLR."},{"key":"2318_CR58","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2023.110275","volume":"263","author":"B Zhou","year":"2023","unstructured":"Zhou, B., Dong, Y., Yang, G., Hou, F., Hu, Z., Xu, S., & Ma, S. (2023). A graph-attention based spatial-temporal learning framework for tourism demand forecasting. Knowledge-Based Systems, 263, 110275.","journal-title":"Knowledge-Based Systems"}],"container-title":["Journal of Intelligent Manufacturing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10845-023-02318-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10845-023-02318-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10845-023-02318-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,25]],"date-time":"2025-02-25T14:16:40Z","timestamp":1740493000000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10845-023-02318-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,2,20]]},"references-count":58,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2025,3]]}},"alternative-id":["2318"],"URL":"https:\/\/doi.org\/10.1007\/s10845-023-02318-7","relation":{},"ISSN":["0956-5515","1572-8145"],"issn-type":[{"type":"print","value":"0956-5515"},{"type":"electronic","value":"1572-8145"}],"subject":[],"published":{"date-parts":[[2024,2,20]]},"assertion":[{"value":"8 June 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"28 December 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 February 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors certify that they have no affiliations with or involvement in any organization or entity with any financial interest or non-financial interest in the subject matter or materials discussed in this manuscript.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}]}}