{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,14]],"date-time":"2026-02-14T08:07:56Z","timestamp":1771056476332,"version":"3.50.1"},"reference-count":54,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2026,1,6]],"date-time":"2026-01-06T00:00:00Z","timestamp":1767657600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2026,1,21]],"date-time":"2026-01-21T00:00:00Z","timestamp":1768953600000},"content-version":"vor","delay-in-days":15,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"DOI":"10.13039\/501100007129","name":"Natural Science Foundation of Shandong Province","doi-asserted-by":"publisher","award":["ZR2020MF137"],"award-info":[{"award-number":["ZR2020MF137"]}],"id":[{"id":"10.13039\/501100007129","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100007129","name":"Natural Science Foundation of Shandong Province","doi-asserted-by":"publisher","award":["ZR2020MF137"],"award-info":[{"award-number":["ZR2020MF137"]}],"id":[{"id":"10.13039\/501100007129","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100007129","name":"Natural Science Foundation of Shandong Province","doi-asserted-by":"publisher","award":["ZR2020MF137"],"award-info":[{"award-number":["ZR2020MF137"]}],"id":[{"id":"10.13039\/501100007129","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Process Lett"],"DOI":"10.1007\/s11063-025-11790-6","type":"journal-article","created":{"date-parts":[[2026,1,6]],"date-time":"2026-01-06T15:01:51Z","timestamp":1767711711000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Dynamic-static partitioning mask and multi-dimensional attention mechanism for skeleton-based action recognition"],"prefix":"10.1007","volume":"58","author":[{"given":"Hui","family":"Cao","sequence":"first","affiliation":[]},{"given":"Yuanyuan","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Tingwei","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,6]]},"reference":[{"key":"11790_CR1","doi-asserted-by":"publisher","first-page":"4039","DOI":"10.1109\/TMM.2023.3321438","volume":"26","author":"Y Wang","year":"2023","unstructured":"Wang Y, Kang H, Wu D, Yang W, Zhang L (2023) Global and local spatio-temporal encoder for 3d human pose estimation. IEEE Trans Multimedia 26:4039\u20134049","journal-title":"IEEE Trans Multimedia"},{"key":"11790_CR2","doi-asserted-by":"publisher","first-page":"3507","DOI":"10.1109\/TIP.2023.3286254","volume":"32","author":"Z Tu","year":"2023","unstructured":"Tu Z, Liu Y, Zhang Y, Mu Q, Yuan J (2023) Dtcm: Joint optimization of dark enhancement and action recognition in videos. IEEE Trans Image Process 32:3507\u20133520","journal-title":"IEEE Trans Image Process"},{"key":"11790_CR3","doi-asserted-by":"publisher","first-page":"525","DOI":"10.1109\/TIP.2023.3345177","volume":"33","author":"Y Zhang","year":"2023","unstructured":"Zhang Y, Xu X, Zhao Y, Wen Y, Tang Z, Liu M (2023) Facial prior guided micro-expression generation. IEEE Trans Image Process 33:525\u2013540","journal-title":"IEEE Trans Image Process"},{"key":"11790_CR4","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIP.2023.3334954","volume":"33","author":"X Wang","year":"2023","unstructured":"Wang X, Zhang W, Wang C, Gao Y, Liu M (2023) Dynamic dense graph convolutional network for skeleton-based human motion prediction. IEEE Trans Image Process 33:1\u201315","journal-title":"IEEE Trans Image Process"},{"issue":"8","key":"11790_CR5","doi-asserted-by":"publisher","first-page":"1987","DOI":"10.1109\/TMM.2018.2790163","volume":"20","author":"F-L Zhang","year":"2018","unstructured":"Zhang F-L, Wu X, Li R-L, Wang J, Zheng Z-H, Hu S-M (2018) Detecting and removing visual distractors for video aesthetic enhancement. IEEE Trans Multimedia 20(8):1987\u20131999","journal-title":"IEEE Trans Multimedia"},{"key":"11790_CR6","doi-asserted-by":"crossref","unstructured":"Chen C, Liu M, Meng X, Xiao W, Ju Q (2020) Refinedetlite: A lightweight one-stage object detection framework for cpu-only devices. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 700\u2013701 (2020)","DOI":"10.1109\/CVPRW50498.2020.00358"},{"key":"11790_CR7","doi-asserted-by":"publisher","first-page":"346","DOI":"10.1016\/j.patcog.2017.02.030","volume":"68","author":"M Liu","year":"2017","unstructured":"Liu M, Liu H, Chen C (2017) Enhanced skeleton visualization for view invariant human action recognition. Pattern Recogn 68:346\u2013362","journal-title":"Pattern Recogn"},{"key":"11790_CR8","doi-asserted-by":"crossref","unstructured":"Yang F, Wu Y, Sakti S, Nakamura S (2019) Make skeleton-based action recognition model smaller, faster and better. In: Proceedings of the 1st ACM International Conference on Multimedia in Asia, pp. 1\u20136","DOI":"10.1145\/3338533.3366569"},{"key":"11790_CR9","doi-asserted-by":"crossref","unstructured":"Liu M, Meng F, Chen C, Wu S (2023) Novel motion patterns matter for practical skeleton-based action recognition. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 37, pp. 1701\u20131709","DOI":"10.1609\/aaai.v37i2.25258"},{"key":"11790_CR10","doi-asserted-by":"crossref","unstructured":"Zhou H, Liu Q, Wang Y (2023) Learning discriminative representations for skeleton based action recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10608\u201310617","DOI":"10.1109\/CVPR52729.2023.01022"},{"key":"11790_CR11","doi-asserted-by":"publisher","DOI":"10.1016\/j.compeleceng.2024.109633","volume":"120","author":"A Roy","year":"2024","unstructured":"Roy A, Tiwari A, Saurav S, Singh S (2024) Enhancing skeleton-based action recognition using a knowledge-driven shift graph convolutional network. Comput Electr Eng 120:109633","journal-title":"Comput Electr Eng"},{"key":"11790_CR12","doi-asserted-by":"crossref","unstructured":"Yan S, Xiong Y, Lin D (2018) Spatial temporal graph convolutional networks for skeleton-based action recognition. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32","DOI":"10.1609\/aaai.v32i1.12328"},{"issue":"2","key":"11790_CR13","doi-asserted-by":"publisher","first-page":"1474","DOI":"10.1109\/TPAMI.2022.3157033","volume":"45","author":"Y-F Song","year":"2022","unstructured":"Song Y-F, Zhang Z, Shan C, Wang L (2022) Constructing stronger and faster baselines for skeleton-based action recognition. IEEE Trans Pattern Anal Mach Intell 45(2):1474\u20131488","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"11790_CR14","doi-asserted-by":"crossref","unstructured":"Shi L, Zhang Y, Cheng J, Lu H (2019) Two-stream adaptive graph convolutional networks for skeleton-based action recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12026\u201312035","DOI":"10.1109\/CVPR.2019.01230"},{"key":"11790_CR15","doi-asserted-by":"publisher","first-page":"9532","DOI":"10.1109\/TIP.2020.3028207","volume":"29","author":"L Shi","year":"2020","unstructured":"Shi L, Zhang Y, Cheng J, Lu H (2020) Skeleton-based action recognition with multi-stream adaptive graph convolutional networks. IEEE Trans Image Process 29:9532\u20139545","journal-title":"IEEE Trans Image Process"},{"key":"11790_CR16","doi-asserted-by":"crossref","unstructured":"Chen Y, Zhang Z, Yuan C, Li B, Deng Y, Hu W (2021) Channel-wise topology refinement graph convolution for skeleton-based action recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 13359\u201313368","DOI":"10.1109\/ICCV48922.2021.01311"},{"key":"11790_CR17","doi-asserted-by":"crossref","unstructured":"Chen Z, Li S, Yang B, Li Q, Liu H (2021) Multi-scale spatial temporal graph convolutional network for skeleton-based action recognition. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, pp. 1113\u20131122","DOI":"10.1609\/aaai.v35i2.16197"},{"key":"11790_CR18","doi-asserted-by":"crossref","unstructured":"Liu Z, Zhang H, Chen Z, Wang Z, Ouyang W (2020) Disentangling and unifying graph convolutions for skeleton-based action recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 143\u2013152","DOI":"10.1109\/CVPR42600.2020.00022"},{"key":"11790_CR19","doi-asserted-by":"crossref","unstructured":"Song Y-F, Zhang Z, Shan C, Wang L (2020) Stronger, faster and more explainable: A graph convolutional baseline for skeleton-based action recognition. In: Proceedings of the 28th ACM International Conference on Multimedia, pp. 1625\u20131633","DOI":"10.1145\/3394171.3413802"},{"issue":"7","key":"11790_CR20","doi-asserted-by":"publisher","first-page":"636","DOI":"10.1080\/13682199.2023.2190927","volume":"71","author":"H Zhou","year":"2023","unstructured":"Zhou H, Xiang X, Qiu Y, Liu X (2023) Graph convolutional network with stc attention and adaptive normalization for skeleton-based action recognition. The Imaging Sci J 71(7):636\u2013646","journal-title":"The Imaging Sci J"},{"key":"11790_CR21","doi-asserted-by":"crossref","unstructured":"Lin L, Zhang J, Liu J (2023) Actionlet-dependent contrastive learning for unsupervised skeleton-based action recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2363\u20132372","DOI":"10.1109\/CVPR52729.2023.00234"},{"key":"11790_CR22","doi-asserted-by":"crossref","unstructured":"Lin J, Gan C, Han S (2019) Tsm: Temporal shift module for efficient video understanding","DOI":"10.1109\/ICCV.2019.00718"},{"key":"11790_CR23","doi-asserted-by":"crossref","unstructured":"Feichtenhofer C, Fan H, Malik J, He K (2019) Slowfast networks for video recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6202\u20136211","DOI":"10.1109\/ICCV.2019.00630"},{"key":"11790_CR24","doi-asserted-by":"publisher","first-page":"454","DOI":"10.1007\/s11263-017-0998-6","volume":"123","author":"C Xu","year":"2017","unstructured":"Xu C, Govindarajan LN, Zhang Y, Cheng L (2017) Lie-x: Depth image based articulated object pose estimation, tracking, and action recognition on lie groups. Int J Comput Vision 123:454\u2013478","journal-title":"Int J Comput Vision"},{"key":"11790_CR25","doi-asserted-by":"crossref","unstructured":"Baek S, Shi Z, Kawade M, Kim T-K (2016) Kinematic-layout-aware random forests for depth-based action recognition. arXiv preprint arXiv:1607.06972","DOI":"10.5244\/C.31.13"},{"key":"11790_CR26","doi-asserted-by":"crossref","unstructured":"Feichtenhofer C, Pinz A, Zisserman A (2016) Convolutional two-stream network fusion for video action recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1933\u20131941","DOI":"10.1109\/CVPR.2016.213"},{"key":"11790_CR27","doi-asserted-by":"crossref","unstructured":"Hu J-F, Zheng W-S, Lai J, Zhang J (2015) Jointly learning heterogeneous features for rgb-d activity recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5344\u20135352","DOI":"10.1109\/CVPR.2015.7299172"},{"key":"11790_CR28","doi-asserted-by":"publisher","DOI":"10.1016\/j.jvcir.2021.103055","volume":"76","author":"L Song","year":"2021","unstructured":"Song L, Yu G, Yuan J, Liu Z (2021) Human pose estimation and its application to action recognition: A survey. J Vis Commun Image Represent 76:103055","journal-title":"J Vis Commun Image Represent"},{"issue":"3","key":"11790_CR29","first-page":"3200","volume":"45","author":"Z Sun","year":"2022","unstructured":"Sun Z, Ke Q, Rahmani H, Bennamoun M, Wang G, Liu J (2022) Human action recognition from various data modalities: A review. IEEE Trans Pattern Anal Mach Intell 45(3):3200\u20133225","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"11790_CR30","doi-asserted-by":"crossref","unstructured":"Chi H-g, Ha MH, Chi S, Lee SW, Huang Q, Ramani K (2022) Infogcn: Representation learning for human skeleton-based action recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 20186\u201320196","DOI":"10.1109\/CVPR52688.2022.01955"},{"key":"11790_CR31","doi-asserted-by":"crossref","unstructured":"Ke Q, Bennamoun M, An S, Sohel F, Boussaid F (2017) A new representation of skeleton sequences for 3d action recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3288\u20133297","DOI":"10.1109\/CVPR.2017.486"},{"key":"11790_CR32","doi-asserted-by":"crossref","unstructured":"Du Y, Wang W, Wang L (2015) Hierarchical recurrent neural network for skeleton based action recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1110\u20131118","DOI":"10.1109\/CVPR.2015.7298714"},{"key":"11790_CR33","doi-asserted-by":"crossref","unstructured":"Cheng K, Zhang Y, Cao C, Shi L, Cheng J, Lu H (2020) Decoupling gcn with dropgraph module for skeleton-based action recognition. In: Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XXIV 16, pp. 536\u2013553. Springer","DOI":"10.1007\/978-3-030-58586-0_32"},{"key":"11790_CR34","doi-asserted-by":"crossref","unstructured":"Gao L, Ji Y, Yang Y, Shen H (2022) Global-local cross-view fisher discrimination for view-invariant action recognition. In: Proceedings of the 30th ACM International Conference on Multimedia, pp. 5255\u20135264","DOI":"10.1145\/3503161.3548280"},{"key":"11790_CR35","doi-asserted-by":"crossref","unstructured":"Prasse K, Jung S, Zhou Y, Keuper M (2023) Local spherical harmonics improve skeleton-based hand action recognition. In: DAGM German Conference on Pattern Recognition, pp. 67\u201382. Springer","DOI":"10.1007\/978-3-031-54605-1_5"},{"key":"11790_CR36","doi-asserted-by":"crossref","unstructured":"Zhang P, Lan C, Zeng W, Xing J, Xue J, Zheng N (2020) Semantics-guided neural networks for efficient skeleton-based human action recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1112\u20131121","DOI":"10.1109\/CVPR42600.2020.00119"},{"key":"11790_CR37","unstructured":"Veli\u010dkovi\u0107 P, Cucurull G, Casanova A, Romero A, Lio P, Bengio Y (2017) Graph attention networks. arXiv preprint arXiv:1710.10903"},{"key":"11790_CR38","first-page":"28877","volume":"34","author":"C Ying","year":"2021","unstructured":"Ying C, Cai T, Luo S, Zheng S, Ke G, He D, Shen Y, Liu T-Y (2021) Do transformers really perform badly for graph representation? Adv Neural Inf Process Syst 34:28877\u201328888","journal-title":"Adv Neural Inf Process Syst"},{"key":"11790_CR39","doi-asserted-by":"crossref","unstructured":"Dong X, Shi P, Liang T, Yang A (2024) Ctaffnet: Cnn\u2013transformer adaptive feature fusion object detection algorithm for complex traffic scenarios. Transportation Research Record, 03611981241258753","DOI":"10.1177\/03611981241258753"},{"key":"11790_CR40","doi-asserted-by":"publisher","DOI":"10.1016\/j.displa.2024.102814","volume":"84","author":"X Dong","year":"2024","unstructured":"Dong X, Shi P, Qi H, Yang A, Liang T (2024) Ts-bev: Bev object detection algorithm based on temporal-spatial feature fusion. Displays 84:102814","journal-title":"Displays"},{"key":"11790_CR41","doi-asserted-by":"crossref","unstructured":"Qian R, Ding S, Liu X, Lin D (2022) Static and dynamic concepts for self-supervised video representation learning. In: European Conference on Computer Vision, pp. 145\u2013164. Springer","DOI":"10.1007\/978-3-031-19809-0_9"},{"key":"11790_CR42","unstructured":"Yue F, Huang S, Chen Q, Hu S, Tan Y, Dang S, Du D (2023) A novel two-stream architecture fusing static and dynamic features for human action recognition. IAENG Int J Comput Sci 50(2)"},{"key":"11790_CR43","first-page":"17623","volume":"36","author":"P Wei","year":"2023","unstructured":"Wei P, Kong L, Qu X, Ren Y, Xu Z, Jiang J, Yin X (2023) Unsupervised video domain adaptation for action recognition: A disentanglement perspective. Adv Neural Inf Process Syst 36:17623\u201317642","journal-title":"Adv Neural Inf Process Syst"},{"key":"11790_CR44","doi-asserted-by":"crossref","unstructured":"Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D (2017) Grad-cam: Visual explanations from deep networks via gradient-based localization. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 618\u2013626","DOI":"10.1109\/ICCV.2017.74"},{"key":"11790_CR45","doi-asserted-by":"crossref","unstructured":"Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7132\u20137141","DOI":"10.1109\/CVPR.2018.00745"},{"key":"11790_CR46","doi-asserted-by":"crossref","unstructured":"Shahroudy A, Liu J, Ng T-T, Wang G (2016) Ntu rgb+ d: A large scale dataset for 3d human activity analysis. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1010\u20131019","DOI":"10.1109\/CVPR.2016.115"},{"issue":"10","key":"11790_CR47","doi-asserted-by":"publisher","first-page":"2684","DOI":"10.1109\/TPAMI.2019.2916873","volume":"42","author":"J Liu","year":"2019","unstructured":"Liu J, Shahroudy A, Perez M, Wang G, Duan L-Y, Kot AC (2019) Ntu rgb+ d 120: A large-scale benchmark for 3d human activity understanding. IEEE Trans Pattern Anal Mach Intell 42(10):2684\u20132701","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"11790_CR48","doi-asserted-by":"crossref","unstructured":"Duan H, Wang J, Chen K, Lin D (2022) Pyskl: Towards good practices for skeleton action recognition. In: Proceedings of the 30th ACM International Conference on Multimedia, pp. 7351\u20137354","DOI":"10.1145\/3503161.3548546"},{"key":"11790_CR49","doi-asserted-by":"crossref","unstructured":"Ye F, Pu S, Zhong Q, Li C, Xie D, Tang H (2020) Dynamic gcn: Context-enriched topology learning for skeleton-based action recognition. In: Proceedings of the 28th ACM International Conference on Multimedia, pp. 55\u201363","DOI":"10.1145\/3394171.3413941"},{"key":"11790_CR50","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2022.109231","volume":"136","author":"L Wu","year":"2023","unstructured":"Wu L, Zhang C, Zou Y (2023) Spatiotemporal focus for skeleton-based action recognition. Pattern Recogn 136:109231","journal-title":"Pattern Recogn"},{"key":"11790_CR51","doi-asserted-by":"publisher","first-page":"8527","DOI":"10.1109\/TMM.2023.3318325","volume":"25","author":"H Tian","year":"2023","unstructured":"Tian H, Ma X, Li X, Li Y (2023) Skeleton-based action recognition with select-assemble-normalize graph convolutional networks. IEEE Trans Multimedia 25:8527\u20138538","journal-title":"IEEE Trans Multimedia"},{"key":"11790_CR52","doi-asserted-by":"crossref","unstructured":"Yun X, Xu C, Riou K, Dong K, Sun Y, Li S, Subrin K, Le\u00a0Callet P (2024) Behavioral recognition of skeletal data based on targeted dual fusion strategy. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 38, pp. 6917\u20136925","DOI":"10.1609\/aaai.v38i7.28517"},{"key":"11790_CR53","doi-asserted-by":"publisher","DOI":"10.1016\/j.compeleceng.2024.109633","volume":"120","author":"A Roy","year":"2024","unstructured":"Roy A, Tiwari A, Saurav S, Singh S (2024) Enhancing skeleton-based action recognition using a knowledge-driven shift graph convolutional network. Comput Electr Eng 120:109633","journal-title":"Comput Electr Eng"},{"key":"11790_CR54","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2024.111106","volume":"159","author":"Z Wu","year":"2025","unstructured":"Wu Z, Ding Y, Wan L, Li T, Nian F (2025) Local and global self-attention enhanced graph convolutional network for skeleton-based action recognition. Pattern Recogn 159:111106","journal-title":"Pattern Recogn"}],"container-title":["Neural Processing Letters"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11063-025-11790-6","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11063-025-11790-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11063-025-11790-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,14]],"date-time":"2026-02-14T07:25:09Z","timestamp":1771053909000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11063-025-11790-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1,6]]},"references-count":54,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2026,2]]}},"alternative-id":["11790"],"URL":"https:\/\/doi.org\/10.1007\/s11063-025-11790-6","relation":{},"ISSN":["1573-773X"],"issn-type":[{"value":"1573-773X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,1,6]]},"assertion":[{"value":"10 December 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 July 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 January 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no relevant financial or non-financial interests to disclose.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest\/Competing interests:"}},{"value":"This study does not involve human participants or animals and therefore does not require ethics approval.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval and consent to participate:"}},{"value":"All authors have read and agreed to the final version of the manuscript and consent to its submission and publication.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication:"}},{"value":"The NTU RGB+D dataset and NTU RGB+D 120 dataset are available at\n                      \n                      .","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Materials availability:"}},{"value":"The code will be made publicly available upon acceptance.","order":6,"name":"Ethics","group":{"name":"EthicsHeading","label":"Code availability:"}}],"article-number":"12"}}