{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,19]],"date-time":"2026-03-19T14:50:03Z","timestamp":1773931803797,"version":"3.50.1"},"reference-count":62,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2024,6,27]],"date-time":"2024-06-27T00:00:00Z","timestamp":1719446400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,6,27]],"date-time":"2024-06-27T00:00:00Z","timestamp":1719446400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["No.61063021"],"award-info":[{"award-number":["No.61063021"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2024,8]]},"DOI":"10.1007\/s00530-024-01395-9","type":"journal-article","created":{"date-parts":[[2024,6,27]],"date-time":"2024-06-27T03:40:43Z","timestamp":1719459643000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":8,"title":["PointDMIG: a dynamic motion-informed graph neural network for 3D action recognition"],"prefix":"10.1007","volume":"30","author":[{"given":"Yao","family":"Du","sequence":"first","affiliation":[]},{"given":"Zhenjie","family":"Hou","sequence":"additional","affiliation":[]},{"given":"Xing","family":"Li","sequence":"additional","affiliation":[]},{"given":"Jiuzhen","family":"Liang","sequence":"additional","affiliation":[]},{"given":"Kaijun","family":"You","sequence":"additional","affiliation":[]},{"given":"Xinwen","family":"Zhou","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,6,27]]},"reference":[{"key":"1395_CR1","doi-asserted-by":"publisher","first-page":"4049","DOI":"10.1109\/TMM.2020.3037481","volume":"23","author":"A Javaheri","year":"2020","unstructured":"Javaheri, A., Brites, C., Pereira, F., Ascenso, J.: Point cloud rendering after coding: impacts on subjective and objective quality. IEEE Trans. Multimedia 23, 4049\u20134064 (2020)","journal-title":"IEEE Trans. Multimedia"},{"key":"1395_CR2","doi-asserted-by":"crossref","unstructured":"Tran, D., Bourdev, L., Fergus, R., Torresani, L., Paluri, M.: Learning spatiotemporal features with 3d convolutional networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 4489\u20134497 (2015)","DOI":"10.1109\/ICCV.2015.510"},{"key":"1395_CR3","doi-asserted-by":"crossref","unstructured":"Karpathy, A., Toderici, G., Shetty, S., Leung, T., Sukthankar, R., Fei-Fei, L.: Large-scale video classification with convolutional neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1725\u20131732 (2014)","DOI":"10.1109\/CVPR.2014.223"},{"key":"1395_CR4","doi-asserted-by":"crossref","unstructured":"Carreira, J., Zisserman, A.: Quo vadis, action recognition? A new model and the kinetics dataset. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE (2017)","DOI":"10.1109\/CVPR.2017.502"},{"key":"1395_CR5","doi-asserted-by":"crossref","unstructured":"Choy, C., Gwak, J., Savarese, S.: 4d spatio-temporal convnets: Minkowski convolutional neural networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3075\u20133084 (2019)","DOI":"10.1109\/CVPR.2019.00319"},{"key":"1395_CR6","doi-asserted-by":"crossref","unstructured":"Luo, W., Yang, B., Urtasun, R.: Fast and furious: Real time end-to-end 3D detection, tracking and motion forecasting with a single convolutional net. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2018)","DOI":"10.1109\/CVPR.2018.00376"},{"key":"1395_CR7","unstructured":"Qi, C.R., Su, H., Mo, K., Guibas, L.J.: PointNet: deep learning on point sets for 3D classification and segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2017)"},{"key":"1395_CR8","unstructured":"Qi, C.R., Yi, L., Su, H., Guibas, L.J.: PointNet++ deep hierarchical feature learning on point sets in a metric space. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, pp. 5105\u20135114 (2017)"},{"key":"1395_CR9","doi-asserted-by":"crossref","unstructured":"Shen, Y., Feng, C., Yang, Y., Tian, D.: Mining point cloud local structures by kernel correlation and graph pooling. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4548\u20134557 (2018)","DOI":"10.1109\/CVPR.2018.00478"},{"issue":"5","key":"1395_CR10","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3326362","volume":"38","author":"Y Wang","year":"2019","unstructured":"Wang, Y., Sun, Y., Liu, Z., Sarma, S.E., Bronstein, M.M., Solomon, J.M.: Dynamic graph CNN for learning on point clouds. ACM Trans Graphics (TOG) 38(5), 1\u201312 (2019)","journal-title":"ACM Trans Graphics (TOG)"},{"key":"1395_CR11","doi-asserted-by":"crossref","unstructured":"Liu, X., Yan, M., Bohg, J.: MeteorNet: deep learning on dynamic 3d point cloud sequences. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 9246\u20139255 (2019)","DOI":"10.1109\/ICCV.2019.00934"},{"key":"1395_CR12","unstructured":"Fan, H., Yu, X., Ding, Y., Yang, Y., Kankanhalli, M.: Pstnet: Point spatio-temporal convolution on point cloud sequences. arXiv e-prints 2205 (2022)"},{"key":"1395_CR13","doi-asserted-by":"crossref","unstructured":"Fan, H., Yang, Y., Kankanhalli, M.: Point 4D transformer networks for spatio-temporal modeling in point cloud videos. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14204\u201314213 (2021)","DOI":"10.1109\/CVPR46437.2021.01398"},{"key":"1395_CR14","doi-asserted-by":"crossref","unstructured":"Li, X., Huang, Q., Yang, T., Wu, Q.: Hyperpointnet for point cloud sequence-based 3D human action recognition. In: 2022 IEEE International Conference on Multimedia and Expo (ICME), pp. 1\u20136. IEEE (2022)","DOI":"10.1109\/ICME52920.2022.9859807"},{"key":"1395_CR15","unstructured":"Li, X., Huang, Q., Wang, Z., Hou, Z., Yang, T.: Sequentialpointnet: a strong frame-level parallel point cloud sequence network for 3D action recognition. arXiv preprint arXiv:2111.08492 (2021)"},{"key":"1395_CR16","unstructured":"De\u00a0Smedt, Q., Wannous, H., Vandeborre, J.-P., Guerry, J., Saux, B.L., Filliat, D.: 3D hand gesture recognition using a depth and skeletal dataset: Shrec\u201917 track. In: Proceedings of the Workshop on 3D Object Retrieval, pp. 33\u201338 (2017)"},{"key":"1395_CR17","doi-asserted-by":"crossref","unstructured":"Lu, L., Lu, Y., Wang, S.: Learning multi-level interaction relations and feature representations for group activity recognition. In: MultiMedia Modeling: 27th International Conference, MMM 2021, Prague, Czech Republic, June 22\u201324, 2021, Proceedings, Part I 27, pp. 617\u2013628 . Springer (2021)","DOI":"10.1007\/978-3-030-67832-6_50"},{"issue":"2","key":"1395_CR18","doi-asserted-by":"publisher","first-page":"524","DOI":"10.1109\/TMM.2019.2930344","volume":"22","author":"L Lu","year":"2019","unstructured":"Lu, L., Lu, Y., Yu, R., Di, H., Zhang, L., Wang, S.: GAIM: graph attention interaction model for collective activity recognition. IEEE Trans. Multimedia 22(2), 524\u2013539 (2019)","journal-title":"IEEE Trans. Multimedia"},{"issue":"3","key":"1395_CR19","doi-asserted-by":"publisher","first-page":"1646","DOI":"10.1109\/TCSVT.2021.3075470","volume":"32","author":"J Gao","year":"2021","unstructured":"Gao, J., Xu, C.: Learning video moment retrieval without a single annotated video. IEEE Trans. Circuits Syst. Video Technol. 32(3), 1646\u20131657 (2021)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"10","key":"1395_CR20","doi-asserted-by":"publisher","first-page":"3476","DOI":"10.1109\/TPAMI.2020.2985708","volume":"43","author":"J Gao","year":"2020","unstructured":"Gao, J., Zhang, T., Xu, C.: Learning to model relationships for zero-shot video classification. IEEE Trans. Pattern Anal. Mach. Intell. 43(10), 3476\u20133491 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1395_CR21","doi-asserted-by":"publisher","first-page":"5410","DOI":"10.1109\/TMM.2023.3333206","volume":"26","author":"Y Hu","year":"2023","unstructured":"Hu, Y., Gao, J., Dong, J., Fan, B., Liu, H.: Exploring rich semantics for open-set action recognition. IEEE Trans. Multimedia 26, 5410\u20135421 (2023)","journal-title":"IEEE Trans. Multimedia"},{"key":"1395_CR22","doi-asserted-by":"publisher","first-page":"15949","DOI":"10.1109\/TPAMI.2023.3311447","volume":"45","author":"J Gao","year":"2023","unstructured":"Gao, J., Chen, M., Xu, C.: Vectorized evidential learning for weakly-supervised temporal action localization. IEEE Trans. Pattern Anal. Mach. Intell. 45, 15949\u201315963 (2023)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1395_CR23","doi-asserted-by":"crossref","unstructured":"Yan, S., Xiong, Y., Lin, D.: Spatial temporal graph convolutional networks for skeleton-based action recognition. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32 (2018)","DOI":"10.1609\/aaai.v32i1.12328"},{"key":"1395_CR24","doi-asserted-by":"crossref","unstructured":"Hang, R., Li, M.: Spatial-temporal adaptive graph convolutional network for skeleton-based action recognition. In: Proceedings of the Asian Conference on Computer Vision, pp. 1265\u20131281 (2022)","DOI":"10.1007\/978-3-031-26316-3_11"},{"key":"1395_CR25","doi-asserted-by":"publisher","first-page":"21546","DOI":"10.1109\/ACCESS.2023.3247820","volume":"11","author":"M Rahevar","year":"2023","unstructured":"Rahevar, M., Ganatra, A., Saba, T., Rehman, A., Bahaj, S.A.: Spatial-temporal dynamic graph attention network for skeleton-based action recognition. IEEE Access 11, 21546\u201321553 (2023)","journal-title":"IEEE Access"},{"issue":"8","key":"1395_CR26","doi-asserted-by":"publisher","first-page":"13794","DOI":"10.1109\/TITS.2021.3128424","volume":"23","author":"F Lu","year":"2021","unstructured":"Lu, F., Chen, G., Li, Z., Zhang, L., Liu, Y., Qu, S., Knoll, A.: MoNet: motion-based point cloud prediction network. IEEE Trans. Intell. Transp. Syst. 23(8), 13794\u201313804 (2021)","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"1395_CR27","doi-asserted-by":"crossref","unstructured":"Huang, R., Zhang, W., Kundu, A., Pantofaru, C., Ross, D.A., Funkhouser, T., Fathi, A.: An lstm approach to temporal 3D object detection in lidar point clouds. In: Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XVIII 16, pp. 266\u2013282 . Springer (2020)","DOI":"10.1007\/978-3-030-58523-5_16"},{"key":"1395_CR28","doi-asserted-by":"crossref","unstructured":"Zhao, Y., Birdal, T., Deng, H., Tombari, F.: 3D point capsule networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1009\u20131018 (2019)","DOI":"10.1109\/CVPR.2019.00110"},{"key":"1395_CR29","doi-asserted-by":"crossref","unstructured":"Wang, Y., Xiao, Y., Xiong, F., Jiang, W., Cao, Z., Zhou, J.T., Yuan, J.: 3Dv: 3D dynamic voxel for action recognition in depth video. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 511\u2013520 (2020)","DOI":"10.1109\/CVPR42600.2020.00059"},{"key":"1395_CR30","unstructured":"Fan, H., Yang, Y.: PointRNN: point recurrent neural network for moving point cloud processing. arXiv preprint arXiv:1910.08287 (2019)"},{"key":"1395_CR31","doi-asserted-by":"crossref","unstructured":"Min, Y., Zhang, Y., Chai, X., Chen, X.: An efficient pointLSTM for point clouds based gesture recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5761\u20135770 (2020)","DOI":"10.1109\/CVPR42600.2020.00580"},{"key":"1395_CR32","doi-asserted-by":"crossref","unstructured":"Li, W., Zhang, Z., Liu, Z.: Action recognition based on a bag of 3D points. In: 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition-workshops, pp. 9\u201314 . IEEE (2010)","DOI":"10.1109\/CVPRW.2010.5543273"},{"key":"1395_CR33","doi-asserted-by":"crossref","unstructured":"Shahroudy, A., Liu, J., Ng, T.-T., Wang, G.: NTU RGB+ D: a large scale dataset for 3D human activity analysis. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1010\u20131019 (2016)","DOI":"10.1109\/CVPR.2016.115"},{"key":"1395_CR34","doi-asserted-by":"crossref","unstructured":"Chen, C., Jafari, R., Kehtarnavaz, N.: UTD-MHAD: a multimodal dataset for human action recognition utilizing a depth camera and a wearable inertial sensor. In: 2015 IEEE International Conference on Image Processing (ICIP), pp. 168\u2013172 . IEEE (2015)","DOI":"10.1109\/ICIP.2015.7350781"},{"key":"1395_CR35","doi-asserted-by":"crossref","unstructured":"Wang, J., Liu, Z., Wu, Y., Yuan, J.: Mining actionlet ensemble for action recognition with depth cameras. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition, pp. 1290\u20131297 . IEEE (2012)","DOI":"10.1109\/CVPR.2012.6247813"},{"key":"1395_CR36","doi-asserted-by":"crossref","unstructured":"Zhang, X., Wang, Y., Gou, M., Sznaier, M., Camps, O.: Efficient temporal sequence comparison and classification using gram matrix embeddings on a Riemannian manifold. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4498\u20134507 (2016)","DOI":"10.1109\/CVPR.2016.487"},{"key":"1395_CR37","doi-asserted-by":"crossref","unstructured":"Klaser, A., Marsza\u0142ek, M., Schmid, C.: A spatio-temporal descriptor based on 3D-gradients. In: BMVC 2008-19th British Machine Vision Conference, pp. 275\u20131 . British Machine Vision Association (2008)","DOI":"10.5244\/C.22.99"},{"key":"1395_CR38","doi-asserted-by":"crossref","unstructured":"Vieira, A.W., Nascimento, E.R., Oliveira, G.L., Liu, Z., Campos, M.F.: Stop: space-time occupancy patterns for 3D action recognition from depth map sequences. In: Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications: 17th Iberoamerican Congress, CIARP 2012, Buenos Aires, Argentina, September 3\u20136, 2012. Proceedings 17, pp. 252\u2013259 . Springer (2012)","DOI":"10.1007\/978-3-642-33275-3_31"},{"issue":"5","key":"1395_CR39","doi-asserted-by":"publisher","first-page":"1051","DOI":"10.1109\/TMM.2018.2818329","volume":"20","author":"P Wang","year":"2018","unstructured":"Wang, P., Li, W., Gao, Z., Tang, C., Ogunbona, P.O.: Depth pooling based large-scale 3-D action recognition with convolutional neural networks. IEEE Trans. Multimedia 20(5), 1051\u20131061 (2018)","journal-title":"IEEE Trans. Multimedia"},{"key":"1395_CR40","doi-asserted-by":"publisher","first-page":"287","DOI":"10.1016\/j.ins.2018.12.050","volume":"480","author":"Y Xiao","year":"2019","unstructured":"Xiao, Y., Chen, J., Wang, Y., Cao, Z., Zhou, J.T., Bai, X.: Action recognition for depth video using multi-view dynamic images. Inf. Sci. 480, 287\u2013304 (2019)","journal-title":"Inf. Sci."},{"issue":"17","key":"1395_CR41","doi-asserted-by":"publisher","first-page":"24119","DOI":"10.1007\/s11042-022-12091-z","volume":"81","author":"A Sanchez-Caballero","year":"2022","unstructured":"Sanchez-Caballero, A., de L\u00f3pez-Diz, S., Fuentes-Jimenez, D., Losada-Guti\u00e9rrez, C., Marr\u00f3n-Romera, M., Casillas-Perez, D., Sarker, M.I.: 3DFCNN: real-time action recognition using 3D deep neural networks with raw depth information. Multimedia Tools Appl. 81(17), 24119\u201324143 (2022)","journal-title":"Multimedia Tools Appl."},{"key":"1395_CR42","unstructured":"Sanchez-Caballero, A., Fuentes-Jim\u00e9nez, D., Losada-Guti\u00e9rrez, C.: Exploiting the convlstm: Human action recognition using raw depth video-based recurrent neural networks. arXiv preprint arXiv:2006.07744 (2020)"},{"key":"1395_CR43","doi-asserted-by":"crossref","unstructured":"Li, M., Chen, S., Chen, X., Zhang, Y., Wang, Y., Tian, Q.: Actional-structural graph convolutional networks for skeleton-based action recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3595\u20133603 (2019)","DOI":"10.1109\/CVPR.2019.00371"},{"key":"1395_CR44","doi-asserted-by":"crossref","unstructured":"Si, C., Chen, W., Wang, W., Wang, L., Tan, T.: An attention enhanced graph convolutional LSTM network for skeleton-based action recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1227\u20131236 (2019)","DOI":"10.1109\/CVPR.2019.00132"},{"key":"1395_CR45","doi-asserted-by":"publisher","first-page":"9532","DOI":"10.1109\/TIP.2020.3028207","volume":"29","author":"L Shi","year":"2020","unstructured":"Shi, L., Zhang, Y., Cheng, J., Lu, H.: Skeleton-based action recognition with multi-stream adaptive graph convolutional networks. IEEE Trans. Image Process. 29, 9532\u20139545 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"1395_CR46","doi-asserted-by":"crossref","unstructured":"Li, L., Wang, M., Ni, B., Wang, H., Yang, J., Zhang, W.: 3D human action representation learning via cross-view consistency pursuit. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4741\u20134750 (2021)","DOI":"10.1109\/CVPR46437.2021.00471"},{"issue":"6","key":"1395_CR47","doi-asserted-by":"publisher","first-page":"3316","DOI":"10.1109\/TPAMI.2021.3053765","volume":"44","author":"M Li","year":"2021","unstructured":"Li, M., Chen, S., Chen, X., Zhang, Y., Wang, Y., Tian, Q.: Symbiotic graph neural networks for 3D skeleton-based human action recognition and motion prediction. IEEE Trans. Pattern Anal. Mach. Intell. 44(6), 3316\u20133333 (2021)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1395_CR48","doi-asserted-by":"publisher","DOI":"10.1016\/j.cviu.2023.103722","volume":"233","author":"AF Bavil","year":"2023","unstructured":"Bavil, A.F., Damirchi, H., Taghirad, H.D.: Action capsules: human skeleton action recognition. Comput. Vis. Image Underst. 233, 103722 (2023)","journal-title":"Comput. Vis. Image Underst."},{"issue":"10","key":"1395_CR49","doi-asserted-by":"publisher","first-page":"4648","DOI":"10.1109\/TIP.2017.2718189","volume":"26","author":"B Zhang","year":"2017","unstructured":"Zhang, B., Yang, Y., Chen, C., Yang, L., Han, J., Shao, L.: Action recognition using 3D histograms of texture and a multi-class boosting classifier. IEEE Trans. Image Process. 26(10), 4648\u20134660 (2017)","journal-title":"IEEE Trans. Image Process."},{"issue":"11","key":"1395_CR50","doi-asserted-by":"publisher","first-page":"5275","DOI":"10.1109\/TIP.2018.2855438","volume":"27","author":"NED Elmadany","year":"2018","unstructured":"Elmadany, N.E.D., He, Y., Guan, L.: Information fusion for human action recognition via biset\/multiset globality locality preserving canonical correlation analysis. IEEE Trans. Image Process. 27(11), 5275\u20135287 (2018)","journal-title":"IEEE Trans. Image Process."},{"issue":"9","key":"1395_CR51","doi-asserted-by":"publisher","first-page":"1806","DOI":"10.1109\/TSMC.2018.2850149","volume":"49","author":"A Kamel","year":"2018","unstructured":"Kamel, A., Sheng, B., Yang, P., Li, P., Shen, R., Feng, D.D.: Deep convolutional neural networks for human action recognition using depth maps and postures. IEEE Trans. Syst. Man Cybern. Syst. 49(9), 1806\u20131819 (2018)","journal-title":"IEEE Trans. Syst. Man Cybern. Syst."},{"issue":"5","key":"1395_CR52","doi-asserted-by":"publisher","first-page":"1317","DOI":"10.1109\/TMM.2018.2875510","volume":"21","author":"NED Elmadany","year":"2018","unstructured":"Elmadany, N.E.D., He, Y., Guan, L.: Multimodal learning for human action recognition via bimodal\/multimodal hybrid centroid canonical correlation analysis. IEEE Trans. Multimedia 21(5), 1317\u20131331 (2018)","journal-title":"IEEE Trans. Multimedia"},{"key":"1395_CR53","doi-asserted-by":"publisher","first-page":"135118","DOI":"10.1109\/ACCESS.2020.3006067","volume":"8","author":"T Yang","year":"2020","unstructured":"Yang, T., Hou, Z., Liang, J., Gu, Y., Chao, X.: Depth sequential information entropy maps and multi-label subspace learning for human action recognition. IEEE Access 8, 135118\u2013135130 (2020)","journal-title":"IEEE Access"},{"issue":"21","key":"1395_CR54","doi-asserted-by":"publisher","first-page":"14551","DOI":"10.1007\/s00521-021-06097-1","volume":"33","author":"J Trelinski","year":"2021","unstructured":"Trelinski, J., Kwolek, B.: CNN-based and DTW features for human activity recognition on depth maps. Neural Comput. Appl. 33(21), 14551\u201314563 (2021)","journal-title":"Neural Comput. Appl."},{"issue":"3","key":"1395_CR55","doi-asserted-by":"publisher","first-page":"1250","DOI":"10.1109\/TCSVT.2021.3077512","volume":"32","author":"H Wu","year":"2021","unstructured":"Wu, H., Ma, X., Li, Y.: Spatiotemporal multimodal learning with 3D CNNs for video action recognition. IEEE Trans. Circuits Syst. Video Technol. 32(3), 1250\u20131261 (2021)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1395_CR56","doi-asserted-by":"crossref","unstructured":"De\u00a0Smedt, Q., Wannous, H., Vandeborre, J.-P.: Skeleton-based dynamic hand gesture recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 1\u20139 (2016)","DOI":"10.1109\/CVPRW.2016.153"},{"key":"1395_CR57","doi-asserted-by":"crossref","unstructured":"Hou, J., Wang, G., Chen, X., Xue, J.-H., Zhu, R., Yang, H.: Spatial-temporal attention res-TCN for skeleton-based dynamic hand gesture recognition. In: Proceedings of the European Conference on Computer Vision (ECCV) Workshops (2018)","DOI":"10.1007\/978-3-030-11024-6_18"},{"key":"1395_CR58","unstructured":"Chen, Y., Zhao, L., Peng, X., Yuan, J., Metaxas, D.N.: Construct dynamic graphs for hand gesture recognition via spatial-temporal attention. arXiv preprint arXiv:1907.08871 (2019)"},{"issue":"4","key":"1395_CR59","doi-asserted-by":"publisher","first-page":"7823","DOI":"10.1109\/LRA.2021.3101822","volume":"6","author":"A Sabater","year":"2021","unstructured":"Sabater, A., Alonso, I., Montesano, L., Murillo, A.C.: Domain and view-point agnostic hand action recognition. IEEE Robot. Autom. Lett. 6(4), 7823\u20137830 (2021)","journal-title":"IEEE Robot. Autom. Lett."},{"issue":"9","key":"1395_CR60","doi-asserted-by":"publisher","first-page":"6227","DOI":"10.1109\/TCSVT.2022.3165069","volume":"32","author":"J-H Song","year":"2022","unstructured":"Song, J.-H., Kong, K., Kang, S.-J.: Dynamic hand gesture recognition using improved spatio-temporal graph convolutional network. IEEE Trans. Circuits Syst. Video Technol. 32(9), 6227\u20136239 (2022)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1395_CR61","doi-asserted-by":"publisher","first-page":"811","DOI":"10.1109\/TMM.2023.3271811","volume":"26","author":"J Liu","year":"2023","unstructured":"Liu, J., Wang, X., Wang, C., Gao, Y., Liu, M.: Temporal decoupling graph convolutional network for skeleton-based gesture recognition. IEEE Trans. Multimedia 26, 811\u2013823 (2023)","journal-title":"IEEE Trans. Multimedia"},{"key":"1395_CR62","doi-asserted-by":"crossref","unstructured":"Bigalke, A., Heinrich, M.P.: Fusing posture and position representations for point cloud-based hand gesture recognition. In: 2021 International Conference on 3D Vision (3DV), pp. 617\u2013626. IEEE (2021)","DOI":"10.1109\/3DV53792.2021.00071"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01395-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-024-01395-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01395-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,22]],"date-time":"2024-11-22T21:01:20Z","timestamp":1732309280000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-024-01395-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,27]]},"references-count":62,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2024,8]]}},"alternative-id":["1395"],"URL":"https:\/\/doi.org\/10.1007\/s00530-024-01395-9","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,6,27]]},"assertion":[{"value":"12 November 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 June 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 June 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"All authors of this research paper declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"192"}}