{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T18:38:37Z","timestamp":1775068717252,"version":"3.50.1"},"reference-count":44,"publisher":"MDPI AG","issue":"5","license":[{"start":{"date-parts":[[2017,5,11]],"date-time":"2017-05-11T00:00:00Z","timestamp":1494460800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Sensors"],"abstract":"<jats:p>Human activity recognition is an important area in computer vision, with its wide range of applications including ambient assisted living. In this paper, an activity recognition system based on skeleton data extracted from a depth camera is presented. The system makes use of machine learning techniques to classify the actions that are described with a set of a few basic postures. The training phase creates several models related to the number of clustered postures by means of a multiclass Support Vector Machine (SVM), trained with Sequential Minimal Optimization (SMO). The classification phase adopts the X-means algorithm to find the optimal number of clusters dynamically. The contribution of the paper is twofold. The first aim is to perform activity recognition employing features based on a small number of informative postures, extracted independently from each activity instance; secondly, it aims to assess the minimum number of frames needed for an adequate classification. The system is evaluated on two publicly available datasets, the Cornell Activity Dataset (CAD-60) and the Telecommunication Systems Team (TST) Fall detection dataset. The number of clusters needed to model each instance ranges from two to four elements. The proposed approach reaches excellent performances using only about 4 s of input data (~100 frames) and outperforms the state of the art when it uses approximately 500 frames on the CAD-60 dataset. The results are promising for the test in real context.<\/jats:p>","DOI":"10.3390\/s17051100","type":"journal-article","created":{"date-parts":[[2017,5,11]],"date-time":"2017-05-11T11:06:46Z","timestamp":1494500806000},"page":"1100","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":49,"title":["A Human Activity Recognition System Based on Dynamic Clustering of Skeleton Data"],"prefix":"10.3390","volume":"17","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-6147-7692","authenticated-orcid":false,"given":"Alessandro","family":"Manzi","sequence":"first","affiliation":[{"name":"The BioRobotics Institute, Scuola Superiore Sant\u2019Anna, Viale Rinaldo Piaggio, 34, 56026 Pontedera (PI), Italy"}]},{"given":"Paolo","family":"Dario","sequence":"additional","affiliation":[{"name":"The BioRobotics Institute, Scuola Superiore Sant\u2019Anna, Viale Rinaldo Piaggio, 34, 56026 Pontedera (PI), Italy"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7432-5033","authenticated-orcid":false,"given":"Filippo","family":"Cavallo","sequence":"additional","affiliation":[{"name":"The BioRobotics Institute, Scuola Superiore Sant\u2019Anna, Viale Rinaldo Piaggio, 34, 56026 Pontedera (PI), Italy"}]}],"member":"1968","published-online":{"date-parts":[[2017,5,11]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"976","DOI":"10.1016\/j.imavis.2009.11.014","article-title":"A survey on vision-based human action recognition","volume":"28","author":"Poppe","year":"2010","journal-title":"Image Vis. Comput."},{"key":"ref_2","unstructured":"Yamato, J., Ohya, J., and Ishii, K. (1992, January 15\u201318). Recognizing human action in time-sequential images using hidden markov model. Proceedings of the 1992 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR\u201992), Champaign, IL, USA."},{"key":"ref_3","unstructured":"Kellokumpu, V., Pietik\u00e4inen, M., and Heikkil\u00e4, J. (2005, January 16\u201318). Human activity recognition using sequences of postures. Proceedings of the IAPR Conference on Machine Vision Applications (IAPR MVA 2005), Tsukuba Science City, Japan."},{"key":"ref_4","doi-asserted-by":"crossref","unstructured":"Scholkopf, B., and Smola, A.J. (2001). Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond, MIT Press.","DOI":"10.7551\/mitpress\/4175.001.0001"},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Willems, G., Tuytelaars, T., and Van Gool, L. (2008, January 12\u201318). An efficient dense and scale-invariant spatio-temporal interest point detector. Proceedings of the European Conference on Computer Vision, Marseille, France.","DOI":"10.1007\/978-3-540-88688-4_48"},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"16","DOI":"10.1145\/1922649.1922653","article-title":"Human activity analysis: A review","volume":"43","author":"Aggarwal","year":"2011","journal-title":"ACM Comput. Surv."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"224","DOI":"10.1016\/j.cviu.2010.10.002","article-title":"A survey of vision-based methods for action representation, segmentation and recognition","volume":"115","author":"Weinland","year":"2011","journal-title":"Comput. Vis. Image Underst."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"887","DOI":"10.1016\/j.cviu.2010.05.002","article-title":"Photometric stereo with an arbitrary number of illuminants","volume":"114","author":"Argyriou","year":"2010","journal-title":"Comput. Vis. Image Underst."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"70","DOI":"10.1016\/j.patrec.2014.04.011","article-title":"Human activity recognition from 3d data: A review","volume":"48","author":"Aggarwal","year":"2014","journal-title":"Pattern Recognit. Lett."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"116","DOI":"10.1145\/2398356.2398381","article-title":"Real-time human pose recognition in parts from single depth images","volume":"56","author":"Shotton","year":"2013","journal-title":"Commun. ACM"},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"1318","DOI":"10.1109\/TCYB.2013.2265378","article-title":"Enhanced computer vision with microsoft kinect sensor: A review","volume":"43","author":"Han","year":"2013","journal-title":"IEEE Trans. Cybern."},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"27","DOI":"10.1109\/MPUL.2011.940428","article-title":"Technology and innovative services","volume":"2","author":"Turchetti","year":"2011","journal-title":"IEEE Pulse"},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Cavallo, F., Aquilano, M., Bonaccorsi, M., Mannari, I., Carrozza, M., and Dario, P. (September, January 30). Multidisciplinary approach for developing a new robotic system for domiciliary assistance to elderly people. Proceedings of the 2011 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Boston, MA, USA.","DOI":"10.1109\/IEMBS.2011.6091318"},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"12959","DOI":"10.3390\/s150612959","article-title":"Visual privacy by context: Proposal and evaluation of a level-based visualisation scheme","volume":"15","author":"Chaaraoui","year":"2015","journal-title":"Sensors"},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"1383","DOI":"10.1109\/TCYB.2013.2276433","article-title":"Multilevel depth and image fusion for human activity detection","volume":"43","author":"Ni","year":"2013","journal-title":"IEEE Trans. Cybern."},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Ni, B., Wang, G., and Moulin, P. (2013). Rgbd-hudaact: A color-depth video database for human daily activity recognition. Consumer Depth Cameras for Computer Vision, Springer.","DOI":"10.1007\/978-1-4471-4640-7_10"},{"key":"ref_17","unstructured":"Sung, J., Ponce, C., Selman, B., and Saxena, A. (2012, January 14\u201318). Unstructured human activity detection from rgbd images. Proceedings of the 2012 IEEE International Conference on Robotics and Automation (ICRA), St. Paul, MN, USA."},{"key":"ref_18","doi-asserted-by":"crossref","unstructured":"Wang, J., Liu, Z., and Wu, Y. (2014). Learning actionlet ensemble for 3D human action recognition. Human Action Recognition with Depth Cameras, Springer.","DOI":"10.1007\/978-3-319-04561-0_2"},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"453","DOI":"10.1016\/j.imavis.2014.04.005","article-title":"Evaluating spatiotemporal interest point features for depth-based action recognition","volume":"32","author":"Zhu","year":"2014","journal-title":"Image Vis. Comput."},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"586","DOI":"10.1109\/THMS.2014.2377111","article-title":"Human activity recognition process using 3D posture data","volume":"45","author":"Gaglio","year":"2015","journal-title":"IEEE Trans. Hum.-Mach. Syst."},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"329","DOI":"10.1016\/j.jvcir.2014.10.009","article-title":"STFC: Spatio-temporal feature chain for skeleton-based human action recognition","volume":"26","author":"Ding","year":"2015","journal-title":"J. Vis. Commun. Image Represent."},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"29","DOI":"10.1016\/j.image.2015.02.004","article-title":"Informative joints based human action recognition using skeleton contexts","volume":"33","author":"Jiang","year":"2015","journal-title":"Signal Process. Image Commun."},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"786","DOI":"10.1016\/j.eswa.2013.08.009","article-title":"Evolutionary joint selection to improve human action recognition with RGB-D devices","volume":"41","author":"Chaaraoui","year":"2014","journal-title":"Expert Syst. Appl."},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Baysal, S., Kurt, M.C., and Duygulu, P. (2010, January 23\u201326). Recognizing human actions using key poses. Proceedings of the 2010 20th International Conference on Pattern Recognition (ICPR), Istanbul, Turkey.","DOI":"10.1109\/ICPR.2010.427"},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Ballan, L., Bertini, M., Bimbo, A.D., Seidenari, L., and Serra, G. (October, January 29). Effective Codebooks for human action categorization. Proceedings of the 2009 IEEE 12th International Conference on Computer Vision Workshops (ICCV Workshops), Kyoto, Japan.","DOI":"10.1109\/ICCVW.2009.5457658"},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Raptis, M., and Sigal, L. (2013, January 25\u201327). Poselet Key-Framing: A Model for Human Activity Recognition. Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition (CVPR \u201913), Oregon, Portland.","DOI":"10.1109\/CVPR.2013.342"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Shan, J., and Akella, S. (2014, January 11\u201313). 3D human action segmentation and recognition using pose kinetic energy. Proceedings of the 2014 IEEE International Workshop on Advanced Robotics and its Social Impacts, Evanston, IL, USA.","DOI":"10.1109\/ARSO.2014.7020983"},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"Zhu, G., Zhang, L., Shen, P., Song, J., Zhi, L., and Yi, K. (2015, January 6\u20139). Human action recognition using key poses and atomic motions. Proceedings of the 2015 IEEE International Conference on Robotics and Biomimetics (ROBIO), Zhuhai, China.","DOI":"10.1109\/ROBIO.2015.7418936"},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Zhu, G., Zhang, L., Shen, P., and Song, J. (2016). An Online Continuous Human Action Recognition Algorithm Based on the Kinect Sensor. Sensors, 16.","DOI":"10.3390\/s16020161"},{"key":"ref_30","doi-asserted-by":"crossref","first-page":"4351435","DOI":"10.1155\/2016\/4351435","article-title":"A Human Activity Recognition System Using Skeleton Data from RGBD Sensors","volume":"2016","author":"Cippitelli","year":"2016","journal-title":"Comput. Intell. Neurosci."},{"key":"ref_31","unstructured":"Hua, G., and J\u00e9gou, H. (15\u201316, January 8\u201310). A 3D Human Posture Approach for Activity Recognition Based on Depth Camera. Proceedings of the Computer Vision\u2014ECCV 2016 Workshops, Amsterdam, The Netherlands."},{"key":"ref_32","unstructured":"Microsoft (2017, May 10). Natural User Interface for Kinect. Available online: https:\/\/msdn.microsoft.com\/en-us\/library\/hh855352.aspx."},{"key":"ref_33","unstructured":"MacQueen, J. (1966, January 7). Some methods for classification and analysis of multivariate observations. Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability, Berkeley, California."},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"881","DOI":"10.1109\/TPAMI.2002.1017616","article-title":"An efficient k-means clustering algorithm: Analysis and implementation","volume":"24","author":"Kanungo","year":"2002","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_35","unstructured":"Arthur, D., and Vassilvitskii, S. (2007, January 7\u20139). K-means++: The advantages of carefull seeding. Proceedings of the Eighteenth Annual ACM-SIAM Symposium on Discrete Algorithms, New Orleans, LA, USA."},{"key":"ref_36","unstructured":"Pelleg, D., and Moore, A.W. (July, January 29). X-means: Extending K-means with Efficient Estimation of the Number of Clusters. Proceedings of the Seventeenth International Conference on Machine Learning, Morgan Kaufmann, Stanford, CA, USA."},{"key":"ref_37","unstructured":"Witten, I.H., Frank, E., and Hall, M.A. (2011). Data Mining: Practical Machine Learning Tools and Techniques, Morgan Kaufmann Publishers Inc.. [3rd ed.]."},{"key":"ref_38","unstructured":"Schoelkopf, B., Burges, C., and Smola, A. (1998). Fast Training of Support Vector Machines using Sequential Minimal Optimization. Advances in Kernel Methods\u2014Support Vector Learning, MIT Press."},{"key":"ref_39","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/1961189.1961199","article-title":"LIBSVM: A library for support vector machines","volume":"2","author":"Chang","year":"2011","journal-title":"ACM Trans. Intell. Syst. Technol."},{"key":"ref_40","unstructured":"Jordan, M.I., Kearns, M.J., and Solla, S.A. (1998). Classification by Pairwise Coupling. Advances in Neural Information Processing Systems, MIT Press."},{"key":"ref_41","doi-asserted-by":"crossref","first-page":"10","DOI":"10.1145\/1656274.1656278","article-title":"The WEKA Data Mining Software: An Update","volume":"11","author":"Hall","year":"2009","journal-title":"SIGKDD Explor. Newsl."},{"key":"ref_42","doi-asserted-by":"crossref","unstructured":"Gasparrini, S., Cippitelli, E., Gambi, E., Spinsante, S., W\u00e5hsl\u00e9n, J., Orhan, I., and Lindh, T. (2016). Proposal and experimental evaluation of fall detection solution based on wearable and depth data fusion. ICT Innovations 2015, Springer.","DOI":"10.1007\/978-3-319-25733-4_11"},{"key":"ref_43","doi-asserted-by":"crossref","unstructured":"Faria, D.R., Premebida, C., and Nunes, U. (2014, January 25\u201329). A probabilistic approach for human everyday activities recognition using body motion from RGB-D images. Proceedings of the 23rd IEEE International Symposium on Robot and Human Interactive Communication, Edinburgh, UK.","DOI":"10.1109\/ROMAN.2014.6926340"},{"key":"ref_44","doi-asserted-by":"crossref","first-page":"3","DOI":"10.3389\/fnbot.2015.00003","article-title":"Self-Organizing Neural Integration of Pose-Motion Features for Human Action Recognition","volume":"9","author":"Parisi","year":"2015","journal-title":"Front. Neurorobot."}],"container-title":["Sensors"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1424-8220\/17\/5\/1100\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T18:35:28Z","timestamp":1760207728000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1424-8220\/17\/5\/1100"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2017,5,11]]},"references-count":44,"journal-issue":{"issue":"5","published-online":{"date-parts":[[2017,5]]}},"alternative-id":["s17051100"],"URL":"https:\/\/doi.org\/10.3390\/s17051100","relation":{},"ISSN":["1424-8220"],"issn-type":[{"value":"1424-8220","type":"electronic"}],"subject":[],"published":{"date-parts":[[2017,5,11]]}}}