{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,4]],"date-time":"2026-04-04T18:15:31Z","timestamp":1775326531766,"version":"3.50.1"},"reference-count":28,"publisher":"Springer Science and Business Media LLC","issue":"16","license":[{"start":{"date-parts":[[2021,4,4]],"date-time":"2021-04-04T00:00:00Z","timestamp":1617494400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2021,4,4]],"date-time":"2021-04-04T00:00:00Z","timestamp":1617494400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"published-print":{"date-parts":[[2021,7]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Automatic facial expression analysis remains challenging due to its low recognition accuracy and poor robustness. In this study, we utilized active learning and support vector machine (SVM) algorithms to classify facial action units (AU) for human facial expression recognition. Active learning was used to detect the targeted facial expression AUs, while an SVM was utilized to classify different AUs and ultimately map them to their corresponding facial expressions. Active learning reduces the number of non-support vectors in the training sample set and shortens the labeling and training times without affecting the performance of the classifier, thereby reducing the cost of labeling samples and improving the training speed. Experimental results show that the proposed algorithm can effectively suppress correlated noise and achieve higher recognition rates than principal component analysis and a human observer on seven different facial expressions.<\/jats:p>","DOI":"10.1007\/s11042-021-10836-w","type":"journal-article","created":{"date-parts":[[2021,4,4]],"date-time":"2021-04-04T11:02:16Z","timestamp":1617534136000},"page":"24287-24301","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":66,"title":["Action unit classification for facial expression recognition using active learning and SVM"],"prefix":"10.1007","volume":"80","author":[{"given":"Li","family":"Yao","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7712-7531","authenticated-orcid":false,"given":"Yan","family":"Wan","sequence":"additional","affiliation":[]},{"given":"Hongjie","family":"Ni","sequence":"additional","affiliation":[]},{"given":"Bugao","family":"Xu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,4,4]]},"reference":[{"key":"10836_CR1","doi-asserted-by":"crossref","unstructured":"Abdul-Majjed IO (2017) Emotion recognition system based on facial expressions using svm. Recent Developments in Intelligent Computing, Communication and Devices. pp. 31\u201335.","DOI":"10.1007\/978-981-10-3779-5_5"},{"key":"10836_CR2","unstructured":"AU R-CNN (2019) Encoding expert prior knowledge into R-CNN for action unit detection. Ma C., Chen L., Yong J. H. Neurocomputing"},{"key":"10836_CR3","doi-asserted-by":"crossref","unstructured":"Benitez-Quiroz CF, Srinivasan R, Martinez AM (2016) EmotioNet: an accurate, real-time algorithm for the automatic annotation of a million facial expressions in the wild [C]\/\/ the IEEE conference on computer vision and pattern recognition (CVPR). IEEE:5562\u20135570","DOI":"10.1109\/CVPR.2016.600"},{"issue":"3","key":"10836_CR4","doi-asserted-by":"publisher","first-page":"622","DOI":"10.1109\/TPAMI.2013.187","volume":"36","author":"L Chen","year":"2014","unstructured":"Chen L (2014) A fair comparison should be based on the same protocol--comments on \"trainable convolution filters and their application to face recognition\"[J]. IEEE Trans Pattern Anal Mach Intell 36(3):622\u2013623","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"2","key":"10836_CR5","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2846092","volume":"34","author":"ZY Cheng","year":"2016","unstructured":"Cheng ZY, Shen JL (2016) On effective location-aware music recommendation[J]. ACM Trans Inf Syst 34(2):1\u201313","journal-title":"ACM Trans Inf Syst"},{"key":"10836_CR6","doi-asserted-by":"crossref","unstructured":"Ekman P (1978) A technique for the measurement of facial action. Palo alto [J]. Facial Action Coding System","DOI":"10.1037\/t27734-000"},{"key":"10836_CR7","unstructured":"Friesen W, Ekman P (1983) EMFACS-7: emotional facial action coding system. Unpublished manual, University of California, California"},{"key":"10836_CR8","unstructured":"Friesen W, Ekman P (1983) EMFACS-7: emotional facial action coding system. Unpublished manual, University of California, California"},{"key":"10836_CR9","first-page":"1","volume":"6","author":"A Gudi","year":"2015","unstructured":"Gudi A, Tasli HE, Den Uyl TM et al (2015) Deep learning based FACS unit occurrence and intensity estimation[C]\/\/2015 11th IEEE international conference and workshops on automatic face and gesture recognition(FG). IEEE 6:1\u20135","journal-title":"IEEE"},{"key":"10836_CR10","doi-asserted-by":"crossref","unstructured":"Huang X, Wang SJ, Zhao G et al. (2015) Facial micro-expression recognition using spatiotemporal local binary pattern with integral projection[C]\/\/ second workshop on computer vision for affective computing at Iccv. IEEE Computer Society","DOI":"10.1109\/ICCVW.2015.10"},{"key":"10836_CR11","doi-asserted-by":"crossref","unstructured":"Huang X, Zhao G, Hong X et al. (2016) Spontaneous facial micro-expression analysis using Spatiotemporal Completed Local Quantized Patterns[J]. Neurocomputing, 2016, 175 (JAN.29PT.A):564\u2013578.","DOI":"10.1016\/j.neucom.2015.10.096"},{"issue":"4","key":"10836_CR12","doi-asserted-by":"publisher","first-page":"299","DOI":"10.1109\/TAFFC.2015.2485205","volume":"7","author":"Y Liu","year":"2016","unstructured":"Liu Y, Zhang J, Yan W et al (2016) A Main directional mean optical flow feature for spontaneous micro-expression recognition [J]. IEEE Trans Affect Comput 7(4):299\u2013310","journal-title":"IEEE Trans Affect Comput"},{"issue":"9","key":"10836_CR13","doi-asserted-by":"publisher","first-page":"2248","DOI":"10.1109\/TMM.2019.2902096","volume":"21","author":"L Liu","year":"2019","unstructured":"Liu L, Li G, Xie Y, Yu Y, Wang Q, Lin L (2019) Facial landmark machines: a backbone-branches architecture with progressive representation learning [J]. IEEE Transactions on Multimedia 21(9):2248\u20132262","journal-title":"IEEE Transactions on Multimedia"},{"key":"10836_CR14","unstructured":"Lu WL Facial expression recognition based on emotional geometry and support vector machines [D]. Fudan University"},{"issue":"1","key":"10836_CR15","first-page":"94","volume":"36","author":"P Lucey","year":"2010","unstructured":"Lucey P, Cohn JF, Kanade T et al (2010) The extended Cohn-Kanade dataset (CK+): a complete dataset for action unit and emotion-specified expression. Computer Vision and Pattern Recognition Workshops IEEE 36(1):94\u2013101","journal-title":"Computer Vision and Pattern Recognition Workshops IEEE"},{"key":"10836_CR16","unstructured":"Merghani W, Davison AK, Yap MHA (n.d.) Review on Facial Micro-Expressions Analysis: Datasets, Features and Metrics. PREPRINT SUBMITTED TO IEEE JOURNAL"},{"issue":"1","key":"10836_CR17","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1587\/transinf.E96.D.81","volume":"96","author":"S Polikovsky","year":"2013","unstructured":"Polikovsky S, Kameda Y, Ohta Y et al (2013) Facial micro-expression detection in hi-speed video based on facial action coding system (FACS)[J]. IEICE Trans Inf Syst 96(1):81\u201392","journal-title":"IEICE Trans Inf Syst"},{"issue":"1","key":"10836_CR18","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1587\/transinf.E96.D.81","volume":"96","author":"S Polikovsky","year":"2013","unstructured":"Polikovsky S, Kameda S, Ohta Y (2013) Facial micro-expression detection in hi-speed video based on facial action coding system (FACS) [J]. IEICE Transactions on Information & Systems 96(1):81\u201392","journal-title":"IEICE Transactions on Information & Systems"},{"key":"10836_CR19","doi-asserted-by":"publisher","first-page":"803","DOI":"10.1016\/j.imavis.2008.08.005","volume":"27","author":"C Shan","year":"2009","unstructured":"Shan C, Gong S, McOwan PW (2009) Facial expression recognition based on local binary patterns: a comprehensive study. Image Vis Comput 27:803\u2013816","journal-title":"Image Vis Comput"},{"key":"10836_CR20","doi-asserted-by":"crossref","unstructured":"Shreve M, Godavarthy S, Manohar V et al. (2009) Towards macro-and micro-expression spotting in video using strain patterns, in: Applications of Computer Vision (WACV), pp. 1\u20136.","DOI":"10.1109\/WACV.2009.5403044"},{"key":"10836_CR21","doi-asserted-by":"crossref","unstructured":"Wang SJ, Yan WJ, Li X et al. (2014) Micro-expression recognition using dynamic textures on tensor independent color space[C]\/\/ international conference on pattern recognition. IEEE","DOI":"10.1109\/ICPR.2014.800"},{"issue":"1","key":"10836_CR22","doi-asserted-by":"publisher","first-page":"25","DOI":"10.1007\/s11063-013-9288-7","volume":"39","author":"SJ Wang","year":"2014","unstructured":"Wang SJ, Chen HL, Yan WJ, Chen YH, Fu X (2014) Face recognition and micro-expression recognition based on discriminant tensor subspace analysis plus extreme learning machine. Neural Process Lett 39(1):25\u201343","journal-title":"Neural Process Lett"},{"key":"10836_CR23","volume-title":"LBP with six intersection points: reducing redundant information in LBP-TOP for micro-expression recognition[C]\/\/ Accv","author":"Y Wang","year":"2015","unstructured":"Wang Y, See J, Phan W et al (2015) LBP with six intersection points: reducing redundant information in LBP-TOP for micro-expression recognition[C]\/\/ Accv. Springer, Cham"},{"issue":"7","key":"10836_CR24","first-page":"109","volume":"39","author":"W Xian","year":"2012","unstructured":"Xian W, Yan Z, Xin M et al (2012) Face recognition algorithm based on improved LBP [J]. Opt Eng 39(7):109\u2013114","journal-title":"Opt Eng"},{"issue":"2","key":"10836_CR25","doi-asserted-by":"publisher","first-page":"254","DOI":"10.1109\/TAFFC.2016.2518162","volume":"8","author":"F Xu","year":"2017","unstructured":"Xu F, Zhang J, Wang JZ (2017) Microexpression identification and categorization using a facial dynamics map[J]. IEEE Trans Affect Comput 8(2):254\u2013267","journal-title":"IEEE Trans Affect Comput"},{"issue":"1","key":"10836_CR26","first-page":"9","volume":"17","author":"X Zhang","year":"2005","unstructured":"Zhang X, Shan S, Cao B et al (2005) Cas-peal: a large-scale Chinese face database and some primary evaluations. Journal of Computer Aided Design & Computer Graphics 17(1):9\u201317","journal-title":"Journal of Computer Aided Design & Computer Graphics"},{"key":"10836_CR27","doi-asserted-by":"crossref","unstructured":"Zhao K, Chu WS, De la Torre F. et al (2015) Joint patch and multi-label learning for facial action unit detection[C]\/\/ Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2207\u20132216","DOI":"10.1109\/CVPR.2015.7298833"},{"key":"10836_CR28","doi-asserted-by":"crossref","unstructured":"Zhao K, Chu WS, Zhang H (2016) Deep Region and Multi-label Learning for Facial Action Unit Detection[C] \/\/ Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2016:3391\u20133399.","DOI":"10.1109\/CVPR.2016.369"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-021-10836-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-021-10836-w\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-021-10836-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,7,5]],"date-time":"2021-07-05T04:11:50Z","timestamp":1625458310000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-021-10836-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,4,4]]},"references-count":28,"journal-issue":{"issue":"16","published-print":{"date-parts":[[2021,7]]}},"alternative-id":["10836"],"URL":"https:\/\/doi.org\/10.1007\/s11042-021-10836-w","relation":{},"ISSN":["1380-7501","1573-7721"],"issn-type":[{"value":"1380-7501","type":"print"},{"value":"1573-7721","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,4,4]]},"assertion":[{"value":"3 June 2019","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 December 2020","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 March 2021","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 April 2021","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}