{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T22:47:11Z","timestamp":1776120431592,"version":"3.50.1"},"reference-count":112,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2020,6,23]],"date-time":"2020-06-23T00:00:00Z","timestamp":1592870400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2020,6,23]],"date-time":"2020-06-23T00:00:00Z","timestamp":1592870400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2021,6]]},"DOI":"10.1007\/s00371-020-01881-x","type":"journal-article","created":{"date-parts":[[2020,6,23]],"date-time":"2020-06-23T13:03:32Z","timestamp":1592917412000},"page":"1447-1466","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":37,"title":["Benchmarking commercial emotion detection systems using realistic distortions of facial image datasets"],"prefix":"10.1007","volume":"37","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-7106-0022","authenticated-orcid":false,"given":"Kangning","family":"Yang","sequence":"first","affiliation":[]},{"given":"Chaofan","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Zhanna","family":"Sarsenbayeva","sequence":"additional","affiliation":[]},{"given":"Benjamin","family":"Tag","sequence":"additional","affiliation":[]},{"given":"Tilman","family":"Dingler","sequence":"additional","affiliation":[]},{"given":"Greg","family":"Wadley","sequence":"additional","affiliation":[]},{"given":"Jorge","family":"Goncalves","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,6,23]]},"reference":[{"key":"1881_CR1","unstructured":"Affectiva: Home\u2014Affectiva : Affectiva. https:\/\/www.affectiva.com\/ (2019)"},{"key":"1881_CR2","doi-asserted-by":"crossref","first-page":"159","DOI":"10.1016\/B978-0-12-800935-2.00008-7","volume-title":"Neuroimaging Personality, Social Cognition, and Character","author":"DN Albohn","year":"2016","unstructured":"Albohn, D.N., Adams Jr., R.B.: Social vision: at the intersection of vision and person perception. In: Cloutier, J., Absher, J.R. (eds.) Neuroimaging Personality, Social Cognition, and Character, pp. 159\u2013186. Elsevier, Amsterdam (2016)"},{"key":"1881_CR3","doi-asserted-by":"crossref","unstructured":"Alm, C.O., Roth, D., Sproat, R.: Emotions from text: machine learning for text-based emotion prediction. In: Proceedings of the Conference on Human Language Technology and Empirical Methods in Natural Language Processing, pp. 579\u2013586. Association for Computational Linguistics (2005)","DOI":"10.3115\/1220575.1220648"},{"key":"1881_CR4","doi-asserted-by":"crossref","unstructured":"AlZoubi, O., Calvo, R.A., Stevens, R.H.: Classification of EEG for affect recognition: an adaptive approach. In: Australasian Joint Conference on Artificial Intelligence, pp. 52\u201361. Springer, Berlin (2009)","DOI":"10.1007\/978-3-642-10439-8_6"},{"key":"1881_CR5","unstructured":"Amazon: Amazon Rekognition-Video and Image-aws. https:\/\/aws.amazon.com\/rekognition\/?nc1=h_ls (2019)"},{"issue":"1","key":"1881_CR6","doi-asserted-by":"crossref","first-page":"87","DOI":"10.1080\/09541449108406221","volume":"3","author":"T Bachmann","year":"1991","unstructured":"Bachmann, T.: Identification of spatially quantised tachistoscopic images of faces: how many pixels does it take to carry identity? Eur. J. Cogn. Psychol. 3(1), 87\u2013103 (1991)","journal-title":"Eur. J. Cogn. Psychol."},{"key":"1881_CR7","unstructured":"Baidu: Baidu ai. https:\/\/ai.baidu.com\/docs#\/Face-Detect-V3\/top (2019)"},{"issue":"1","key":"1881_CR8","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1177\/1529100619832930","volume":"20","author":"LF Barrett","year":"2019","unstructured":"Barrett, L.F., Adolphs, R., Marsella, S., Martinez, A.M., Pollak, S.D.: Emotional expressions reconsidered: challenges to inferring emotion from human facial movements. Psychol. Sci. Public Interest 20(1), 1\u201368 (2019)","journal-title":"Psychol. Sci. Public Interest"},{"key":"1881_CR9","doi-asserted-by":"crossref","unstructured":"Bartlett, M.S., Littlewort, G., Fasel, I., Movellan, J.R.: Real time face detection and facial expression recognition: Development and applications to human computer interaction. In: 2003 Conference on Computer Vision and Pattern Recognition Workshop, vol. 5, pp. 53\u201353. IEEE (2003)","DOI":"10.1109\/CVPRW.2003.10057"},{"key":"1881_CR10","doi-asserted-by":"crossref","unstructured":"Bernin, A., M\u00fcller, L., Ghose, S., von Luck, K., Grecos, C., Wang, Q., Vogt, F.: Towards more robust automatic facial expression recognition in smart environments. In: Proceedings of the 10th International Conference on PErvasive Technologies Related to Assistive Environments, pp. 37\u201344. ACM (2017)","DOI":"10.1145\/3056540.3056546"},{"issue":"11","key":"1881_CR11","doi-asserted-by":"crossref","first-page":"1021","DOI":"10.1007\/s00371-011-0611-x","volume":"27","author":"S Berretti","year":"2011","unstructured":"Berretti, S., Amor, B.B., Daoudi, M., Del Bimbo, A.: 3d facial expression recognition using sift descriptors of automatically detected keypoints. Vis. Comput. 27(11), 1021 (2011)","journal-title":"Vis. Comput."},{"issue":"12","key":"1881_CR12","doi-asserted-by":"crossref","first-page":"1333","DOI":"10.1007\/s00371-013-0869-2","volume":"29","author":"S Berretti","year":"2013","unstructured":"Berretti, S., Del Bimbo, A., Pala, P.: Automatic facial expression recognition in real-time from dynamic sequences of 3d face scans. Vis. Comput. 29(12), 1333\u20131350 (2013)","journal-title":"Vis. Comput."},{"key":"1881_CR13","unstructured":"Bettadapura, V.: Face expression recognition and analysis: the state of the art. arXiv preprint arXiv:1203.6722 (2012)"},{"key":"1881_CR14","unstructured":"Bourel, F., Chibelushi, C.C., Low, A.A.: Robust facial expression recognition using a state-based model of spatially-localised facial dynamics. In: Proceedings of Fifth IEEE International Conference on Automatic Face Gesture Recognition, pp. 113\u2013118. IEEE (2002)"},{"key":"1881_CR15","doi-asserted-by":"crossref","unstructured":"Boyat, A.K., Joshi, B.K.: A review paper: noise models in digital image processing. arXiv preprint arXiv:1505.03489 (2015)","DOI":"10.5121\/sipij.2015.6206"},{"key":"1881_CR16","doi-asserted-by":"crossref","unstructured":"Br\u00f6hl, C., Mertens, A., Ziefle, M.: How do users interact with mobile devices? an analysis of handheld positions for different technology generations. In: International Conference on Human Aspects of IT for the Aged Population, pp. 3\u201316. Springer, Berlin (2017)","DOI":"10.1007\/978-3-319-58536-9_1"},{"key":"1881_CR17","doi-asserted-by":"crossref","unstructured":"Bryant, D., Howard, A.: A comparative analysis of emotion-detecting AI systems with respect to algorithm performance and dataset diversity. In: Proceedings of the 2019 AAAI\/ACM Conference on AI, Ethics, and Society, pp. 377\u2013382. ACM (2019)","DOI":"10.1145\/3306618.3314284"},{"issue":"1","key":"1881_CR18","doi-asserted-by":"crossref","first-page":"18","DOI":"10.1109\/T-AFFC.2010.1","volume":"1","author":"RA Calvo","year":"2010","unstructured":"Calvo, R.A., D\u2019Mello, S.: Affect detection: an interdisciplinary review of models, methods, and their applications. IEEE Trans. Affect. Comput. 1(1), 18\u201337 (2010)","journal-title":"IEEE Trans. Affect. Comput."},{"issue":"1","key":"1881_CR19","doi-asserted-by":"crossref","first-page":"314","DOI":"10.3390\/app10010314","volume":"10","author":"E Carlotta Olivetti","year":"2020","unstructured":"Carlotta Olivetti, E., Violante, M.G., Vezzetti, E., Marcolin, F., Eynard, B.: Engagement evaluation in a virtual learning environment via facial expression recognition and self-reports: a preliminary approach. Appl. Sci. 10(1), 314 (2020)","journal-title":"Appl. Sci."},{"issue":"2","key":"1881_CR20","doi-asserted-by":"crossref","first-page":"205","DOI":"10.1037\/0022-3514.70.2.205","volume":"70","author":"JM Carroll","year":"1996","unstructured":"Carroll, J.M., Russell, J.A.: Do facial expressions signal specific emotions? judging emotion from the face in context. J. Personal. Soc. Psychol. 70(2), 205 (1996)","journal-title":"J. Personal. Soc. Psychol."},{"key":"1881_CR21","doi-asserted-by":"crossref","unstructured":"Celma, O.: Music recommendation. In: Gerstner, R. (ed.) Music Recommendation and Discovery, pp. 43\u201385. Springer, Berlin (2010)","DOI":"10.1007\/978-3-642-13287-2_3"},{"key":"1881_CR22","doi-asserted-by":"crossref","unstructured":"Chang, W.Y., Hsu, S.H., Chien, J.H.: Fatauva-net: an integrated deep learning framework for facial attribute recognition, action unit detection, and valence-arousal estimation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 17\u201325 (2017)","DOI":"10.1109\/CVPRW.2017.246"},{"issue":"3","key":"1881_CR23","doi-asserted-by":"crossref","first-page":"619","DOI":"10.1016\/j.psychres.2014.11.035","volume":"225","author":"Y Chen","year":"2015","unstructured":"Chen, Y., McBain, R., Norton, D.: Specific vulnerability of face perception to noise: a similar effect in schizophrenia patients and healthy individuals. Psychiatry Res. 225(3), 619\u2013624 (2015)","journal-title":"Psychiatry Res."},{"key":"1881_CR24","doi-asserted-by":"crossref","unstructured":"Cheng, Y., Jiang, B., Jia, K.: A deep structure for facial expression recognition under partial occlusion. In: 2014 Tenth International Conference on Intelligent Information Hiding and Multimedia Signal Processing, pp. 211\u2013214. IEEE (2014)","DOI":"10.1109\/IIH-MSP.2014.59"},{"issue":"2","key":"1881_CR25","doi-asserted-by":"crossref","first-page":"117","DOI":"10.1023\/B:JONB.0000023655.25550.be","volume":"28","author":"M Coulson","year":"2004","unstructured":"Coulson, M.: Attributing emotion to static body postures: recognition accuracy, confusions, and viewpoint dependence. J. Nonverbal Behav. 28(2), 117\u2013139 (2004)","journal-title":"J. Nonverbal Behav."},{"key":"1881_CR26","doi-asserted-by":"crossref","DOI":"10.1037\/10001-000","volume-title":"The Expression of the Emotions in Man and Animals","author":"C Darwin","year":"1872","unstructured":"Darwin, C.: The Expression of the Emotions in Man and Animals, Anniversary edn. Harper Perennial, London (1872). (P. Ekman, ed)","edition":"Anniversary"},{"key":"1881_CR27","doi-asserted-by":"crossref","unstructured":"Dodge, S., Karam, L.: Understanding how image quality affects deep neural networks. In: 2016 Eighth International Conference on Quality of Multimedia Experience (QoMEX), pp. 1\u20136. IEEE (2016)","DOI":"10.1109\/QoMEX.2016.7498955"},{"key":"1881_CR28","doi-asserted-by":"crossref","unstructured":"Dodge, S., Karam, L.: Can the early human visual system compete with deep neural networks? In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2798\u20132804 (2017)","DOI":"10.1109\/ICCVW.2017.329"},{"key":"1881_CR29","doi-asserted-by":"crossref","unstructured":"Dodge, S., Karam, L.: A study and comparison of human and deep learning recognition performance under visual distortions. In: 2017 26th International Conference on Computer Communication and Networks (ICCCN), pp. 1\u20137. IEEE (2017)","DOI":"10.1109\/ICCCN.2017.8038465"},{"key":"1881_CR30","doi-asserted-by":"crossref","unstructured":"Dupr\u00e9, D., Andelic, N., Morrison, G., McKeown, G.: Accuracy of three commercial automatic emotion recognition systems across different individuals and their facial expressions. In: 2018 IEEE International Conference on Pervasive Computing and Communications Workshops (PerCom Workshops), pp. 627\u2013632. IEEE (2018)","DOI":"10.1109\/PERCOMW.2018.8480127"},{"key":"1881_CR31","first-page":"45","volume-title":"Handbook of Methods in Nonverbal Behavior Research","author":"P Ekman","year":"1982","unstructured":"Ekman, P.: Methods for measuring facial action. In: Scherer, K.R., Ekman, P. (eds.) Handbook of Methods in Nonverbal Behavior Research, pp. 45\u201390. Cambridge University Press, Cambridge (1982)"},{"key":"1881_CR32","first-page":"19","volume":"3","author":"P Ekman","year":"1984","unstructured":"Ekman, P.: Expression and the nature of emotion. Approaches Emot. 3, 19\u2013344 (1984)","journal-title":"Approaches Emot."},{"issue":"3\u20134","key":"1881_CR33","doi-asserted-by":"crossref","first-page":"169","DOI":"10.1080\/02699939208411068","volume":"6","author":"P Ekman","year":"1992","unstructured":"Ekman, P.: An argument for basic emotions. Cogn. Emot. 6(3\u20134), 169\u2013200 (1992)","journal-title":"Cogn. Emot."},{"key":"1881_CR34","doi-asserted-by":"crossref","first-page":"45","DOI":"10.1002\/0470013494.ch3","volume-title":"Handbook of Cognition and Emotion","author":"P Ekman","year":"1999","unstructured":"Ekman, P.: Basic emotions. In: Dalgleish, T., Power, M.J. (eds.) Handbook of Cognition and Emotion, vol. 98, pp. 45\u201360. Wiley, New York (1999)"},{"issue":"2","key":"1881_CR35","doi-asserted-by":"crossref","first-page":"124","DOI":"10.1037\/h0030377","volume":"17","author":"P Ekman","year":"1971","unstructured":"Ekman, P., Friesen, W.V.: Constants across cultures in the face and emotion. J. Personal. Soc. Psychol. 17(2), 124 (1971)","journal-title":"J. Personal. Soc. Psychol."},{"key":"1881_CR36","volume-title":"Unmasking the Face: A Guide to Recognizing Emotions from Facial Clues","author":"P Ekman","year":"2003","unstructured":"Ekman, P., Friesen, W.V.: Unmasking the Face: A Guide to Recognizing Emotions from Facial Clues. ISHK, Los Altos (2003)"},{"key":"1881_CR37","volume-title":"Emotion in the Human Face: Guide-Lines for Research and an Integration of Findings: Guidelines for Research and an Integration of Findings","author":"P Ekman","year":"1972","unstructured":"Ekman, P., Friesen, W.V., Ellsworth, P.: Emotion in the Human Face: Guide-Lines for Research and an Integration of Findings: Guidelines for Research and an Integration of Findings. Pergamon, Oxford (1972)"},{"key":"1881_CR38","first-page":"77","volume-title":"Facial Action Coding System: The Manual on CD ROM","author":"P Ekman","year":"2002","unstructured":"Ekman, P., Friesen, W.V., Hager, J.C.: Facial Action Coding System: The Manual on CD ROM, pp. 77\u2013254. A Human Face, Salt Lake City (2002)"},{"issue":"4","key":"1881_CR39","doi-asserted-by":"crossref","first-page":"712","DOI":"10.1037\/0022-3514.53.4.712","volume":"53","author":"P Ekman","year":"1987","unstructured":"Ekman, P., Friesen, W.V., O\u2019sullivan, M., Chan, A., Diacoyanni-Tarlatzis, I., Heider, K., Krause, R., LeCompte, W.A., Pitcairn, T., Ricci-Bitti, P.E., et al.: Universals and cultural differences in the judgments of facial expressions of emotion. J. Personal. Soc. Psychol. 53(4), 712 (1987)","journal-title":"J. Personal. Soc. Psychol."},{"issue":"3","key":"1881_CR40","doi-asserted-by":"crossref","first-page":"572","DOI":"10.1016\/j.patcog.2010.09.020","volume":"44","author":"M El Ayadi","year":"2011","unstructured":"El Ayadi, M., Kamel, M.S., Karray, F.: Survey on speech emotion recognition: features, classification schemes, and databases. Pattern Recognit. 44(3), 572\u2013587 (2011)","journal-title":"Pattern Recognit."},{"issue":"1","key":"1881_CR41","doi-asserted-by":"crossref","first-page":"3","DOI":"10.1016\/0301-0511(91)90003-Y","volume":"32","author":"AJ Fridlund","year":"1991","unstructured":"Fridlund, A.J.: Evolution and facial action in reflex, social motive, and paralanguage. Biol. Psychol. 32(1), 3\u2013100 (1991)","journal-title":"Biol. Psychol."},{"key":"1881_CR42","volume-title":"Facial Action Coding System: A Technique for the Measurement of Facial Movement","author":"E Friesen","year":"1978","unstructured":"Friesen, E., Ekman, P.: Facial Action Coding System: A Technique for the Measurement of Facial Movement, vol. 3. Consulting Psychologists Press, Palo Alto (1978)"},{"key":"1881_CR43","unstructured":"Friesen, W.V., Ekman, P., et al.: Emfacs-7: emotional facial action coding system, vol. 2, no. 36, p. 1. Unpublished manuscript, University of California at San Francisco (1983)"},{"key":"1881_CR44","unstructured":"Gedraite, E.S., Hadad, M.: Investigation on the effect of a Gaussian blur in image filtering and segmentation. In: Proceedings ELMAR-2011, pp. 393\u2013396. IEEE (2011)"},{"key":"1881_CR45","volume-title":"Behavioral Medicine","author":"MD Gellman","year":"2013","unstructured":"Gellman, M.D.: Behavioral Medicine. Springer, Berlin (2013)"},{"issue":"6","key":"1881_CR46","doi-asserted-by":"crossref","first-page":"1094","DOI":"10.1080\/02699930701626582","volume":"22","author":"E Goeleven","year":"2008","unstructured":"Goeleven, E., De Raedt, R., Leyman, L., Verschuere, B.: The Karolinska directed emotional faces: a validation study. Cogn. Emot. 22(6), 1094\u20131118 (2008)","journal-title":"Cogn. Emot."},{"key":"1881_CR47","doi-asserted-by":"crossref","unstructured":"Goncalves, J., Pandab, P., Ferreira, D., Ghahramani, M., Zhao, G., Kostakos, V.: Projective testing of diurnal collective emotion. In: Proceedings of the 2014 ACM International Joint Conference on Pervasive and Ubiquitous Computing, UbiComp\u2019 14, pp. 487\u2013497. New York, NY, USA (2014)","DOI":"10.1145\/2632048.2636067"},{"key":"1881_CR48","doi-asserted-by":"crossref","unstructured":"Gong, B., Wang, Y., Liu, J., Tang, X.: Automatic facial expression recognition on a single 3d face by exploring shape deformation. In: Proceedings of the 17th ACM International Conference on Multimedia, pp. 569\u2013572 (2009)","DOI":"10.1145\/1631272.1631358"},{"key":"1881_CR49","unstructured":"Google: Vision AI-derive image insights via ml-cloud vision api-google cloud. https:\/\/cloud.google.com\/vision\/ (2019)"},{"issue":"5","key":"1881_CR50","doi-asserted-by":"crossref","first-page":"551","DOI":"10.1080\/026999399379186","volume":"13","author":"JJ Gross","year":"1999","unstructured":"Gross, J.J.: Emotion regulation: past, present, future. Cogn. Emot. 13(5), 551\u2013573 (1999)","journal-title":"Cogn. Emot."},{"key":"1881_CR51","doi-asserted-by":"crossref","unstructured":"Gu, Y., Li, X., Huang, K., Fu, S., Yang, K., Chen, S., Zhou, M., Marsic, I.: Human conversation analysis using attentive multimodal networks with hierarchical encoder\u2013decoder. In: 2018 ACM Multimedia Conference on Multimedia Conference, pp. 537\u2013545. ACM (2018)","DOI":"10.1145\/3240508.3240714"},{"key":"1881_CR52","unstructured":"Gu, Y., Yang, K., Fu, S., Chen, S., Li, X., Marsic, I.: Hybrid attention based multimodal network for spoken language classification. In: Proceedings of the Conference. Association for Computational Linguistics. Meeting, vol. 2018, pp. 2379\u20132390. NIH Public Access (2018)"},{"key":"1881_CR53","doi-asserted-by":"crossref","unstructured":"Gu, Y., Yang, K., Fu, S., Chen, S., Li, X., Marsic, I.: Multimodal affective analysis using hierarchical attention strategy with word-level alignment. arXiv preprint arXiv:1805.08660 (2018)","DOI":"10.18653\/v1\/P18-1207"},{"issue":"3","key":"1881_CR54","first-page":"187","volume":"2","author":"A Heraz","year":"2007","unstructured":"Heraz, A., Frasson, C.: Predicting the three major dimensions of the learner\u2019s emotions from brainwaves. Int. J. Comput. Sci. 2(3), 187\u2013193 (2007)","journal-title":"Int. J. Comput. Sci."},{"issue":"4","key":"1881_CR55","doi-asserted-by":"crossref","first-page":"2213","DOI":"10.1137\/120888302","volume":"6","author":"L Hou","year":"2013","unstructured":"Hou, L., Ji, H., Shen, Z.: Recovering over-\/underexposed regions in photographs. SIAM J. Imaging Sci. 6(4), 2213\u20132235 (2013)","journal-title":"SIAM J. Imaging Sci."},{"key":"1881_CR56","doi-asserted-by":"crossref","unstructured":"Howard, A., Zhang, C., Horvitz, E.: Addressing bias in machine learning algorithms: a pilot study on emotion recognition for intelligent systems. In: 2017 IEEE Workshop on Advanced Robotics and Its Social Impacts (ARSO), pp. 1\u20137. IEEE (2017)","DOI":"10.1109\/ARSO.2017.8025197"},{"key":"1881_CR57","doi-asserted-by":"crossref","unstructured":"Huang, D., De la Torre, F.: Bilinear kernel reduced rank regression for facial expression synthesis. In: European Conference on Computer Vision, pp. 364\u2013377. Springer, Berlin (2010)","DOI":"10.1007\/978-3-642-15552-9_27"},{"key":"1881_CR58","volume-title":"The Face of Emotion","author":"CE Izard","year":"1971","unstructured":"Izard, C.E.: The Face of Emotion. Appleton-Century Crofts, New York (1971)"},{"issue":"19","key":"1881_CR59","doi-asserted-by":"crossref","first-page":"7241","DOI":"10.1073\/pnas.1200155109","volume":"109","author":"RE Jack","year":"2012","unstructured":"Jack, R.E., Garrod, O.G., Yu, H., Caldara, R., Schyns, P.G.: Facial expressions of emotion are not culturally universal. Proc. Natl. Acad. Sci. 109(19), 7241\u20137244 (2012)","journal-title":"Proc. Natl. Acad. Sci."},{"key":"1881_CR60","volume-title":"Facial Expression of Emotion","author":"D Keltner","year":"2000","unstructured":"Keltner, D., Ekman, P., Gonzaga, G., Beer, J.: Facial Expression of Emotion. Guilford Publications, New York (2000)"},{"key":"1881_CR61","doi-asserted-by":"crossref","unstructured":"Khamis, M., Baier, A., Henze, N., Alt, F., Bulling, A.: Understanding face and eye visibility in front-facing cameras of smartphones used in the wild. In: Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems, p. 280. ACM (2018)","DOI":"10.1145\/3173574.3173854"},{"key":"1881_CR62","doi-asserted-by":"crossref","first-page":"32672","DOI":"10.1038\/srep32672","volume":"6","author":"SR Kheradpisheh","year":"2016","unstructured":"Kheradpisheh, S.R., Ghodrati, M., Ganjtabesh, M., Masquelier, T.: Deep networks can resemble human feed-forward vision in invariant object recognition. Sci. Rep. 6, 32672 (2016)","journal-title":"Sci. Rep."},{"key":"1881_CR63","doi-asserted-by":"crossref","unstructured":"Kim, Y., Lee, H., Provost, E.M.: Deep learning for robust feature generation in audiovisual emotion recognition. In: 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 3687\u20133691. IEEE (2013)","DOI":"10.1109\/ICASSP.2013.6638346"},{"key":"1881_CR64","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: Advances in Neural Information Processing Systems, pp. 1097\u20131105 (2012)"},{"issue":"8","key":"1881_CR65","doi-asserted-by":"crossref","first-page":"1377","DOI":"10.1080\/02699930903485076","volume":"24","author":"O Langner","year":"2010","unstructured":"Langner, O., Dotsch, R., Bijlstra, G., Wigboldus, D.H., Hawk, S.T., Van Knippenberg, A.: Presentation and validation of the radboud faces database. Cogn. Emot. 24(8), 1377\u20131388 (2010)","journal-title":"Cogn. Emot."},{"key":"1881_CR66","doi-asserted-by":"crossref","unstructured":"Le, H.V., Mayer, S., Wolf, K., Henze, N.: Finger placement and hand grasp during smartphone interaction. In: Proceedings of the 2016 CHI Conference Extended Abstracts on Human Factors in Computing Systems, pp. 2576\u20132584. ACM (2016)","DOI":"10.1145\/2851581.2892462"},{"issue":"4","key":"1881_CR67","doi-asserted-by":"crossref","first-page":"227","DOI":"10.1037\/npe0000028","volume":"7","author":"P Lewinski","year":"2014","unstructured":"Lewinski, P., den Uyl, T.M., Butler, C.: Automated facial coding: validation of basic emotions and FACS AUs in facereader. J. Neurosci. Psychol. Econ. 7(4), 227 (2014)","journal-title":"J. Neurosci. Psychol. Econ."},{"key":"1881_CR68","volume-title":"Handbook of Emotions","author":"M Lewis","year":"2010","unstructured":"Lewis, M., Haviland-Jones, J.M., Barrett, L.F.: Handbook of Emotions. Guilford Press, New York (2010)"},{"key":"1881_CR69","doi-asserted-by":"crossref","unstructured":"Litman, D.J., Forbes-Riley, K.: Predicting student emotions in computer-human tutoring dialogues. In: Proceedings of the 42nd Annual Meeting on Association for Computational Linguistics, p. 351. Association for Computational Linguistics (2004)","DOI":"10.3115\/1218955.1219000"},{"key":"1881_CR70","doi-asserted-by":"crossref","unstructured":"Liu, C., Freeman, W.T., Szeliski, R., Kang, S.B.: Noise estimation from a single image. In: 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR\u201906), vol.\u00a01, pp. 901\u2013908. IEEE (2006)","DOI":"10.1109\/CVPR.2006.207"},{"key":"1881_CR71","doi-asserted-by":"crossref","unstructured":"Ma, C., Osherenko, A., Prendinger, H., Ishizuka, M.: A chat system based on emotion estimation from text and embodied conversational messengers. In: Proceedings of the 2005 International Conference on Active Media Technology, 2005. (AMT 2005), pp. 546\u2013548. IEEE (2005)","DOI":"10.1007\/11558651_56"},{"key":"1881_CR72","doi-asserted-by":"crossref","unstructured":"Maalej, A., Amor, B.B., Daoudi, M., Srivastava, A., Berretti, S.: Local 3d shape analysis for facial expression recognition. In: 2010 20th International Conference on Pattern Recognition, pp. 4129\u20134132. IEEE (2010)","DOI":"10.1109\/ICPR.2010.1003"},{"key":"1881_CR73","doi-asserted-by":"crossref","unstructured":"Mao, X., Xue, Y., Li, Z., Huang, K., Lv, S.: Robust facial expression recognition based on RPCA and AdaBoost. In: 2009 10th Workshop on Image Analysis for Multimedia Interactive Services, pp. 113\u2013116. IEEE (2009)","DOI":"10.1109\/WIAMIS.2009.5031445"},{"issue":"5","key":"1881_CR74","doi-asserted-by":"crossref","first-page":"4237","DOI":"10.4249\/scholarpedia.4237","volume":"3","author":"D Matsumoto","year":"2008","unstructured":"Matsumoto, D., Ekman, P.: Facial expression analysis. Scholarpedia 3(5), 4237 (2008)","journal-title":"Scholarpedia"},{"key":"1881_CR75","first-page":"211","volume-title":"Handbook of Emotions","author":"D Matsumoto","year":"2008","unstructured":"Matsumoto, D., Keltner, D., Shiota, M.N., O\u2019Sullivan, M., Frank, M.: Facial expressions of emotion. In: Lewis, M., Haviland-Jones, J.M., Barrett, L.F. (eds.) Handbook of Emotions, vol. 3, pp. 211\u2013234. Guilford Press, New York (2008)"},{"key":"1881_CR76","doi-asserted-by":"crossref","unstructured":"Matthews, O., Sarsenbayeva, Z., Jiang, W., Newn, J., Velloso, E., Clinch, S., Goncalves, J.: Inferring the mood of a community from their walking speed: a preliminary study. In: Proceedings of the 2018 ACM International Joint Conference and 2018 International Symposium on Pervasive and Ubiquitous Computing, UbiComp\u2019 18, pp. 1144\u20131149 (2018)","DOI":"10.1145\/3267305.3274759"},{"key":"1881_CR77","unstructured":"McDaniel, B., D\u2019Mello, S., King, B., Chipman, P., Tapp, K., Graesser, A.: Facial features for affective state detection in learning environments. In: Proceedings of the Annual Meeting of the Cognitive Science Society, vol. 29 (2007)"},{"key":"1881_CR78","unstructured":"Microsoft: Face API\u2014facial recognition software\u2013microsoft azure. https:\/\/azure.microsoft.com\/en-au\/services\/cognitive-services\/face\/ (2019)"},{"key":"1881_CR79","doi-asserted-by":"crossref","unstructured":"Narwekar, A.A., Girju, R.: Uiuc at semeval-2018 task 1: recognizing affect with ensemble models. In: Proceedings of The 12th International Workshop on Semantic Evaluation, pp. 377\u2013384 (2018)","DOI":"10.18653\/v1\/S18-1057"},{"issue":"1","key":"1881_CR80","doi-asserted-by":"crossref","first-page":"8","DOI":"10.1177\/1754073912457227","volume":"5","author":"NL Nelson","year":"2013","unstructured":"Nelson, N.L., Russell, J.A.: Universality revisited. Emot. Rev. 5(1), 8\u201315 (2013)","journal-title":"Emot. Rev."},{"key":"1881_CR81","doi-asserted-by":"crossref","first-page":"1516","DOI":"10.3389\/fpsyg.2014.01516","volume":"5","author":"M Olszanowski","year":"2015","unstructured":"Olszanowski, M., Pochwatko, G., Kuklinski, K., Scibor-Rylski, M., Lewinski, P., Ohme, R.K.: Warsaw set of emotional facial expression pictures: a validation study of facial display photographs. Front. Psychol. 5, 1516 (2015)","journal-title":"Front. Psychol."},{"key":"1881_CR82","unstructured":"Opencv: Cascade Classifier Training. https:\/\/docs.opencv.org\/3.1.0\/dc\/d88\/tutorial_traincascade.html#gsc.tab=0"},{"key":"1881_CR83","doi-asserted-by":"crossref","unstructured":"Panigrahi, S.K., Gupta, S., Sahu, P.K.: Phases under Gaussian additive noise. In: 2016 International Conference on Communication and Signal Processing (ICCSP), pp. 1771\u20131776. IEEE (2016)","DOI":"10.1109\/ICCSP.2016.7754471"},{"key":"1881_CR84","volume-title":"Software Testing","author":"R Patton","year":"2006","unstructured":"Patton, R.: Software Testing. Pearson Education India, New Delhi (2006)"},{"key":"1881_CR85","doi-asserted-by":"crossref","first-page":"98","DOI":"10.1016\/j.inffus.2017.02.003","volume":"37","author":"S Poria","year":"2017","unstructured":"Poria, S., Cambria, E., Bajpai, R., Hussain, A.: A review of affective computing: from unimodal analysis to multimodal fusion. Inf. Fusion 37, 98\u2013125 (2017)","journal-title":"Inf. Fusion"},{"key":"1881_CR86","doi-asserted-by":"crossref","unstructured":"Poria, S., Cambria, E., Gelbukh, A.: Deep convolutional neural network textual features and multiple kernel learning for utterance-level multimodal sentiment analysis. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp. 2539\u20132544 (2015)","DOI":"10.18653\/v1\/D15-1303"},{"key":"1881_CR87","doi-asserted-by":"crossref","unstructured":"Poria, S., Chaturvedi, I., Cambria, E., Hussain, A.: Convolutional MKL based multimodal emotion recognition and sentiment analysis. In: 2016 IEEE 16th International Conference on Data Mining (ICDM), pp. 439\u2013448. IEEE (2016)","DOI":"10.1109\/ICDM.2016.0055"},{"key":"1881_CR88","doi-asserted-by":"crossref","unstructured":"Rodner, E., Simon, M., Fisher, R.B., Denzler, J.: Fine-grained recognition in the noisy wild: sensitivity analysis of convolutional neural networks approaches. arXiv preprint arXiv:1610.06756 (2016)","DOI":"10.5244\/C.30.60"},{"key":"1881_CR89","doi-asserted-by":"publisher","unstructured":"Rodriguez, P., Cucurull, G., Gonz\u00e0lez, J., Gonfaus, J.M., Nasrollahi, K., Moeslund, T.B., Roca, F.X.: Deep pain: Exploiting long short-term memory networks for facial expression classification. IEEE Trans. Cybern. 2017. https:\/\/doi.org\/10.1109\/TCYB.2017.2662199","DOI":"10.1109\/TCYB.2017.2662199"},{"issue":"1","key":"1881_CR90","doi-asserted-by":"crossref","first-page":"102","DOI":"10.1037\/0033-2909.115.1.102","volume":"115","author":"JA Russell","year":"1994","unstructured":"Russell, J.A.: Is there universal recognition of emotion from facial expression? a review of the cross-cultural studies. Psychol. Bull. 115(1), 102 (1994)","journal-title":"Psychol. Bull."},{"key":"1881_CR91","volume-title":"Oxford Companion to Emotion and the Affective Sciences","author":"D Sander","year":"2014","unstructured":"Sander, D., Scherer, K.: Oxford Companion to Emotion and the Affective Sciences. Oxford University Press, Oxford (2014)"},{"key":"1881_CR92","doi-asserted-by":"crossref","unstructured":"Sarsenbayeva, Z., Ferreira, D., van Berkel, N., Luo, C., Vaisanen, M., Kostakos, V., Goncalves, J.: Vision-based happiness inference: a feasibility case-study. In: Proceedings of the 2017 ACM International Joint Conference on Pervasive and Ubiquitous Computing, UbiComp \u201917, pp. 494\u2013499. ACM, New York, NY, USA (2017)","DOI":"10.1145\/3123024.3124438"},{"key":"1881_CR93","doi-asserted-by":"crossref","unstructured":"Sarsenbayeva, Z., Marini, G., van Berkel, N., Luo, C., Jiang, W., Yang, K., Wadley, G., Dingler, T., Kostakos, V., Goncalves, J.: Does smartphone use drive our emotions or vice versa? a causal analysis. In: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, CHI\u2019 20, pp. 1\u201315. New York, NY, USA (2020)","DOI":"10.1145\/3313831.3376163"},{"key":"1881_CR94","doi-asserted-by":"crossref","unstructured":"Schuller, B., Stadermann, J., Rigoll, G.: Affect-robust speech recognition by dynamic emotional adaptation. In: Proceedings of Speech Prosody 2006, Dresden (2006)","DOI":"10.21437\/SpeechProsody.2006-221"},{"key":"1881_CR95","doi-asserted-by":"crossref","unstructured":"Sharma, P., Eseng\u00f6n\u00fcl, M., Khanal, S.R., Khanal, T.T., Filipe, V., Reis, M.J.: Student concentration evaluation index in an e-learning context using facial emotion analysis. In: International Conference on Technology and Innovation in Learning, Teaching and Education, pp. 529\u2013538. Springer, Berlin (2018)","DOI":"10.1007\/978-3-030-20954-4_40"},{"issue":"4","key":"1881_CR96","doi-asserted-by":"crossref","first-page":"1446","DOI":"10.3758\/s13428-017-0996-1","volume":"50","author":"S St\u00f6ckli","year":"2018","unstructured":"St\u00f6ckli, S., Schulte-Mecklenbeck, M., Borer, S., Samson, A.C.: Facial expression analysis with affdex and facet: a validation study. Behav. Res. Methods 50(4), 1446\u20131460 (2018)","journal-title":"Behav. Res. Methods"},{"key":"1881_CR97","unstructured":"Swinton, R., El\u00a0Kaliouby, R.: Measuring emotions through a mobile device across borders, ages, genders and more. In: Proceedings of the ESOMAR Congress, Atlanta, pp. 1\u201312 (2012)"},{"issue":"6","key":"1881_CR98","doi-asserted-by":"crossref","first-page":"1635","DOI":"10.1109\/TIP.2010.2042645","volume":"19","author":"X Tan","year":"2010","unstructured":"Tan, X., Triggs, W.: Enhanced local texture feature sets for face recognition under difficult lighting conditions. IEEE Trans. Image Process. 19(6), 1635\u20131650 (2010)","journal-title":"IEEE Trans. Image Process."},{"key":"1881_CR99","unstructured":"Technology, M.: Emotion recognition\u2014face++ AI open platform. https:\/\/www.faceplusplus.com\/emotion-recognition\/ (2019)"},{"issue":"11","key":"1881_CR100","doi-asserted-by":"crossref","first-page":"1958","DOI":"10.1109\/TPAMI.2008.128","volume":"30","author":"A Torralba","year":"2008","unstructured":"Torralba, A., Fergus, R., Freeman, W.T.: 80 million tiny images: a large data set for nonparametric object and scene recognition. IEEE Trans. Pattern Anal. Mach. Intell. 30(11), 1958\u20131970 (2008)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1881_CR101","doi-asserted-by":"crossref","unstructured":"Towner, H., Slater, M.: Reconstruction and recognition of occluded facial expressions using PCA. In: International Conference on Affective Computing and Intelligent Interaction, pp. 36\u201347. Springer, Berlin (2007)","DOI":"10.1007\/978-3-540-74889-2_4"},{"key":"1881_CR102","unstructured":"Useche, O., El-Sheikh, E.: An intelligent system framework for measuring attention levels of students in online course environments. In: Proceedings on the International Conference on Artificial Intelligence (ICAI), p. 452. The Steering Committee of The World Congress in Computer Science, Computer... (2015)"},{"key":"1881_CR103","doi-asserted-by":"crossref","unstructured":"Valstar, M.F., Jiang, B., Mehu, M., Pantic, M., Scherer, K.: The first facial expression recognition and analysis challenge. In: Face and Gesture 2011, pp. 921\u2013926. IEEE (2011)","DOI":"10.1109\/FG.2011.5771374"},{"key":"1881_CR104","doi-asserted-by":"crossref","unstructured":"van Berkel, N., Goncalves, J., Koval, P., Hosio, S., Dingler, T., Ferreira, D., Kostakos, V.: Context-informed scheduling and analysis: improving accuracy of mobile self-reports. In: Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, CHI\u2019 19 (2019)","DOI":"10.1145\/3290605.3300281"},{"key":"1881_CR105","doi-asserted-by":"crossref","unstructured":"van\u00a0der Haar, D.T.: Student emotion recognition in computer science education: a blessing or curse? In: International Conference on Human\u2013Computer Interaction, pp. 301\u2013311. Springer, Berlin (2019)","DOI":"10.1007\/978-3-030-21814-0_23"},{"issue":"4","key":"1881_CR106","doi-asserted-by":"crossref","first-page":"907","DOI":"10.1037\/a0023853","volume":"11","author":"J Van Der Schalk","year":"2011","unstructured":"Van Der Schalk, J., Hawk, S.T., Fischer, A.H., Doosje, B.: Moving faces, looking places: validation of the Amsterdam dynamic facial expression set (ADFES). Emotion 11(4), 907 (2011)","journal-title":"Emotion"},{"issue":"11","key":"1881_CR107","doi-asserted-by":"crossref","first-page":"2218","DOI":"10.3390\/app9112218","volume":"9","author":"MG Violante","year":"2019","unstructured":"Violante, M.G., Marcolin, F., Vezzetti, E., Ulrich, L., Billia, G., Di Grazia, L.: 3d facial expression recognition for defining users\u2019 inner requirements-an emotional design case study. Appl. Sci. 9(11), 2218 (2019)","journal-title":"Appl. Sci."},{"key":"1881_CR108","doi-asserted-by":"crossref","unstructured":"Visuri, A., Sarsenbayeva, Z., Goncalves, J., Karapanos, E., Jones, S.: Impact of mood changes on application selection. In: Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing: Adjunct, pp. 535\u2013540. ACM (2016)","DOI":"10.1145\/2968219.2968317"},{"issue":"3","key":"1881_CR109","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/1276377.1276379","volume":"26","author":"L Yuan","year":"2007","unstructured":"Yuan, L., Sun, J., Quan, L., Shum, H.Y.: Image deblurring with blurred\/noisy image pairs. ACM Trans. Graph. (TOG) 26(3), 1 (2007)","journal-title":"ACM Trans. Graph. (TOG)"},{"issue":"1","key":"1881_CR110","doi-asserted-by":"crossref","first-page":"39","DOI":"10.1109\/TPAMI.2008.52","volume":"31","author":"Z Zeng","year":"2008","unstructured":"Zeng, Z., Pantic, M., Roisman, G.I., Huang, T.S.: A survey of affect recognition methods: audio, visual, and spontaneous expressions. IEEE Trans. Pattern Anal. Mach. Intell. 31(1), 39\u201358 (2008)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"2","key":"1881_CR111","doi-asserted-by":"crossref","first-page":"25","DOI":"10.1145\/3158369","volume":"51","author":"L Zhang","year":"2018","unstructured":"Zhang, L., Verma, B., Tjondronegoro, D., Chandran, V.: Facial expression analysis under partial occlusion: a survey. ACM Comput. Surv. (CSUR) 51(2), 25 (2018)","journal-title":"ACM Comput. Surv. (CSUR)"},{"key":"1881_CR112","doi-asserted-by":"crossref","first-page":"30","DOI":"10.1016\/j.future.2015.12.001","volume":"66","author":"Y Zhang","year":"2017","unstructured":"Zhang, Y., Chen, M., Huang, D., Wu, D., Li, Y.: idoctor: personalized and professionalized medical recommendations based on hybrid matrix factorization. Future Gener. Comput. Syst. 66, 30\u201335 (2017)","journal-title":"Future Gener. Comput. Syst."}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-020-01881-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-020-01881-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-020-01881-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,8,8]],"date-time":"2024-08-08T06:51:34Z","timestamp":1723099894000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-020-01881-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,6,23]]},"references-count":112,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2021,6]]}},"alternative-id":["1881"],"URL":"https:\/\/doi.org\/10.1007\/s00371-020-01881-x","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,6,23]]},"assertion":[{"value":"23 June 2020","order":1,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Compliance with ethical standards"}},{"value":"Authors, Kangning Yang, Chaofan Wang, Zhanna Sarsenbayeva, Benjamin Tag, Tilman Dingler, Greg Wadley, and Jorge Goncalves, declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}