{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,2]],"date-time":"2025-10-02T16:23:00Z","timestamp":1759422180906,"version":"3.37.0"},"publisher-location":"London","reference-count":77,"publisher":"Springer London","isbn-type":[{"type":"print","value":"9781848003057"},{"type":"electronic","value":"9781848003064"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"DOI":"10.1007\/978-1-84800-306-4_14","type":"book-chapter","created":{"date-parts":[[2008,12,1]],"date-time":"2008-12-01T22:01:16Z","timestamp":1228168876000},"page":"241-265","source":"Crossref","is-referenced-by-count":19,"title":["Emotion Recognition Based on Multimodal Information"],"prefix":"10.1007","author":[{"given":"Zhihong","family":"Zeng","sequence":"first","affiliation":[]},{"given":"Maja","family":"Pantic","sequence":"additional","affiliation":[]},{"given":"Thomas S.","family":"Huang","sequence":"additional","affiliation":[]}],"member":"297","reference":[{"key":"14_CR1_14","doi-asserted-by":"publisher","first-page":"644","DOI":"10.1046\/j.0956-7976.2003.psci_1479.x","volume":"14","author":"R.B Adams","year":"2003","unstructured":"Adams, R. B&Kleck, R.E(2003). Perceived gaze direction and the processing of facial displays of emotion. Psychological Science, 14, 644\u2013647.","journal-title":"Psychological Science"},{"issue":"2","key":"14_CR2_14","doi-asserted-by":"publisher","first-page":"256","DOI":"10.1037\/0033-2909.111.2.256","volume":"111","author":"N. Ambady","year":"1992","unstructured":"Ambady, N.,&Rosenthal, R. (1992). Thin slices of expressive behavior as predictors of interpersonal consequences: A meta-analysis. Psychological Bulletin, 111(2), 256\u2013274.","journal-title":"Psychological Bulletin"},{"key":"14_CR3_14","first-page":"318","volume-title":"Emotion analysis in man-machine interaction systems","author":"T. Balomenos","year":"2005","unstructured":"Balomenos, T., Raouzaiou, A., Ioannou, S., Drosopoulos, A., Karpouzis, K.,&Kollias, S. (2005).Emotion analysis in man-machine interaction systems (LNCS 3361; pp. 318\u2013328). New York:Springer."},{"key":"14_CR4_14","doi-asserted-by":"crossref","unstructured":"Bartlett, M. S., Littlewort, G., Frank, M., Lainscsek, C., Fasel, I.,&Movellan, J. (2005), Recognizing facial expression: machine learning and application to spontaneous behavior. In IEEE International Conference on Computer Vision and Pattern Recognition (pp. 568\u2013573).","DOI":"10.1109\/CVPR.2005.297"},{"key":"14_CR5_14","doi-asserted-by":"publisher","first-page":"117","DOI":"10.1016\/S0167-6393(02)00079-1","volume":"40","author":"A. Batliner","year":"2003","unstructured":"Batliner, A., Fischer, K., Hubera, R., Spilkera, J.,&Noth, E. (2003). How to find trouble in communication. Speech Communication, 40, 117\u2013143.","journal-title":"Speech Communication"},{"key":"14_CR6_14","doi-asserted-by":"crossref","unstructured":"Busso, C., Deng, Z., Yildirim, S., Bulut, M., Lee, C. M., et al. (2004), Analysis of emotion recognition using facial expressions, speech and multimodal information. In Proceedings of the International Conference on Multimodal Interfaces (pp. 205\u2013211).","DOI":"10.1145\/1027933.1027968"},{"key":"14_CR7_14","doi-asserted-by":"crossref","unstructured":"Caridakis, G., Malatesta, L., Kessous, L., Amir, N., Paouzaiou, A.&Karpouzis, K. (2006). Modeling naturalistic affective states via facial and vocal expression recognition. In Proceedings of the International Conference on Multimodal Interfaces (pp. 146\u2013154).","DOI":"10.1145\/1180995.1181029"},{"key":"14_CR8_14","doi-asserted-by":"crossref","unstructured":"Chen, L., Huang, T. S., Miyasato, T.,&Nakatsu, R. (1998). Multimodal human emotion\/expression recognition. In Proceedings of the International Conference on Automatic Face and Gesture Recognition (pp. 396\u2013401).","DOI":"10.1109\/AFGR.1998.670976"},{"key":"14_CR9_14","volume-title":"Joint processing of audio-visual information for the recognition of emotional expressions in human-computer interaction, PhD thesis","author":"L. S. Chen","year":"2000","unstructured":"Chen, L. S. (2000), Joint processing of audio-visual information for the recognition of emotional expressions in human-computer interaction, PhD thesis, University of Illinois at Urbana-Champaign, USA."},{"key":"14_CR10_14","doi-asserted-by":"crossref","unstructured":"Cohn, J. F. (2006). Foundations of human computing: Facial expression and emotion. In Proceedings of the International Conference on Multimodal Interfaces (pp. 233\u2013238).","DOI":"10.1145\/1180995.1181043"},{"key":"14_CR11_14","doi-asserted-by":"crossref","unstructured":"Cohn, J. F., Reed, L. I., Ambadar, Z., Xiao, J.,&Moriyama, T. (2004). Automatic analysis and recognition of brow actions and head motion in spontaneous facial behavior. In Proceedings of the International Conference on Systems, Man&Cybernetics, 1 (pp. 610\u2013616).","DOI":"10.1109\/ICSMC.2004.1398367"},{"key":"14_CR12_14","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1142\/S021969130400041X","volume":"2","author":"J. F. Cohn","year":"2004","unstructured":"Cohn, J. F.,&Schmidt, K. L.(2004). The timing of facial motion in posed and spontaneous smiles.International Journal of Wavelets, Multiresolution and Information Processing, 2, 1\u201312.","journal-title":"International Journal of Wavelets, Multiresolution and Information Processing"},{"key":"14_CR13_14","doi-asserted-by":"publisher","first-page":"371","DOI":"10.1016\/j.neunet.2005.03.002","volume":"18","author":"R. Cowie","year":"2005","unstructured":"Cowie, R., Douglas-Cowie, E.,&Cox, C. (2005). Beyond emotion archetypes: Databases for emotion modeling using neural networks. Neural Networks, 18, 371\u2013388.","journal-title":"Neural Networks"},{"key":"14_CR14_14","unstructured":"Cowie, R., Douglas-Cowie, E., Savvidou, S., McMahon, E., Sawey, M.,&Schr\u00f6der, M. (2000).\u2018Feeltrace\u2019 An instrument for recording perceived emotion in real time. In Proceedings of the ISCA Workshop on Speech and Emotion (pp. 19\u201324)."},{"key":"14_CR15_14","doi-asserted-by":"crossref","unstructured":"Cowie, R., Douglas-Cowie, E., Tsapatsoulis, N., Votsis, G., Kollias, S., Fellenz, W.,&Taylor J. G. (2001), Emotion recognition in human-computer interaction, IEEE Signal Processing Magazine, January (pp. 32\u201380).","DOI":"10.1109\/79.911197"},{"issue":"1\u20132","key":"14_CR16_14","doi-asserted-by":"publisher","first-page":"33","DOI":"10.1016\/S0167-6393(02)00070-5","volume":"40","author":"E. Douglas-Cowie","year":"2003","unstructured":"Douglas-Cowie, E., Campbell, N., Cowie, R.,&Roach, P. (2003). Emotional speech: Towards a new generation of database. Speech Communication, 40(1\u20132), 33\u201360.","journal-title":"Speech Communication"},{"issue":"7","key":"14_CR17_14","doi-asserted-by":"publisher","first-page":"1272","DOI":"10.1109\/JPROC.2002.801449","volume":"90","author":"Z. Duric","year":"2002","unstructured":"Duric, Z., Gray, WD ., Heishman, R., Li, F., Rosenfeld, A., Schoelles, M. J., Schunn, C.,&Wechsler, H. (2002). Integrating perceptual and cognitive modeling for adaptive and intelligent human\u2014computer interaction. Proceedings of the IEEE, 90(7), 1272\u20131289.","journal-title":"Proceedings of the IEEE"},{"key":"14_CR18_14","volume-title":"Emotion in the human face","year":"1982","unstructured":"Ekman, P. (Ed.) (1982). Emotion in the human face (2nd ed.). New York: Cambridge University Press.","edition":"2"},{"key":"14_CR19_14","volume-title":"Unmasking the face","author":"P. Ekman","year":"1975","unstructured":"Ekman, P.,&Friesen, W. V. (1975). Unmasking the face. Englewood Cliffs, NJ: Prentice-Hall."},{"key":"14_CR20_14","volume-title":"Facial Action Coding System","author":"P. Ekman","year":"2002","unstructured":"Ekman, P., Friesen, W. V.,&Hager, J. C. (2002). Facial Action Coding System. Salt Lake City, UT:A Human Face."},{"key":"14_CR21_14","doi-asserted-by":"crossref","DOI":"10.1093\/acprof:oso\/9780195179644.001.0001","volume-title":"What the face reveals: Basic and applied studies of spontaneous expression using the facial action coding system","author":"P. Ekman","year":"2005","unstructured":"Ekman P.,&Rosenberg, E. L. (2005). What the face reveals: Basic and applied studies of spontaneous expression using the facial action coding system (2nd ed.). Oxford University Press,University of Illinois at Urbana-Champaign, USA.","edition":"2"},{"key":"14_CR22_14","doi-asserted-by":"publisher","first-page":"389","DOI":"10.1016\/j.neunet.2005.03.006","volume":"18","author":"F. Fragopanagos","year":"2005","unstructured":"Fragopanagos, F.,&Taylor, J. G. (2005), Emotion recognition in human\u2014computer interaction.Neural Networks, 18, 389\u2013405.","journal-title":"Neural Networks"},{"key":"14_CR23_14","unstructured":"Go, H. J., Kwak, K. C., Lee, D. J.,&Chun, M.G. (2003). Emotion recognition from facial image and speech signal. In Proceedings of the International Conference of the Society of Instrument and Control Engineers (pp. 2890\u20132895)."},{"key":"14_CR24_14","doi-asserted-by":"crossref","unstructured":"Graciarena, M., Shriberg, E., Stolcke, A., Enos, J. H. F.,&Kajarekar, S. (2006). Combining prosodic, lexical and cepstral systems for deceptive speech detection. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing, I, 1033\u20131036.","DOI":"10.1109\/ICASSP.2006.1660200"},{"key":"14_CR25_14","doi-asserted-by":"crossref","unstructured":"Gunes, H.,&Piccardi, M. (2005). Affect recognition from face and body: early fusion vs. late fusion. In Proceedings of the International Conference on Systems, Man and Cybernetics (pp. 3437\u20133443).","DOI":"10.1109\/ICSMC.2005.1571679"},{"key":"14_CR26_14","first-page":"1148","volume":"1","author":"H. Gunes","year":"2006","unstructured":"Gunes, H.,&Piccardi, M. (2006). A bimodal face and body gesture database for automatic analysis of human nonverbal affective behavior. International Conference on Pattern Recognition, 1,1148\u20131153.","journal-title":"International Conference on Pattern Recognition"},{"key":"14_CR27_14","doi-asserted-by":"crossref","first-page":"369","DOI":"10.1093\/oso\/9780198529613.001.0001","volume-title":"The new handbook of methods in nonverbal behavior research","author":"J. A. Harrigan","year":"2005","unstructured":"Harrigan, J. A., Rosenthal, R.,&Scherer, K. R. (2005). The new handbook of methods in nonverbal behavior research (pp. 369\u2013397). Oxford University Press, USA."},{"key":"14_CR28_14","doi-asserted-by":"crossref","unstructured":"Hoch, S., Althoff, F., McGlaun, G.,&Rigoll, G. (2005), Bimodal fusion of emotional data in an automotive environment. In ICASSP, II (pp. 1085\u20131088).","DOI":"10.1109\/ICASSP.2005.1415597"},{"issue":"5","key":"14_CR29_14","first-page":"862","volume":"36","author":"Q. Ji","year":"2006","unstructured":"Ji, Q., Lan, P.,&Looney, C. (2006). A probabilistic framework for modeling and real-time monitoring human fatigue. IEEE SMC-Part A, 36(5), 862\u2013875.","journal-title":"IEEE SMC-Part A"},{"issue":"8","key":"14_CR30_14","doi-asserted-by":"publisher","first-page":"724","DOI":"10.1016\/j.ijhcs.2007.02.003","volume":"65","author":"A. Kapoor","year":"2007","unstructured":"Kapoor, A., Burleson, W.,&Picard, R. W. (2007). Automatic prediction of frustration. International Journal of Human\u2014Computer Studies, 65(8), 724\u2013736.","journal-title":"International Journal of Human\u2014Computer Studies"},{"key":"14_CR31_14","doi-asserted-by":"crossref","unstructured":"Kapoor, A.,&Picard, R. W. (2005). Multimodal affect recognition in learning environment. In ACM International Conference on Multimedia (pp. 677\u2013682).","DOI":"10.1145\/1101149.1101300"},{"key":"14_CR32_14","first-page":"91","volume-title":"Modeling naturalistic affective states via facial, vocal, and bodily expression recognition","author":"K. Karpouzis","year":"2007","unstructured":"Karpouzis, K., Caridakis, G., Kessous, L., Amir, N., Raouzaiou, A., Malatesta, L.,&Kollias, S.(2007). Modeling naturalistic affective states via facial, vocal, and bodily expression recognition (LNAI 4451; pp. 91\u2013112). New York: Springer."},{"key":"14_CR33_14","doi-asserted-by":"crossref","DOI":"10.1002\/0471660264","volume-title":"Combining pattern classifier: Methods and algorithms","author":"L.I. Kuncheva","year":"2004","unstructured":"Kuncheva, L. I. (2004). Combining pattern classifier: Methods and algorithms. Hoboken, NJ: John Wiley and Sons."},{"issue":"2","key":"14_CR34_14","doi-asserted-by":"publisher","first-page":"293","DOI":"10.1109\/TSA.2004.838534","volume":"13","author":"C. M. Lee","year":"2005","unstructured":"Lee, C. M.,&Narayanan, S. S. (2005). Toward detecting emotions in spoken dialogs. IEEE Transactions on Speech and Audio Processing, 13(2), 293\u2013303.","journal-title":"IEEE Transactions on Speech and Audio Processing"},{"issue":"9","key":"14_CR35_14","doi-asserted-by":"publisher","first-page":"847","DOI":"10.1016\/j.ijhcs.2006.04.001","volume":"64","author":"W. Liao","year":"2006","unstructured":"Liao, W., Zhang, W., Zhu, Z., Ji, Q.,&Gray, W. (2006), Toward a decision-theoretic framework for affect recognition and user assistance. International Journal of Human-Computer Studies,64(9), 847\u2013873.","journal-title":"International Journal of Human-Computer Studies"},{"key":"14_CR36_14","doi-asserted-by":"crossref","unstructured":"Lisetti, C. L.,&Nasoz, F. (2002). MAUI: A multimodal affective user interface. In Proceedings of the International Conference on Multimedia (pp. 161\u2013170).","DOI":"10.1145\/641007.641038"},{"key":"14_CR37_14","doi-asserted-by":"publisher","first-page":"1672","DOI":"10.1155\/S1110865704406192","volume":"11","author":"C. L. Lisetti","year":"2004","unstructured":"Lisetti, C. L.,&Nasoz, F. (2004). Using noninvasive wearable computers to recognize human emotions from physiological signals. EURASIP Journal on Applied Signal Processing, 11, 1672\u20131687.","journal-title":"EURASIP Journal on Applied Signal Processing"},{"key":"14_CR38_14","doi-asserted-by":"crossref","unstructured":"Litman, D. J.,&Forbes-Riley, K. (2004), Predicting student emotions in computer-human tutoring dialogues. In Proceedings of the 42nd Annual Meeting of the Association for Computational Linguistics (ACL), July (pp. 352\u2013359).","DOI":"10.3115\/1218955.1219000"},{"key":"14_CR39_14","doi-asserted-by":"crossref","unstructured":"Littlewort, G. C., Bartlett, M. S.,&Lee, K. (2007). Faces of pain: Automated measurement of spontaneous facial expressions of genuine and posed pain. In Proceedings of the ACM International Conference on Multimodal Interfaces (pp. 15\u201321).","DOI":"10.1145\/1322192.1322198"},{"key":"14_CR40_14","doi-asserted-by":"crossref","unstructured":"Maat, L.,&Pantic, M. (2006). Gaze-X: Adaptive affective multimodal interface for single-user office scenarios, In Proceedings of the ACM International Conference on Multimodal Interfaces (pp. 171\u2013178).","DOI":"10.1145\/1180995.1181032"},{"key":"14_CR41_14","doi-asserted-by":"crossref","unstructured":"Pal, P., Iyer, A. N.,&Yantorno, R. E. (2006). Emotion detection from infant facial expressions and cries. In Proceedings of the International Conference on Acoustics, Speech&Signal Processing, 2 (pp. 721\u2013724).","DOI":"10.1109\/ICASSP.2006.1660444"},{"key":"14_CR42_14","first-page":"377","volume-title":"Face recognition","author":"M. Pantic","year":"2007","unstructured":"Pantic, M.,&Bartlett, M. S. (2007). Machine analysis of facial expressions. In K. Delac and M. Grgic, (Eds.), Face recognition (pp. 377\u2013416). Vienna, Austria: I-Tech Education."},{"key":"14_CR43_14","doi-asserted-by":"crossref","unstructured":"Pantic, M., Pentland, A., Nijholt, A.,&Huang, T. S. (2006). Human computing and machine understanding of human behavior: A survey. In International Conference on Multimodal Interfaces (pp. 239\u2013248).","DOI":"10.1145\/1180995.1181044"},{"issue":"9, Sept.","key":"14_CR44_14","doi-asserted-by":"publisher","first-page":"1370","DOI":"10.1109\/JPROC.2003.817122","volume":"91","author":"M. Pantic","year":"2003","unstructured":"Pantic M.,&Rothkrantz, L. J. M. (2003), Toward an affect-sensitive multimodal human-computer interaction, Proceedings of the IEEE, 91(9, Sept.), 1370\u20131390.","journal-title":"Proceedings of the IEEE"},{"key":"14_CR45_14","doi-asserted-by":"crossref","unstructured":"Pantic, M.,&Rothkrantz, L. J. M. (2004). Case-based reasoning for user-profiled recognition of emotions from face images. In International Conference on Multimedia and Expo (pp. 391\u2013394).","DOI":"10.1109\/ICME.2004.1394211"},{"key":"14_CR46_14","doi-asserted-by":"crossref","unstructured":"Pantic, M., Valstar, M. F, Rademaker, R.,&Maat, L. (2005). Web-based database for facial expression analysis. In International Conference on Multimedia and Expo (pp. 317\u2013321).","DOI":"10.1109\/ICME.2005.1521424"},{"key":"14_CR47_14","doi-asserted-by":"crossref","unstructured":"Patras, I.,&Pantic, M. (2004). Particle filtering with factorized likelihoods for tracking facial features, In Proceedings of the IEEE International Conference on Face and Gesture Recognition (pp. 97\u2013102).","DOI":"10.1109\/AFGR.2004.1301515"},{"key":"14_CR48_14","doi-asserted-by":"crossref","first-page":"33","DOI":"10.1109\/MC.2005.104","volume":"38","author":"A. Pentland","year":"2005","unstructured":"Pentland, A. (2005). Socially aware, computation and communication, IEEE Computer, 38, 33\u201340.","journal-title":"IEEE Computer"},{"key":"14_CR49_14","doi-asserted-by":"crossref","unstructured":"Petridis, S.,&Pantic, M. (2008). Audiovisual discrimination between laughter and speech, In IEEE Int'l Conf. Acoustics, Speech, and Signal Processing (pp. 5117\u20135120).","DOI":"10.1109\/ICASSP.2008.4518810"},{"key":"14_CR50_14","doi-asserted-by":"crossref","DOI":"10.7551\/mitpress\/1140.001.0001","volume-title":"Affective computing","author":"R. W. Picard","year":"1997","unstructured":"Picard, R. W. (1997). Affective computing. Cambridge, MA: MIT Press."},{"issue":"10","key":"14_CR51_14","doi-asserted-by":"publisher","first-page":"1175","DOI":"10.1109\/34.954607","volume":"23","author":"R. W. Picard","year":"2001","unstructured":"Picard, R. W., Vyzas, E.,&Healey, J. (2001). Toward machine emotional intelligence: Analysis of affective physiological state. IEEE Transactions on Pattern Analysis and Machine Intelligence,23(10), 1175\u20131191.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"14_CR52_14","doi-asserted-by":"publisher","first-page":"590","DOI":"10.2307\/2670179","volume":"94","author":"M. K. Pitt","year":"1999","unstructured":"Pitt, M. K.,&Shephard, N. (1999). Filtering via simulation: auxiliary particle filtering. Journal of the American Statistical Association, 94, 590\u2013599.","journal-title":"Journal of the American Statistical Association"},{"issue":"5","key":"14_CR53_14","doi-asserted-by":"publisher","first-page":"776","DOI":"10.1037\/0012-1649.40.5.776","volume":"40","author":"G. I. Roisman","year":"2004","unstructured":"Roisman, G. I., Tsai, J. L.,&Chiang, K. S. (2004). The emotional integration of childhood experience: Physiological, facial expressive, and self-reported emotional response during the adult attachment interview. Developmental Psychology, 40(5), 776\u2013789.","journal-title":"Developmental Psychology"},{"key":"14_CR54_14","doi-asserted-by":"publisher","first-page":"329","DOI":"10.1146\/annurev.psych.54.101601.145102","volume":"54","author":"J. A. Russell","year":"2003","unstructured":"Russell, J. A., Bachorowski, J.,&Fernandez-Dols, J. (2003). Facial and vocal expressions of emotion. Ann. Rev. Psychol. 54, 329\u2013349.","journal-title":"Ann. Rev. Psychol."},{"key":"14_CR55_14","doi-asserted-by":"publisher","first-page":"637","DOI":"10.1002\/0470013494.ch30","volume-title":"Handbook of cognition and emotion","author":"K. R. Scherer","year":"1999","unstructured":"Scherer K. R. (1999). Appraisal theory. In T. Dalgleish&M. J. Power (Eds.), Handbook of cognition and emotion, New York: Wiley, 637\u2013663."},{"key":"14_CR56_14","doi-asserted-by":"crossref","unstructured":"Schuller, B., Villar, R. J., Rigoll, G.,&Lang, M. (2005). Meta-classifiers in acoustic and linguistic feature fusion-based affect recognition. In International Conference on Acoustics, Speech, and Signal Processing (pp. 325\u2013328).","DOI":"10.1109\/ICASSP.2005.1415116"},{"key":"14_CR57_14","doi-asserted-by":"crossref","unstructured":"Sebe, N., Cohen, I., Gevers, T.,&Huang, T. S. (2006). Emotion recognition based on joint visual and audio cues. In International Conference on Pattern Recognition (pp. 1136\u20131139).","DOI":"10.1109\/ICPR.2006.489"},{"key":"14_CR58_14","volume-title":"Handbook of Pattern Recognition and Computer Vision","author":"N. Sebe","year":"2005","unstructured":"Sebe, N., Cohen, I.,&Huang, T. S. (2005). Multimodal emotion recognition. In Handbook of Pattern Recognition and Computer Vision. Singapore: World Scientific."},{"key":"14_CR59_14","unstructured":"Song, M., Bu, J., Chen, C.,&Li, N. (2004), Audio-visual based emotion recognition\u2014A new approach. In International Conference on Computer Vision and Pattern Recognition (pp. 1020\u20131025)."},{"key":"14_CR60_14","volume-title":"The merging of senses","author":"B. Stein","year":"1993","unstructured":"Stein, B.,&Meredith, M. A. (1993). The merging of senses. Cambridge, MA: MIT Press."},{"key":"14_CR61_14","first-page":"225","volume-title":"Handbook of affective sciences","author":"G. Stemmler","year":"2003","unstructured":"Stemmler, G. (2003). Methodological considerations in the psychophysiological study of emotion.In R. J. Davidson, K. R. Scherer,&H. H. Goldsmith (Eds.), Handbook of affective sciences (pp. 225\u2013255). Oxford University Press, USA."},{"key":"14_CR62_14","unstructured":"Tao, H.,&Huang, T. S. (1999), Explanation-based facial motion tracking using a piecewise Bezier volume deformation mode. In CVPR'99, 1 (pp. 611\u2013617)."},{"key":"14_CR63_14","doi-asserted-by":"publisher","first-page":"144","DOI":"10.1016\/j.specom.2007.01.001","volume":"49","author":"K. P. Truong","year":"2007","unstructured":"Truong, K. P.,&van Leeuwen, D. A. (2007) Automatic discrimination between laughter and speech, Speech Communication, 49, 144\u2013158.","journal-title":"Speech Communication"},{"key":"14_CR64_14","doi-asserted-by":"crossref","unstructured":"Valstar, M. F., Gunes, H.,&Pantic, M. (2007). How to distinguish posed from spontaneous smiles using geometric features. In ACM Int'l Conf. Multimodal Interfaces (pp. 38\u201345).","DOI":"10.1145\/1322192.1322202"},{"key":"14_CR65_14","doi-asserted-by":"crossref","unstructured":"Valstar, M., Pantic, M., Ambadar, Z.,&Cohn, J. F. (2006). Spontaneous vs. posed facial behavior: Automatic analysis of brow actions. In International Conference on Multimedia Interfaces (pp. 162\u2013170).","DOI":"10.1145\/1180995.1181031"},{"key":"14_CR66_14","first-page":"149","volume":"3","author":"M. F. Valstar","year":"2006","unstructured":"Valstar, M. F.,&Pantic, M. (2006). Fully automatic facial action unit detection and temporal analysis. Proceedings of IEEE International Conference on Computer Vision and Pattern Recognition, 3,149.","journal-title":"Proceedings of IEEE International Conference on Computer Vision and Pattern Recognition"},{"key":"14_CR67_14","unstructured":"Wang, Y.,&Guan, L. (2005), Recognizing human emotion from audiovisual information. In ICASSP, II (pp. 1125\u20131128)."},{"key":"14_CR68_14","first-page":"113","volume-title":"Emotion: Theory, research and experience. The measurement of emotions","author":"C. M. Whissell","year":"1989","unstructured":"Whissell, C. M. (1989). The dictionary of affect in language. In R. Plutchik&H. Kellerman (Eds.).Emotion: Theory, research and experience. The measurement of emotions (vol. 4; pp. 113\u2013131).New York: Academic Press."},{"issue":"1","key":"14_CR69_14","doi-asserted-by":"publisher","first-page":"85","DOI":"10.1002\/ima.10048","volume":"13","author":"J. Xiao","year":"2003","unstructured":"Xiao, J., Moriyama, T., Kanade, T.,&Cohn, J. F. (2003). Robust full-motion recovery of head by dynamic templates and re-registration techniques. International Journal of Imaging Systems and Technology, 13(1), 85\u201394.","journal-title":"International Journal of Imaging Systems and Technology"},{"key":"14_CR70_14","doi-asserted-by":"crossref","unstructured":"Yoshimoto, D., Shapiro, A., O'Brian, K.,&Gottman, J. M. (2005). Nonverbal communication coding systems of committed couples. In New handbook of methods in nonverbal behavior research, J.A. Harrigan, R. Rosenthal, and K. R. Scherer (Eds.) (pp. 369\u2013397), USA.","DOI":"10.1093\/oso\/9780198529613.003.0010"},{"key":"14_CR71_14","doi-asserted-by":"crossref","unstructured":"Zeng, Z., Hu, Y., Liu, M., Fu, Y.,&Huang, T. S.(2006), Training combination strategy of multi-stream fused hidden markov model for audio-visual affect recognition. In Proceedings of the ACM International Conference on Multimedia (pp. 65\u201368).","DOI":"10.1145\/1180639.1180661"},{"key":"14_CR72_14","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1007\/978-3-540-72348-6_4","volume-title":"Artificial Intelligence for Human Computing","author":"Z. Zeng","year":"2007a","unstructured":"Zeng, Z., Hu, Y., Roisman, G. I., Wen, Z., Fu, Y.,&Huang, T. S. (2007a), Audio-visual spontaneous emotion recognition. In T. S. Huang, A. Nijholt, M. Pantic,&A. Pentland (Eds.) Artificial Intelligence for Human Computing (LNAI 4451, pp. 72\u201390). New York, Springer."},{"key":"14_CR73_14","doi-asserted-by":"crossref","unstructured":"Zeng, Z., Pantic, M., Roisman, G. I.,&Huang, T. S. (2008a). A survey of affect recognition methods: Audio, visual and spontaneous expressions. IEEE Transactions on Pattern Analysis and Machine Intelligence (in press).","DOI":"10.1145\/1322192.1322216"},{"key":"14_CR74_14","doi-asserted-by":"crossref","unstructured":"Zeng, Z., Tu, J., Liu, M., Zhang, T., Rizzolo, N., Zhang, Z., Huang, T. S., Roth, D.,&Levinson, S.(2004), Bimodal HCI-related Emotion Recognition, In International Conference on Multi-modal Interfaces (pp. 137\u2013143).","DOI":"10.1145\/1027933.1027958"},{"issue":"4","key":"14_CR75_14","doi-asserted-by":"publisher","first-page":"570","DOI":"10.1109\/TMM.2008.921737","volume":"10","author":"Z. Zeng","year":"2008b","unstructured":"Zeng, Z., Tu, J., Pianfetti, B.,&Huang, T. S. (2008b). Audio-visual affective expression recognition through multi-stream fused HMM. IEEE Transactions on Multimedia, June 2008, 10(4),570\u2013577.","journal-title":"IEEE Transactions on Multimedia"},{"issue":"2","key":"14_CR76_14","doi-asserted-by":"publisher","first-page":"424","DOI":"10.1109\/TMM.2006.886310","volume":"9","author":"Z. Zeng","year":"2007b","unstructured":"Zeng, Z., Tu, J., Liu, M., Huang, T. S., Pianfetti, B., Roth, D.,&Levinson, S. (2007b). Audio-visual affect recognition. IEEE Transactions on Multimedia, 9(2), 424\u2013428.","journal-title":"IEEE Transactions on Multimedia"},{"issue":"5","key":"14_CR77_14","doi-asserted-by":"publisher","first-page":"699","DOI":"10.1109\/TPAMI.2005.93","volume":"27","author":"Y. Zhang","year":"2005","unstructured":"Zhang, Y.,&Ji, Q. (2005). Active and dynamic information fusion for facial expression understanding from image sequences. IEEE Transactions on Pattern Analysis and Machine Intelligence, 27(5), 699\u2013714.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"}],"container-title":["Affective Information Processing"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-1-84800-306-4_14.pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,5]],"date-time":"2025-02-05T01:53:36Z","timestamp":1738720416000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-1-84800-306-4_14"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[null]]},"ISBN":["9781848003057","9781848003064"],"references-count":77,"URL":"https:\/\/doi.org\/10.1007\/978-1-84800-306-4_14","relation":{},"subject":[]}}