{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,6]],"date-time":"2024-09-06T06:22:58Z","timestamp":1725603778343},"publisher-location":"Berlin, Heidelberg","reference-count":20,"publisher":"Springer Berlin Heidelberg","isbn-type":[{"type":"print","value":"9783642231223"},{"type":"electronic","value":"9783642231230"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2011]]},"DOI":"10.1007\/978-3-642-23123-0_35","type":"book-chapter","created":{"date-parts":[[2011,8,31]],"date-time":"2011-08-31T18:29:37Z","timestamp":1314815377000},"page":"346-355","source":"Crossref","is-referenced-by-count":6,"title":["Fusion of Audio- and Visual Cues for Real-Life Emotional Human Robot Interaction"],"prefix":"10.1007","author":[{"given":"Ahmad","family":"Rabie","sequence":"first","affiliation":[]},{"given":"Uwe","family":"Handmann","sequence":"additional","affiliation":[]}],"member":"297","reference":[{"key":"35_CR1","doi-asserted-by":"crossref","first-page":"214","DOI":"10.1145\/1088463.1088501","volume-title":"Proc. Int. Conf. Multimodal Interfaces","author":"A. Battocchi","year":"2005","unstructured":"Battocchi, A., Pianesi, F., Goren-Bar, D.: A first evaluation study of a database of kinetic facial expressions (dafex). In: Proc. Int. Conf. Multimodal Interfaces, pp. 214\u2013221. ACM Press, New York (2005)"},{"key":"35_CR2","volume-title":"Unmasking the Face: A Guide to Recognizing Emotions from Facial Expressions","author":"P. Ekman","year":"1975","unstructured":"Ekman, P., Friesen, W.: Unmasking the Face: A Guide to Recognizing Emotions from Facial Expressions. Prentice Hall, Englewood Cliffs (1975)"},{"key":"35_CR3","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1145\/1178745.1178762","volume-title":"Proc. ACM Int. Workshop on Human-Centered Multimedia","author":"M. Paleari","year":"2006","unstructured":"Paleari, M., Lisetti, C.L.: Toward multimodal fusion of affective cues. In: Proc. ACM Int. Workshop on Human-Centered Multimedia, pp. 99\u2013108. ACM, New York (2006)"},{"key":"35_CR4","doi-asserted-by":"crossref","unstructured":"Busso, C., Deng, Z., Yildirim, S., Bulut, M., Lee, C.M., Kazemzadeh, A., Lee, S., Neumann, U., Narayanan, S.: Analysis of emotion recognition using facial expressions, speech and multimodal information. In: Proc. Int. Conf. Multimodal Interfaces (2004)","DOI":"10.1145\/1027933.1027968"},{"key":"35_CR5","doi-asserted-by":"crossref","first-page":"146","DOI":"10.1145\/1180995.1181029","volume-title":"Proc. Int. Conf. Multimodal Interfaces","author":"G. Caridakis","year":"2006","unstructured":"Caridakis, G., Malatesta, L., Kessous, L., Amir, N., Raouzaiou, A., Karpouzis, K.: Modeling naturalistic affective states via facial and vocal expressions recognition. In: Proc. Int. Conf. Multimodal Interfaces, pp. 146\u2013154. ACM, New York (2006)"},{"key":"35_CR6","doi-asserted-by":"crossref","first-page":"139","DOI":"10.1145\/1180995.1181028","volume-title":"Proc. Int. Conf. on Multimodal Interfaces","author":"Z. Zeng","year":"2006","unstructured":"Zeng, Z., Hu, Y., Fu, Y., Huang, T.S., Roisman, G.I., Wen, Z.: Audio-visual emotion recognition in adult attachment interview. In: Proc. Int. Conf. on Multimodal Interfaces, pp. 139\u2013145. ACM, New York (2006)"},{"key":"35_CR7","doi-asserted-by":"crossref","unstructured":"Massaro, D.W., Egan, P.B.: Perceiving affect from the voice and the face. Psychonomoic Bulletin and Review\u00a0(3), 215\u2013221","DOI":"10.3758\/BF03212421"},{"key":"35_CR8","doi-asserted-by":"publisher","first-page":"321","DOI":"10.1080\/026999300378842","volume":"14","author":"B. Gelder de","year":"2000","unstructured":"de Gelder, B., Vroomen, J.: Bimodal emotion perception: integration across separate modalities, cross-modal perceptula grouping or perception of multimodal events? Cognition and Emotion\u00a014, 321\u2013324 (2000)","journal-title":"Cognition and Emotion"},{"key":"35_CR9","unstructured":"Schwartz, J.L.: Why the FLMP should not be applied to McGurk data. or how to better compare models in the bazesian framework. In: Proc. Int. Conf. Audio-Visual Speech Processing, pp. 77\u201382 (2003)"},{"key":"35_CR10","doi-asserted-by":"crossref","unstructured":"Fagel, S.: Emotional mcgurk effect. In: Proc. Int. Conf. on Speech Prosody, Dresden, Germany (2006)","DOI":"10.21437\/SpeechProsody.2006-212"},{"key":"35_CR11","unstructured":"Rabie, A., Lang, C., Hanheide, M., Castrillon-Santana, M., Sagerer, G.: Automatic initialization for facial analysis in interactive robotics (2008)"},{"key":"35_CR12","doi-asserted-by":"crossref","unstructured":"Hegel, F., Spexard, T., Vogt, T., Horstmann, G., Wrede, B.: Playing a different imitation game: Interaction with an empathic android robot. In: Proc. Int. Conf. Humanoid Robots, pp. 56\u201361 (2006)","DOI":"10.1109\/ICHR.2006.321363"},{"key":"35_CR13","doi-asserted-by":"publisher","first-page":"681","DOI":"10.1109\/34.927467","volume":"23","author":"T.F. Cootes","year":"2001","unstructured":"Cootes, T.F., Edwards, G.J., Taylor, C.J.: Active appearance models. PAMI\u00a023, 681\u2013685 (2001)","journal-title":"PAMI"},{"key":"35_CR14","doi-asserted-by":"publisher","first-page":"130","DOI":"10.1016\/j.jvcir.2006.11.004","volume":"18","author":"M. Castrill\u00f3n","year":"2007","unstructured":"Castrill\u00f3n, M., D\u00e9niz, O., Guerra, C., Hern\u00e1ndez, M.: Encara2: Real-time detection of multiple faces at different resolutions in video streams. Journal of Visual Communication and Image Representation\u00a018, 130\u2013140 (2007)","journal-title":"Journal of Visual Communication and Image Representation"},{"key":"35_CR15","doi-asserted-by":"crossref","unstructured":"Hanheide, M., Wrede, S., Lang, C., Sagerer, G.: Who am i talking with? a face memory for social robots (2008)","DOI":"10.1109\/ROBOT.2008.4543772"},{"key":"35_CR16","unstructured":"Vogt, T., Andr\u00e9, E., Bee, N.: Emovoice \u2014 A framework for online recognition of emotions from voice. In: Proc. Workshop on Perception and Interactive Technologies for Speech-Based Systems, Irsee, Germany (2008)"},{"key":"35_CR17","unstructured":"Hall, M.A.: Correlation-based feature subset selection for machine learning. Master\u2019s thesis, University of Waikato, New Zealand (1998)"},{"key":"35_CR18","unstructured":"Vogt, T., Andr\u00e9, E.: Comparing feature sets for acted and spontaneous speech in view of automatic emotion recognition. In: Proc. of IEEE Int. Conf. on Multimedia & Expo., Amsterdam, The Netherlands (2005)"},{"key":"35_CR19","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1109\/TPAMI.2008.52","volume":"31","author":"Z. Zeng","year":"2009","unstructured":"Zeng, Z., Pantic, M., Roisman, G.I., Huang, T.S.: A survey of affect recognition methods: Audio, visual, and spontaneous expressions. IEEE Transaction on Pattern Analysis and Macine Intellegence\u00a031, 39\u201358 (2009)","journal-title":"IEEE Transaction on Pattern Analysis and Macine Intellegence"},{"key":"35_CR20","doi-asserted-by":"crossref","unstructured":"Rabie, A., Vogt, T., Hanheide, M., Wrede, B.: Evaluation and discussion of multi-modal emotion recognition. In: ICCEE (2009)","DOI":"10.1109\/ICCEE.2009.192"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition"],"original-title":[],"link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-642-23123-0_35.pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,4,10]],"date-time":"2024-04-10T19:30:16Z","timestamp":1712777416000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-642-23123-0_35"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2011]]},"ISBN":["9783642231223","9783642231230"],"references-count":20,"URL":"https:\/\/doi.org\/10.1007\/978-3-642-23123-0_35","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2011]]}}}