{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,18]],"date-time":"2026-03-18T14:07:20Z","timestamp":1773842840106,"version":"3.50.1"},"publisher-location":"Cham","reference-count":138,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783319310558","type":"print"},{"value":"9783319310565","type":"electronic"}],"license":[{"start":{"date-parts":[[2016,1,1]],"date-time":"2016-01-01T00:00:00Z","timestamp":1451606400000},"content-version":"unspecified","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2016]]},"DOI":"10.1007\/978-3-319-31056-5_11","type":"book-chapter","created":{"date-parts":[[2016,3,21]],"date-time":"2016-03-21T05:03:17Z","timestamp":1458536597000},"page":"205-238","source":"Crossref","is-referenced-by-count":42,"title":["Analysis of Emotional Speech\u2014A Review"],"prefix":"10.1007","author":[{"given":"P.","family":"Gangamohan","sequence":"first","affiliation":[]},{"given":"Sudarsana Reddy","family":"Kadiri","sequence":"additional","affiliation":[]},{"given":"B.","family":"Yegnanarayana","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2016,3,22]]},"reference":[{"key":"11_CR1","doi-asserted-by":"crossref","unstructured":"Airas M, Alku P (2004) Emotions in short vowel segments: effects of the glottal flow as reflected by the normalized amplitude quotient. In: Affective dialogue systems. Springer, pp 13\u201324","DOI":"10.1007\/978-3-540-24842-2_2"},{"key":"11_CR2","doi-asserted-by":"crossref","unstructured":"Airas M, Pulakka H, B\u00e4ckstr\u00f6m T, Alku P (2005) A toolkit for voice inverse filtering and parametrization. In: INTERSPEECH. Lisbon, Portugal, pp 2145\u20132148","DOI":"10.21437\/Interspeech.2005-434"},{"issue":"5","key":"11_CR3","doi-asserted-by":"publisher","first-page":"623","DOI":"10.1007\/s12046-011-0041-5","volume":"36","author":"P Alku","year":"2011","unstructured":"Alku P (2011) Glottal inverse filtering analysis of human voice production a review of estimation and parameterization methods of the glottal excitation and their applications. Sadhana 36(5):623\u2013650","journal-title":"Sadhana"},{"issue":"5","key":"11_CR4","doi-asserted-by":"crossref","first-page":"240","DOI":"10.1159\/000266415","volume":"48","author":"Paavo Alku","year":"1996","unstructured":"Alku P, Vilkman E (1996) A comparison of glottal voice source quantification parameters in breathy, normal and pressed phonation of female and male speakers. Folia Phoniatrica et Logopaedica 48:240\u2013254","journal-title":"Folia Phoniatrica et Logopaedica"},{"key":"11_CR5","unstructured":"Amer MR, Siddiquie B, Richey C, Divakaran A (2014) Emotion recognition in speech using deep networks. In: ICASSP. Florence, Italy, pp 3752\u20133756"},{"key":"11_CR6","doi-asserted-by":"crossref","unstructured":"Amir N, Kerret O, Karlinski D (2001) Classifying emotions in speech: a comparison of methods. In: INTERSPEECH. Aalborg, Denmark, pp 127\u2013130","DOI":"10.21437\/Eurospeech.2001-44"},{"key":"11_CR7","doi-asserted-by":"crossref","unstructured":"Ang j, Dhillon R, Krupski A, Shriberg E, Stolcke A (2002) Prosody-based automatic detection of annoyance and frustration in human-computer dialog. In: INTERSPEECH. Denver, Colorado, USA","DOI":"10.21437\/ICSLP.2002-559"},{"key":"11_CR8","doi-asserted-by":"crossref","unstructured":"Arias JP, Busso C, Yoma NB (2013) Energy and F0 contour modeling with functional data analysis for emotional speech detection. In: INTERSPEECH. Lyon, France, pp 2871\u20132875","DOI":"10.21437\/Interspeech.2013-253"},{"issue":"1","key":"11_CR9","doi-asserted-by":"publisher","first-page":"278","DOI":"10.1016\/j.csl.2013.07.002","volume":"28","author":"JP Arias","year":"2014","unstructured":"Arias JP, Busso C, Yoma NB (2014) Shape-based modeling of the fundamental frequency contour for emotion detection in speech. Comput Speech Lang 28(1):278\u2013294","journal-title":"Comput Speech Lang"},{"key":"11_CR10","doi-asserted-by":"crossref","unstructured":"Atassi H, Esposito A (2008) A speaker independent approach to the classification of emotional vocal expressions. In: IEEE international conference on tools with artificial intelligence (ICTAI\u201908), vol 2. Dayton, Ohio, USA, pp 147\u2013152","DOI":"10.1109\/ICTAI.2008.158"},{"key":"11_CR11","doi-asserted-by":"crossref","first-page":"255","DOI":"10.1007\/978-3-642-12397-9_21","volume-title":"Development of Multimodal Interfaces: Active Listening and Synchrony","author":"Hicham Atassi","year":"2010","unstructured":"Atassi H, Riviello M, Sm\u00e9kal Z, Hussain A, Esposito A (2010) Emotional vocal expressions recognition using the COST 2102 Italian database of emotional speech. In: Esposito A, Campbell N, Vogel C, Hussain A, Nijholt A (eds) Development of multimodal interfaces: active listening and synchrony. Lecture notes in computer science, vol 5967. Springer, Berlin, pp 255\u2013267"},{"issue":"2","key":"11_CR12","doi-asserted-by":"publisher","first-page":"53","DOI":"10.1111\/1467-8721.00013","volume":"8","author":"J Bachorowski","year":"1999","unstructured":"Bachorowski J (1999) Vocal expression and perception of emotion. Curr Dir Psychol Sci 8(2):53\u201357","journal-title":"Curr Dir Psychol Sci"},{"issue":"3","key":"11_CR13","doi-asserted-by":"publisher","first-page":"614","DOI":"10.1037\/0022-3514.70.3.614","volume":"70","author":"R Banse","year":"1996","unstructured":"Banse R, Scherer KR (1996) Acoustic profiles in vocal emotion expression. J Personal Soc Psychol 70(3):614\u2013636","journal-title":"J Personal Soc Psychol"},{"key":"11_CR14","doi-asserted-by":"crossref","unstructured":"Batliner A, Schuller B, Seppi D, Steidl S, Devillers L, Vidrascu L, Vogt T, Aharonson V, Amir N (2011) The automatic recognition of emotions in speech. In: Petta P, Pelachaud C, Cowie R (eds) Emotion-oriented systems. Springer, pp 71\u201399","DOI":"10.1007\/978-3-642-15184-2_6"},{"key":"11_CR15","doi-asserted-by":"publisher","first-page":"387","DOI":"10.1177\/0022002183014004001","volume":"14","author":"RAMG Bezooijen","year":"1983","unstructured":"Bezooijen RAMG, Otto SA, Heenan TA (1983) Recognition of vocal expressions of emotion: a three-nation study to identify universal characteristics. J Cross-Cult Psychol 14:387\u2013406","journal-title":"J Cross-Cult Psychol"},{"issue":"9\/10","key":"11_CR16","first-page":"341","volume":"5","author":"P Boersma","year":"2001","unstructured":"Boersma P, Heuven VV (2001) Speak and unSpeak with PRAAT. Glot Int 5(9\/10):341\u2013347","journal-title":"Glot Int"},{"key":"11_CR17","doi-asserted-by":"crossref","unstructured":"Burkhardt F, Paeschke A, Rolfes M, Sendlmeier WF, Weiss B (2005) A database of German emotional speech. In: INTERSPEECH. Lisbon, Portugal, pp 1517\u20131520","DOI":"10.21437\/Interspeech.2005-446"},{"issue":"4","key":"11_CR18","doi-asserted-by":"publisher","first-page":"335","DOI":"10.1007\/s10579-008-9076-6","volume":"42","author":"C Busso","year":"2008","unstructured":"Busso C, Bulut M, Lee C, Kazemzadeh A, Mower E, Kim S, Chang JN, Lee S, Narayanan S (2008) IEMOCAP: interactive emotional dyadic motion capture database. Lang Res Eval 42(4):335\u2013359","journal-title":"Lang Res Eval"},{"key":"11_CR19","doi-asserted-by":"crossref","unstructured":"Chastagnol C, Devillers L (2011) Analysis of anger across several agent-customer interactions in French call centers. In: ICASSP. Prague, Czech Republic, pp 4960\u20134963","DOI":"10.1109\/ICASSP.2011.5947469"},{"issue":"5","key":"11_CR20","doi-asserted-by":"publisher","first-page":"2394","DOI":"10.1121\/1.402044","volume":"90","author":"DG Childers","year":"1991","unstructured":"Childers DG, Lee CK (1991) Vocal quality factors: analysis, synthesis, and perception. J Acoust Soc Am 90(5):2394\u20132410","journal-title":"J Acoust Soc Am"},{"issue":"1\u20132","key":"11_CR21","doi-asserted-by":"publisher","first-page":"5","DOI":"10.1016\/S0167-6393(02)00071-7","volume":"40","author":"R Cowie","year":"2003","unstructured":"Cowie R, Cornelius RR (2003) Describing the emotional states that are expressed in speech. Speech Commun 40(1\u20132):5\u201332","journal-title":"Speech Commun"},{"key":"11_CR22","doi-asserted-by":"crossref","unstructured":"Darwin C (1872) The expression of emotion in man and animals. reprinted by University of Chicago Press, Murray, London, UK (1975)","DOI":"10.1037\/10001-000"},{"key":"11_CR23","volume-title":"The communication of emotional meaning","author":"JR Davitz","year":"1964","unstructured":"Davitz JR (1964) Personality, perceptual, and cognitive correlates of emotional sensitivity. In: Davitz JR (ed) The communication of emotional meaning. McGraw-Hill, New York"},{"key":"11_CR24","doi-asserted-by":"crossref","unstructured":"Dellaert F, Polzin T, Waibel A (1996) Recognizing emotion in speech. In: international conference on spoken language processing (ICSLP). Philadelphia, USA, pp 1970\u20131973","DOI":"10.1109\/ICSLP.1996.608022"},{"key":"11_CR25","doi-asserted-by":"crossref","unstructured":"Devillers L, Vidrascu L (2006) Real-life emotions detection with lexical and paralinguistic cues on human-human call center dialogs. In: INTERSPEECH. Pittsburgh, PA, USA, pp 801\u2013804","DOI":"10.21437\/Interspeech.2006-275"},{"issue":"1\u20132","key":"11_CR26","doi-asserted-by":"publisher","first-page":"33","DOI":"10.1016\/S0167-6393(02)00070-5","volume":"40","author":"E Douglas-Cowie","year":"2003","unstructured":"Douglas-Cowie E, Campbell N, Cowie R, Roach P (2003) Emotional speech: towards a new generation of databases. Speech Commun 40(1\u20132):33\u201360","journal-title":"Speech Commun"},{"key":"11_CR27","doi-asserted-by":"publisher","first-page":"169","DOI":"10.1080\/02699939208411068","volume":"6","author":"P Ekman","year":"1992","unstructured":"Ekman P (1992) An argument for basic emotions. Cognit Emot 6:169\u2013200","journal-title":"Cognit Emot"},{"key":"11_CR28","doi-asserted-by":"crossref","unstructured":"Engberg IS, Hansen AV, Andersen O, Dalsgaard P (1997) Design, recording and verification of a Danish emotional speech database. In: EUROSPEECH. Rhodes, Greece, pp 1695\u20131698","DOI":"10.21437\/Eurospeech.1997-482"},{"key":"11_CR29","doi-asserted-by":"crossref","unstructured":"Erden M, Arslan LM (2011) Automatic detection of anger in human-human call center dialogs. In: INTERSPEECH. Florence, Italy, pp 81\u201384","DOI":"10.21437\/Interspeech.2011-21"},{"key":"11_CR30","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1159\/000091404","volume":"63","author":"D Erickson","year":"2006","unstructured":"Erickson D, Yoshida K, Menezes C, Fujino A, Mochida T, Shibuya Y (2006) Exploratory study of some acoustic and articulatory characteristics of sad speech. Phonetica 63:1\u20135","journal-title":"Phonetica"},{"issue":"5","key":"11_CR31","doi-asserted-by":"publisher","first-page":"974","DOI":"10.1109\/TASL.2009.2038658","volume":"18","author":"D Erro","year":"2010","unstructured":"Erro D, Navas E, Hern\u00e1ez I, Saratxaga I (2010) Emotion conversion based on prosodic unit selection. IEEE Trans Audio Speech Lang Process 18(5):974\u2013983","journal-title":"IEEE Trans Audio Speech Lang Process"},{"key":"11_CR32","unstructured":"Espinosa HP, Garcia JO, Pineda LV (2010) Features selection for primitives estimation on emotional speech. In: ICASSP. Florence, Italy, pp 5138\u20135141"},{"key":"11_CR33","unstructured":"Eyben F, Wollmer M, Schuller B (2009) OpenEarIntroducing the Munich open-source emotion and affect recognition toolkit. In: International conference on affective computing and intelligent interaction and workshops (ACII). Amsterdam, Netherlands, pp 1\u20136"},{"key":"11_CR34","unstructured":"Eyben F, Batliner A, Schuller B, Seppi D, Steidl S (2010) Cross-corpus classification of realistic emotions\u2014some pilot experiments. In: International workshop on EMOTION (satellite of LREC): corpora for research on emotion and affect. Valletta, Malta, pp 77\u201382"},{"key":"11_CR35","doi-asserted-by":"crossref","unstructured":"Eyben F, W\u00f6llmer M, Schuller B (2010) OpenSMILE: The Munich versatile and fast open-source audio feature extractor. In: International conference on multimedia. Firenze, Italy, pp 1459\u20131462","DOI":"10.1145\/1873951.1874246"},{"key":"11_CR36","doi-asserted-by":"publisher","first-page":"85","DOI":"10.1080\/03637754109374888","volume":"8","author":"G Fairbanks","year":"1941","unstructured":"Fairbanks G, Hoaglin LW (1941) An experimental study of the durational characteristics of the voice during the expression of emotion. Speech Monogr 8:85\u201391","journal-title":"Speech Monogr"},{"key":"11_CR37","doi-asserted-by":"publisher","first-page":"87","DOI":"10.1080\/03637753909374863","volume":"6","author":"G Fairbanks","year":"1939","unstructured":"Fairbanks G, Pronovost W (1939) An experimental study of the pitch characteristics of the voice during the expression of emotion. Speech Monogr 6:87\u2013104","journal-title":"Speech Monogr"},{"key":"11_CR38","first-page":"21","volume":"26","author":"G Fant","year":"1985","unstructured":"Fant G, Lin Q, Gobl C (1985) Notes on glottal flow interaction. Speech Transm Lab Q Progress Status Rep, KTH 26:21\u201325","journal-title":"Speech Transm Lab Q Progress Status Rep, KTH"},{"issue":"9\u201310","key":"11_CR39","doi-asserted-by":"publisher","first-page":"1088","DOI":"10.1016\/j.specom.2011.05.003","volume":"53","author":"R Fernandez","year":"2011","unstructured":"Fernandez R, Picard R (2011) Recognizing affect from speech prosody using hierarchical graphical models. Speech Commun 53(9\u201310):1088\u20131103","journal-title":"Speech Commun"},{"key":"11_CR40","first-page":"293","volume":"16","author":"I Fonagy","year":"1963","unstructured":"Fonagy I, Magdics K (1963) Emotional patterns in intonation and music. Kommunikationforsch 16:293\u2013326","journal-title":"Kommunikationforsch"},{"key":"11_CR41","doi-asserted-by":"crossref","unstructured":"Gangamohan P, Mittal VK, Yegnanarayana B (2012) A flexible analysis and synthesis tool (FAST) for studying the characteristic features of emotion in speech. In: IEEE international conference on consumer communications and networking conference. Las Vegas, USA pp 266\u2013270","DOI":"10.1109\/CCNC.2012.6181096"},{"key":"11_CR42","doi-asserted-by":"crossref","unstructured":"Gangamohan P, Sudarsana RK, Yegnanarayana B (2013) Analysis of emotional speech at subsegmental level. In: INTERSPEECH. Lyon, France, pp 1916\u20131920","DOI":"10.21437\/Interspeech.2013-20"},{"key":"11_CR43","doi-asserted-by":"crossref","unstructured":"Gangamohan P, Sudarsana RK, Suryakanth VG, Yegnanarayana B (2014) Excitation source features for discrimination of anger and happy emotions. In: INTERSPEECH. Singapore, pp 1253\u20131257","DOI":"10.21437\/Interspeech.2014-314"},{"issue":"2","key":"11_CR44","doi-asserted-by":"publisher","first-page":"132","DOI":"10.1109\/T-AFFC.2010.14","volume":"1","author":"M Gnjatovic","year":"2010","unstructured":"Gnjatovic M, R\u00f6sner D (2010) Inducing genuine emotions in simulated speech-based human-machine interaction: the nimitek corpus. IEEE Trans Affect Comput 1(2):132\u2013144","journal-title":"IEEE Trans Affect Comput"},{"key":"11_CR45","first-page":"123","volume":"1","author":"C Gobl","year":"1988","unstructured":"Gobl C (1988) Voice source dynamics in connected speech. Speech Trans Lab Q Progress Status Rep, KTH 1:123\u2013159","journal-title":"Speech Trans Lab Q Progress Status Rep, KTH"},{"key":"11_CR46","first-page":"9","volume":"4","author":"C Gobl","year":"1989","unstructured":"Gobl C (1989) A preliminary study of acoustic voice quality correlates. Speech Trans Lab Q Progress Status Rep, KTH 4:9\u201321","journal-title":"Speech Trans Lab Q Progress Status Rep, KTH"},{"issue":"4","key":"11_CR47","doi-asserted-by":"publisher","first-page":"481","DOI":"10.1016\/0167-6393(92)90055-C","volume":"11","author":"C Gobl","year":"1992","unstructured":"Gobl C, Chasaide AN (1992) Acoustic characteristics of voice quality. Speech Commun 11(4):481\u2013490","journal-title":"Speech Commun"},{"issue":"1\u20132","key":"11_CR48","doi-asserted-by":"publisher","first-page":"189","DOI":"10.1016\/S0167-6393(02)00082-1","volume":"40","author":"C Gobl","year":"2003","unstructured":"Gobl C, Chasaide AN (2003) The role of voice quality in communicating emotion, mood and attitude. Speech Commun 40(1\u20132):189\u2013212","journal-title":"Speech Commun"},{"issue":"3","key":"11_CR49","doi-asserted-by":"publisher","first-page":"414","DOI":"10.1016\/j.specom.2011.10.005","volume":"54","author":"I Grichkovtsova","year":"2012","unstructured":"Grichkovtsova I, Morel M, Lacheret A (2012) The role of voice quality and prosodic contour in affective speech perception. Speech Commun 54(3):414\u2013429","journal-title":"Speech Commun"},{"issue":"10\u201311","key":"11_CR50","doi-asserted-by":"publisher","first-page":"787","DOI":"10.1016\/j.specom.2007.01.010","volume":"49","author":"M Grimm","year":"2007","unstructured":"Grimm M, Kroschel K, Mower E, Narayanan S (2007) Primitives-based evaluation and estimation of emotions in speech. Speech Commun 49(10\u201311):787\u2013800","journal-title":"Speech Commun"},{"key":"11_CR51","doi-asserted-by":"crossref","unstructured":"Grimm M, Kroschel K, Narayanan S (2008) The Vera am Mittag German audio-visual emotional speech database. In: International conference on multimedia and expo. Hannover, Germany, pp 865\u2013868","DOI":"10.1109\/ICME.2008.4607572"},{"issue":"4","key":"11_CR52","doi-asserted-by":"publisher","first-page":"2061","DOI":"10.1121\/1.3203668","volume":"126","author":"S Guruprasad","year":"2009","unstructured":"Guruprasad S, Yegnanarayana B (2009) Perceived loudness of speech based on the characteristics of glottal excitation source. J Acoust Soc Am 126(4):2061\u20132071","journal-title":"J Acoust Soc Am"},{"issue":"4","key":"11_CR53","doi-asserted-by":"publisher","first-page":"307","DOI":"10.1109\/89.506935","volume":"4","author":"JH Hansen","year":"1996","unstructured":"Hansen JH, Womack BD (1996) Feature analysis and neural network-based classification of speech under stress. IEEE Trans Speech Audio Process 4(4):307\u2013313","journal-title":"IEEE Trans Speech Audio Process"},{"issue":"1","key":"11_CR54","doi-asserted-by":"publisher","first-page":"466","DOI":"10.1121\/1.417991","volume":"101","author":"HM Hanson","year":"1997","unstructured":"Hanson HM (1997) Glottal characteristics of female speakers: acoustic correlates. J Acoust Soc Am 101(1):466\u2013481","journal-title":"J Acoust Soc Am"},{"key":"11_CR55","doi-asserted-by":"crossref","unstructured":"Hassan A, Damper RI (2010) Multi-class and hierarchical SVMs for emotion recognition. In: INTERSPEECH. Chiba, Japan, pp 2354\u20132357","DOI":"10.21437\/Interspeech.2010-644"},{"key":"11_CR56","doi-asserted-by":"crossref","unstructured":"He L, Lech M, Allen N (2010) On the importance of glottal flow spectral energy for the recognition of emotions in speech. In: INTERSPEECH. Chiba, Japan, pp 2346\u20132349","DOI":"10.21437\/Interspeech.2010-642"},{"key":"11_CR57","unstructured":"Hershey JR, Olsen PA (2007) Approximating the Kullback Leibler divergence between Gaussian mixture models. In: ICASSP, vol 4. Montreal, Quebec, Canada, pp 317\u2013320"},{"key":"11_CR58","doi-asserted-by":"crossref","unstructured":"Huber R, Batliner A, Buckow J, N\u00f6th E, Warnke V, Niemann H (2000) Recognition of emotion in a realistic dialogue scenario. In: Proceedings of international conference on spoken language processing. Beijing, China, pp 665\u2013668","DOI":"10.21437\/ICSLP.2000-165"},{"key":"11_CR59","doi-asserted-by":"crossref","unstructured":"H\u00fcbner D, Vlasenko B, Grosser T, Wendemuth A (2010) Determining optimal features for emotion recognition from speech by applying an evolutionary algorithm. In: INTERSPEECH. Chiba, Japan, pp 2358\u20132361","DOI":"10.21437\/Interspeech.2010-645"},{"key":"11_CR60","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4899-2209-0","volume-title":"Human emotions","author":"CE Izard","year":"1977","unstructured":"Izard CE (1977) Human emotions. Plenum Press, New York"},{"key":"11_CR61","doi-asserted-by":"crossref","unstructured":"Jeon JH, Xia R, Liu Y (2011) Sentence level emotion recognition based on decisions from subsentence segments. In: ICASSP. Lyon, France, pp 4940\u20134943","DOI":"10.1109\/ICASSP.2011.5947464"},{"key":"11_CR62","doi-asserted-by":"crossref","unstructured":"Jeon JH, Le D, Xia R, Liu Y (2013) A preliminary study of cross-lingual emotion recognition from speech: automatic classification versus human perception. In: INTERSPEECH. Prague, Czech Republic, pp 2837\u20132840","DOI":"10.21437\/Interspeech.2013-246"},{"key":"11_CR63","doi-asserted-by":"crossref","unstructured":"Joachims T (1998) Text categorization with support vector machines: Learning with many relevant features. In: European conference on machine learning. London, UK, pp 137\u2013142","DOI":"10.1007\/BFb0026683"},{"key":"11_CR64","unstructured":"Kadiri SR, Gangamohan P, Mittal VK, Yegnanarayana B (2014) Naturalistic audio-visual emotion database. In: International conference on natural language processing. Goa, India, pp 127\u2013134"},{"key":"11_CR65","unstructured":"Kadiri SR, Gangamohan P, Yegnanarayana B (2014) Discriminating neutral and emotional speech using neural networks. In: Interenational conference on natural language processing. Goa, India, pp 119\u2013126"},{"key":"11_CR66","doi-asserted-by":"crossref","unstructured":"Kadiri SR, Gangamohan P, Gangashetty SV, Yegnanarayana B (2015) Analysis of excitation source features of speech for emotion recognition. In: INTERSPEECH. Dresden, Germany, pp 1032\u20131036","DOI":"10.21437\/Interspeech.2015-329"},{"key":"11_CR67","doi-asserted-by":"crossref","unstructured":"Keller E (2005) The analysis of voice quality in speech processing. In: G\u00e8rard C, Anna E, Marcos F, Maria M (eds) Lecture notes in computer science. Springer, pp 54\u201373","DOI":"10.1007\/11520153_4"},{"key":"11_CR68","doi-asserted-by":"crossref","unstructured":"Kim W, Hansen JHL (2010) Angry emotion detection from real-life conversational speech by leveraging content structure. In: ICASSP. Dallas, Texas, USA, pp 5166\u20135169","DOI":"10.1109\/ICASSP.2010.5495021"},{"key":"11_CR69","doi-asserted-by":"crossref","unstructured":"Kim J, Lee S, Narayanan S (2010) An exploratory study of manifolds of emotional speech. In: ICASSP. Dallas, Texas, USA, pp 5142\u20135145","DOI":"10.1109\/ICASSP.2010.5495032"},{"key":"11_CR70","doi-asserted-by":"crossref","unstructured":"Kim J, Park J, Oh Y (2011) On-line speaker adaptation based emotion recognition using incremental emotional information. In: ICASSP. Prague, Czech Republic, pp 4948\u20134951","DOI":"10.1109\/ICASSP.2011.5947466"},{"key":"11_CR71","unstructured":"Klasmeyer G, Sendlmeier WF (2000) Voice and emotional states. In: Voice quality measurement. Springer, Berlin, Germany, pp 339\u2013358"},{"issue":"3","key":"11_CR72","doi-asserted-by":"publisher","first-page":"971","DOI":"10.1121\/1.383940","volume":"67","author":"DH Klatt","year":"1980","unstructured":"Klatt DH (1980) Software for a cascade\/parallel formant synthesizer. J Acoust Soc Am 67(3):971\u2013995","journal-title":"J Acoust Soc Am"},{"issue":"2","key":"11_CR73","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1007\/s10772-011-9125-1","volume":"15","author":"SG Koolagudi","year":"2012","unstructured":"Koolagudi SG, Sreenivasa Rao K (2012) Emotion recognition from speech: a review. Int J Speech Technol 15(2):99\u2013117","journal-title":"Int J Speech Technol"},{"key":"11_CR74","first-page":"485","volume-title":"Communications in Computer and Information Science","author":"Shashidhar G. Koolagudi","year":"2009","unstructured":"Koolagudi SG, Maity S, Vuppala AK, Chakrabarti S, Sreenivasa Rao K (2009) IITKGP-SESC: speech database for emotion analysis. In: Communications in computer and information science, pp 485\u2013492"},{"issue":"1","key":"11_CR75","doi-asserted-by":"publisher","first-page":"43","DOI":"10.3109\/13682826809011440","volume":"3","author":"DM Laver John","year":"1968","unstructured":"Laver John DM (1968) Voice quality and indexical information. Int J Lang Commun Disord 3(1):43\u201354","journal-title":"Int J Lang Commun Disord"},{"issue":"9\u201310","key":"11_CR76","doi-asserted-by":"publisher","first-page":"1162","DOI":"10.1016\/j.specom.2011.06.004","volume":"53","author":"C Lee","year":"2011","unstructured":"Lee C, Mower E, Busso C, Lee S, Narayanan S (2011) Emotion recognition using a hierarchical binary decision tree approach. Speech Commun 53(9\u201310):1162\u20131171","journal-title":"Speech Commun"},{"issue":"2","key":"11_CR77","doi-asserted-by":"publisher","first-page":"293","DOI":"10.1109\/TSA.2004.838534","volume":"13","author":"CM Lee","year":"2005","unstructured":"Lee CM, Narayanan S (2005) Toward detecting emotions in spoken dialogs. IEEE Trans Speech Audio Process 13(2):293\u2013303","journal-title":"IEEE Trans Speech Audio Process"},{"key":"11_CR78","doi-asserted-by":"crossref","unstructured":"Lee CM, Yildirim S, Bulut M, Kazemzadeh A, Busso C, Deng Z, Lee S, Narayanan S (2004) Emotion recognition based on phoneme classes. In: INTERSPEECH. JejuIsland, Korea, pp 205\u2013211","DOI":"10.21437\/Interspeech.2004-322"},{"issue":"7","key":"11_CR79","doi-asserted-by":"publisher","first-page":"922","DOI":"10.1121\/1.1918222","volume":"34","author":"P Lieberman","year":"1962","unstructured":"Lieberman P, Michaels SB (1962) Some aspects of fundamental frequency and envelope amplitude as related to the emotional content of speech. J Acoust Soc Am 34(7):922\u2013927","journal-title":"J Acoust Soc Am"},{"key":"11_CR80","doi-asserted-by":"crossref","unstructured":"Lin J, Wu C, Wei W (2013) Emotion recognition of conversational affective speech using temporal course modeling. In: INTERSPEECH. Lyon, France, pp 1336\u20131340","DOI":"10.21437\/Interspeech.2013-354"},{"key":"11_CR81","doi-asserted-by":"crossref","unstructured":"Luengo I, Navas E, Hern\u00e1ez I, S\u00e1nchez J (2005) Automatic emotion recognition using prosodic parameters. In: INTERSPEECH. Lisbon, Portugal, pp 493\u2013496","DOI":"10.21437\/Interspeech.2005-324"},{"key":"11_CR82","doi-asserted-by":"crossref","unstructured":"Lugger M, Yang B (2007) The relevance of voice quality features in speaker independent emotion recognition. In: ICASSP, vol 4. Honolulu, Hawaii, USA, pp 17\u201320","DOI":"10.1109\/ICASSP.2007.367152"},{"key":"11_CR83","doi-asserted-by":"publisher","first-page":"561","DOI":"10.1109\/PROC.1975.9792","volume":"63","author":"J Makhoul","year":"1975","unstructured":"Makhoul J (1975) Linear prediction: a tutorial review. Proc IEEE 63:561\u2013580","journal-title":"Proc IEEE"},{"key":"11_CR84","unstructured":"Mansoorizadeh M, Charkari NM (2007) Speech emotion recognition: comparison of speech segmentation approaches. In: Proceedings of IKT, Mashad, Iran"},{"key":"11_CR85","unstructured":"McGilloway S, Cowie R, Douglas-Cowie E, Gielen S, Westerdijk M, Stroeve S (2000) Approaching automatic recognition of emotion from voice: a rough benchmark. In: ISCA tutorial and research workshop (ITRW) on speech and emotion. Newcastle, Northern Ireland, UK"},{"issue":"5","key":"11_CR86","doi-asserted-by":"publisher","first-page":"3050","DOI":"10.1121\/1.4796110","volume":"133","author":"VK Mittal","year":"2013","unstructured":"Mittal VK, Yegnanarayana B (2013) Effect of glottal dynamics in the production of shouted speech. J Acoust Soc Am 133(5):3050\u20133061","journal-title":"J Acoust Soc Am"},{"issue":"2","key":"11_CR87","doi-asserted-by":"publisher","first-page":"98","DOI":"10.1016\/j.specom.2006.11.004","volume":"49","author":"D Morrison","year":"2007","unstructured":"Morrison D, Wang R, De Silva LC (2007) Ensemble methods for spoken emotion recognition in call-centres. Speech Commun 49(2):98\u2013112","journal-title":"Speech Commun"},{"issue":"2","key":"11_CR88","doi-asserted-by":"publisher","first-page":"1097","DOI":"10.1121\/1.405558","volume":"93","author":"IR Murray","year":"1993","unstructured":"Murray IR, Arnott JL (1993) Toward the simulation of emotion in synthetic speech: a review of the literature on human vocal emotion. J Acoust Soc Am 93(2):1097\u20131108","journal-title":"J Acoust Soc Am"},{"issue":"8","key":"11_CR89","doi-asserted-by":"crossref","first-page":"1602","DOI":"10.1109\/TASL.2008.2004526","volume":"16","author":"K. Sri Rama Murty","year":"2008","unstructured":"Murty KSR, Yegnanarayana B (2008) Epoch extraction from speech signals. IEEE Trans Audio Speech Lang Process 16(8):1602\u20131613","journal-title":"IEEE Transactions on Audio, Speech, and Language Processing"},{"key":"11_CR90","doi-asserted-by":"crossref","unstructured":"Nogueiras A, Moreno A, Bonafonte A, Mari\u00f1o JB (2001) Speech emotion recognition using hidden Markov models. In: EUROSPEECH. Aalborg, Denmark, pp 2679\u20132682","DOI":"10.21437\/Eurospeech.2001-627"},{"issue":"4","key":"11_CR91","doi-asserted-by":"publisher","first-page":"603","DOI":"10.1016\/S0167-6393(03)00099-2","volume":"41","author":"TL Nwe","year":"2003","unstructured":"Nwe TL, Foo SW, De Silva LC (2003) Speech emotion recognition using hidden Markov models. Speech Commun 41(4):603\u2013623","journal-title":"Speech Commun"},{"key":"11_CR92","first-page":"33","volume":"123","author":"K Oatley","year":"1989","unstructured":"Oatley K (1989) The importance of being emotional. New Sci 123:33\u201336","journal-title":"New Sci"},{"key":"11_CR93","unstructured":"Pereira C (2000) Dimensions of emotional meaning in speech. In: ISCA tutorial and research workshop (ITRW) on speech and emotion. Northern Ireland, UK"},{"key":"11_CR94","doi-asserted-by":"crossref","unstructured":"Polzehl T, Sundaram S, Ketabdar H, Wagner M, Metze F (2009) Emotion classification in children\u2019s speech using fusion of acoustic and linguistic features. In: INTERSPEECH. Brighton, UK, pp 340\u2013343","DOI":"10.21437\/Interspeech.2009-110"},{"key":"11_CR95","unstructured":"Prasanna SRM, Govind D (2010) Analysis of excitation source information in emotional speech. In: INTERSPEECH. Chiba, Japan, pp 781\u2013784"},{"issue":"6","key":"11_CR96","doi-asserted-by":"publisher","first-page":"1632","DOI":"10.1121\/1.1913513","volume":"53","author":"M Rothenberg","year":"1973","unstructured":"Rothenberg M (1973) A new inverse-filtering technique for deriving the glottal air flow waveform during voicing. J Acoust Soc Am 53(6):1632\u20131645","journal-title":"J Acoust Soc Am"},{"key":"11_CR97","doi-asserted-by":"crossref","unstructured":"Rozgic V, Ananthakrishnan S, Saleem S, Kumar R, Vembu AN, Prasad R (2012) Emotion recognition using acoustic and lexical features. In: INTERSPEECH. Portland, USA","DOI":"10.21437\/Interspeech.2012-118"},{"key":"11_CR98","volume-title":"Speech evaluation in psychiatry","author":"KR Scherer","year":"1981","unstructured":"Scherer KR (1981) Speech and emotional states. In: Darby JK (ed) Speech evaluation in psychiatry. Grune and Stratton, New York"},{"key":"11_CR99","volume-title":"Approaches to emotion","author":"KR Scherer","year":"1984","unstructured":"Scherer KR (1984) On the nature and function of emotion: a component process approach. In: Scherer KR, Ekman P (eds) Approaches to emotion. Lawrence Elbraum, Hillsdale"},{"issue":"1\u20132","key":"11_CR100","doi-asserted-by":"publisher","first-page":"227","DOI":"10.1016\/S0167-6393(02)00084-5","volume":"40","author":"KR Scherer","year":"2003","unstructured":"Scherer KR (2003) Vocal communication of emotion: a review of research paradigms. Speech Commun 40(1\u20132):227\u2013256","journal-title":"Speech Commun"},{"issue":"6","key":"11_CR101","doi-asserted-by":"crossref","first-page":"497","DOI":"10.1037\/h0061489","volume":"29","author":"H. Scholsberg","year":"1941","unstructured":"Scholsberg H (1941) A scale for the judgment of facial expressions. J Exp Psychol 29(6):497\u2013510","journal-title":"Journal of Experimental Psychology"},{"issue":"2","key":"11_CR102","doi-asserted-by":"crossref","first-page":"81","DOI":"10.1037\/h0054570","volume":"61","author":"Harold Schlosberg","year":"1954","unstructured":"Schlosberg H (1954) Three dimensions of emotion. J Psychol Rev 61(2):81\u201388","journal-title":"Psychological Review"},{"key":"11_CR103","doi-asserted-by":"crossref","unstructured":"Schr\u00f6der M (2001) Emotional speech synthesis-a review. In: INTERSPEECH. Aalborg,Denmark, pp 561\u2013564","DOI":"10.21437\/Eurospeech.2001-150"},{"key":"11_CR104","unstructured":"Schr\u00f6der M (2004) Speech and emotion research: an overview of research frameworks and a dimensional approach to emotional speech synthesis. PhD thesis, Saarland University"},{"key":"11_CR105","doi-asserted-by":"crossref","unstructured":"Schr\u00f6der M, Cowie R, Douglas-Cowie E, Westerdijk M, Gielen SC (2001) Acoustic correlates of emotion dimensions in view of speech synthesis. In: INTERSPEECH. Aalborg, Denmark, pp 87\u201390","DOI":"10.21437\/Eurospeech.2001-34"},{"issue":"4","key":"11_CR106","doi-asserted-by":"publisher","first-page":"192","DOI":"10.1109\/T-AFFC.2011.17","volume":"2","author":"B Schuller","year":"2011","unstructured":"Schuller B (2011) Recognizing affect from linguistic information in 3D continuous space. IEEE Trans Affect Comput 2(4):192\u2013205","journal-title":"IEEE Trans Affect Comput"},{"key":"11_CR107","doi-asserted-by":"crossref","unstructured":"Schuller B, Rigoll G (2006) Timing levels in segment-based speech emotion recognition. In: INTERSPEECH. Pittsburgh, Pennsylvania, pp 17\u201321","DOI":"10.21437\/Interspeech.2006-502"},{"key":"11_CR108","doi-asserted-by":"crossref","unstructured":"Schuller B, Rigoll G, Lang M (2004) Speech emotion recognition combining acoustic features and linguistic information in a hybrid support vector machine-belief network architecture. In: ICASSP vol 1. Montreal, Quebec, Canada, pp 577\u2013580","DOI":"10.1109\/ICASSP.2004.1326051"},{"key":"11_CR109","doi-asserted-by":"crossref","unstructured":"Schuller B, M\u00fcller R, Lang M, Rigoll G (2005) Speaker independent emotion recognition by early fusion of acoustic and linguistic features within ensembles. In: INTERSPEECH. Lisbon, Portugal, pp 805\u2013808","DOI":"10.21437\/Interspeech.2005-379"},{"key":"11_CR110","doi-asserted-by":"crossref","unstructured":"Schuller B, Villar RJ, Rigoll G, Lang MK (2005) Meta-classifiers in acoustic and linguistic feature fusion-based affect recognition. In: ICASSP. Philadelphia, Pennsylvania, USA, pp 325\u2013328","DOI":"10.1109\/ICASSP.2005.1415116"},{"key":"11_CR111","doi-asserted-by":"crossref","unstructured":"Schuller B, Batliner A, Steidl S, Seppi D (2009) Emotion recognition from speech: putting ASR in the loop. In: ICASSP. Taipei, Taiwan, pp 4585\u20134588","DOI":"10.1109\/ICASSP.2009.4960651"},{"issue":"9\u201310","key":"11_CR112","doi-asserted-by":"publisher","first-page":"1062","DOI":"10.1016\/j.specom.2011.01.011","volume":"53","author":"B Schuller","year":"2011","unstructured":"Schuller B, Batliner A, Steidl S, Seppi D (2011) Recognising realistic emotions and affect in speech: state of the art and lessons learnt from the first challenge. Speech Commun 53(9\u201310):1062\u20131087","journal-title":"Speech Commun"},{"issue":"2","key":"11_CR113","doi-asserted-by":"publisher","first-page":"119","DOI":"10.1109\/T-AFFC.2010.8","volume":"1","author":"B Schuller","year":"2010","unstructured":"Schuller B, Vlasenko B, Eyben F, Wollmer M, Stuhlsatz A, Wendemuth A, Rigoll G (2010) Cross-corpus acoustic emotion recognition: variances and strategies. IEEE Trans Affect Comput 1(2):119\u2013131","journal-title":"IEEE Trans Affect Comput"},{"issue":"3","key":"11_CR114","doi-asserted-by":"publisher","first-page":"201","DOI":"10.1016\/j.specom.2007.01.006","volume":"49","author":"M Shami","year":"2007","unstructured":"Shami M, Verhelst W (2007) An evaluation of the robustness of existing supervised machine learning approaches to the classification of emotions in speech. Speech Commun 49(3):201\u2013212","journal-title":"Speech Commun"},{"issue":"6","key":"11_CR115","doi-asserted-by":"crossref","first-page":"1061","DOI":"10.1037\/0022-3514.52.6.1061","volume":"52","author":"Phillip Shaver","year":"1987","unstructured":"Shaver P, Schwartz J, kirson D, O\u2019Connor C (1987) Emotion, knowledge: further exploration of a prototype approach. J Personal Soc Psychol 52:1061\u20131086","journal-title":"Journal of Personality and Social Psychology"},{"issue":"1","key":"11_CR116","doi-asserted-by":"publisher","first-page":"32","DOI":"10.1109\/T-AFFC.2011.26","volume":"3","author":"I Sneddon","year":"2012","unstructured":"Sneddon I, McRorie M, McKeown G, Hanratty J (2012) The Belfast induced natural emotion database. IEEE Trans Affect Comput 3(1):32\u201341","journal-title":"IEEE Trans Affect Comput"},{"key":"11_CR117","unstructured":"Steidl S (2009) Automatic classification of emotion related user states in spontaneous children\u2019s speech. PhD thesis, Universit\u00e4t Erlangen-N\u00fcrnberg, Germany"},{"issue":"1","key":"11_CR118","doi-asserted-by":"crossref","first-page":"783954","DOI":"10.1186\/1687-4722-2010-783954","volume":"2010","author":"Stefan Steidl","year":"2010","unstructured":"Steidl S, Batliner A, Seppi D, Schuller B (2010) On the impact of children\u2019s emotional speech on acoustic and language models. EURASIP J Audio, Speech, and Music Processing","journal-title":"EURASIP Journal on Audio, Speech, and Music Processing"},{"key":"11_CR119","doi-asserted-by":"publisher","first-page":"161","DOI":"10.1080\/02699939208411067","volume":"6","author":"N Stein","year":"1992","unstructured":"Stein N, Oatley K (1992) Basic emotions: theory and measurement. Cognit Emot 6:161\u2013168","journal-title":"Cognit Emot"},{"key":"11_CR120","doi-asserted-by":"crossref","unstructured":"Sun R, Moore II E (2012) A preliminary study on cross-databases emotion recognition using the glottal features in speech. In: INTERSPEECH. Portland, USA, pp 1628\u20131631","DOI":"10.21437\/Interspeech.2012-447"},{"key":"11_CR121","doi-asserted-by":"crossref","unstructured":"Sun R, Moore II E, Torres JF (2009) Investigating glottal parameters for differentiating emotional categories with similar prosodics. In: ICASSP. Taipei, Taiwan, pp 4509\u20134512","DOI":"10.1109\/ICASSP.2009.4960632"},{"issue":"3","key":"11_CR122","doi-asserted-by":"publisher","first-page":"162","DOI":"10.1109\/T-AFFC.2011.14","volume":"2","author":"J Sundberg","year":"2011","unstructured":"Sundberg J, Patel S, Bjorkner E, Scherer KR (2011) Interdependencies among voice source parameters in emotional speech. IEEE Trans Affect Comput 2(3):162\u2013174","journal-title":"IEEE Trans Affect Comput"},{"key":"11_CR123","doi-asserted-by":"crossref","unstructured":"Tahon M, Degottex G, Devillers L (2012) Usual voice quality features and glottal features for emotional valence detection. In: Speech Prosody. Shanghai, China, pp 693\u2013696","DOI":"10.21437\/SpeechProsody.2012-173"},{"key":"11_CR124","volume-title":"Principles of voice production","author":"IR Titze","year":"1994","unstructured":"Titze IR (1994) Principles of voice production. Prentice-Hall, Englewood Cliffs"},{"issue":"9","key":"11_CR125","doi-asserted-by":"publisher","first-page":"1049","DOI":"10.1016\/j.specom.2012.04.006","volume":"54","author":"P Truong Khiet","year":"2012","unstructured":"Truong Khiet P, van Leeuwen David A, de Jong Franciska M G (2012) Speech-based recognition of self-reported and observed emotion in a dimensional space. Speech Commun 54(9):1049\u20131063","journal-title":"Speech Commun"},{"key":"11_CR126","unstructured":"Ververidis D, Kotropoulos C (2003) A review of emotional speech databases. In: Proceedings of panhellenic conference on informatics (PCI). Thessaloniki, Greece, pp 560\u2013574"},{"key":"11_CR127","doi-asserted-by":"crossref","unstructured":"Ververidis D, Kotropoulos C (2005) Emotional speech classification using Gaussian mixture models. In: International symposium on circuits and systems. Kobe, Japan, pp 2871\u20132874","DOI":"10.1109\/ISCAS.2005.1465226"},{"key":"11_CR128","doi-asserted-by":"crossref","unstructured":"Vlasenko B, Prylipko D, Philippou-H\u00fcbner D, Wendemuth A (2011) Vowels formants analysis allows straightforward detection of high arousal acted and spontaneous emotions. In: INTERSPEECH. Florence, Italy, pp 1577\u20131580","DOI":"10.21437\/Interspeech.2011-474"},{"key":"11_CR129","doi-asserted-by":"crossref","unstructured":"Vroomen J, Collier R, Mozziconacci S (1993) Duration and intonation in emotional speech. In: EUROSPEECH, vol 1. Berlin, Germany, pp 577\u2013580","DOI":"10.21437\/Eurospeech.1993-136"},{"issue":"3","key":"11_CR130","doi-asserted-by":"publisher","first-page":"1523","DOI":"10.1121\/1.427149","volume":"106","author":"G \u0160vec Jan","year":"1999","unstructured":"\u0160vec Jan G, Schutte Harm K, Miller Donald G (1999) On pitch jumps between chest and falsetto registers in voice: data from living and excised human larynges. J Acoust Soc Am 106(3):1523\u20131531","journal-title":"J Acoust Soc Am"},{"issue":"1","key":"11_CR131","doi-asserted-by":"publisher","first-page":"30","DOI":"10.1016\/j.jvoice.2008.04.004","volume":"24","author":"T Waaramaa","year":"2010","unstructured":"Waaramaa T, Laukkanen AM, Airas M, Alku P (2010) Perception of emotional valences and activity levels from vowel segments of continuous speech. J Voice 24(1):30\u201338","journal-title":"J Voice"},{"key":"11_CR132","first-page":"1369","volume":"40","author":"CE Williams","year":"1969","unstructured":"Williams CE, Stevens KN (1969) On determining the emotional state of pilots during flight: an exploratory study. Aerosp Med 40:1369\u20131372","journal-title":"Aerosp Med"},{"issue":"2","key":"11_CR133","doi-asserted-by":"publisher","first-page":"1238","DOI":"10.1121\/1.1913238","volume":"52","author":"CE Williams","year":"1972","unstructured":"Williams CE, Stevens KN (1972) Emotions and speech: some acoustical correlates. J Acoust Soc Am 52(2):1238\u20131250","journal-title":"J Acoust Soc Am"},{"issue":"5","key":"11_CR134","doi-asserted-by":"publisher","first-page":"768","DOI":"10.1016\/j.specom.2010.08.013","volume":"53","author":"S Wu","year":"2011","unstructured":"Wu S, Falk TH, Chan W (2011) Automatic speech emotion recognition using modulation spectral features. Speech Commun 53(5):768\u2013785","journal-title":"Speech Commun"},{"issue":"6","key":"11_CR135","doi-asserted-by":"publisher","first-page":"782","DOI":"10.1016\/j.specom.2013.02.007","volume":"55","author":"B Yegnanarayana","year":"2013","unstructured":"Yegnanarayana B, Dhananjaya N (2013) Spectro-temporal analysis of speech signals using zero-time windowing and group delay function. Speech Commun 55(6):782\u2013795","journal-title":"Speech Commun"},{"issue":"4","key":"11_CR136","doi-asserted-by":"publisher","first-page":"614","DOI":"10.1109\/TASL.2008.2012194","volume":"17","author":"B Yegnanarayana","year":"2009","unstructured":"Yegnanarayana B, Murty KSR (2009) Event-based instantaneous fundamental frequency estimation from speech signals. IEEE Trans Audio Speech Lang Process 17(4):614\u2013624","journal-title":"IEEE Trans Audio Speech Lang Process"},{"key":"11_CR137","doi-asserted-by":"crossref","unstructured":"Yeh L, Chi T (2010) Spectro-temporal modulations for robust speech emotion recognition. In: INTERSPEECH. Chiba, Japan, pp 789\u2013792","DOI":"10.21437\/Interspeech.2010-286"},{"issue":"1","key":"11_CR138","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1109\/TPAMI.2008.52","volume":"31","author":"Z Zeng","year":"2009","unstructured":"Zeng Z, Pantic M, Roisman GI, Huang TS (2009) A survey of affect recognition methods: Audio, visual, and spontaneous expressions. IEEE Trans Pattern Anal Mach Intell 31(1):39\u201358","journal-title":"IEEE Trans Pattern Anal Mach Intell"}],"container-title":["Intelligent Systems Reference Library","Toward Robotic Socially Believable Behaving Systems - Volume I"],"original-title":[],"link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-319-31056-5_11","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,6,15]],"date-time":"2024-06-15T04:41:21Z","timestamp":1718426481000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-319-31056-5_11"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2016]]},"ISBN":["9783319310558","9783319310565"],"references-count":138,"URL":"https:\/\/doi.org\/10.1007\/978-3-319-31056-5_11","relation":{},"ISSN":["1868-4394","1868-4408"],"issn-type":[{"value":"1868-4394","type":"print"},{"value":"1868-4408","type":"electronic"}],"subject":[],"published":{"date-parts":[[2016]]}}}