{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T16:33:50Z","timestamp":1775579630585,"version":"3.50.1"},"reference-count":121,"publisher":"IEEE","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2011,3]]},"DOI":"10.1109\/fg.2011.5771357","type":"proceedings-article","created":{"date-parts":[[2011,5,24]],"date-time":"2011-05-24T19:14:25Z","timestamp":1306264465000},"page":"827-834","source":"Crossref","is-referenced-by-count":198,"title":["Emotion representation, analysis and synthesis in continuous space: A survey"],"prefix":"10.1109","author":[{"given":"Hatice","family":"Gunes","sequence":"first","affiliation":[]},{"given":"Bjorn","family":"Schuller","sequence":"additional","affiliation":[]},{"given":"Maja","family":"Pantic","sequence":"additional","affiliation":[]},{"given":"Roddy","family":"Cowie","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.0507650102"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1007\/11527886_8"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1080\/02699930302294"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1037\/0022-3514.71.3.421"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1177\/0146167299025002007"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.2466\/pms.1967.24.3.711"},{"key":"ref37","first-page":"48","article-title":"Recognizing affective dimensions from body posture","author":"kleinsmith","year":"0","journal-title":"Proc ACII"},{"key":"ref36","first-page":"2362","article-title":"Context-sensitive multimodal emotion recognition from speech and facial expression using bidirectional lstm modeling","author":"wollmer","year":"0","journal-title":"Proc INTERSPEECH"},{"key":"ref35","first-page":"42","article-title":"The emotional and communicative significance of head nods and shakes in a naturalistic database","author":"cowie","year":"0","journal-title":"Proc LREC Int'l Workshop Emotion"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1016\/j.biopsycho.2003.07.005"},{"key":"ref28","author":"schroder","year":"2003","journal-title":"Speech and emotion research An overview of research frameworks and a dimensional approach to emotional speech synthesis"},{"key":"ref27","first-page":"86","article-title":"The prosody of excitement in horse race commentaries","author":"trouvain","year":"0","journal-title":"Proc ISCA workshop Speech Emotion"},{"key":"ref29","first-page":"1","article-title":"Perception of non-verbal emotional listener feedback","author":"schroder","year":"2006","journal-title":"Speech Prosody"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-006-6106-y"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/79.911197"},{"key":"ref21","first-page":"19","article-title":"Feeltrace: An instrument for recording perceived emotion in real time","author":"cowie","year":"0","journal-title":"Proc ISCA workshop Speech Emotion"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1044\/jshr.1103.481"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/T-AFFC.2010.1"},{"key":"ref101","first-page":"91","article-title":"Adaptive expressiveness virtual conversational agents that can align to their interaction partner","author":"buschmeier","year":"0","journal-title":"Proc 1st Int Conf Autonomous Agents Multiagent Syst"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1007\/BF00992539"},{"key":"ref100","doi-asserted-by":"publisher","DOI":"10.1109\/ACII.2009.5349530"},{"key":"ref25","first-page":"101","article-title":"The Communication of Emotional Meaning","author":"davitz","year":"1964","journal-title":"chapter Auditory correlates of vocal expression of emotional feeling"},{"key":"ref50","first-page":"185","article-title":"Affective Computing: Focus on Emotion Expression, Synthesis, and Recognition","author":"gunes","year":"2008","journal-title":"Ch From the Lab to the Real World Affect Recognition using Multiple Cues and Modalities"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2008.52"},{"key":"ref59","doi-asserted-by":"publisher","DOI":"10.5772\/4755"},{"key":"ref58","first-page":"597","article-title":"Abandoning emotion classes-towards continuous emotion recognition with modelling of long-range dependencies","author":"wollmer","year":"0","journal-title":"Proc INTERSPEECH"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1109\/JSTSP.2010.2057200"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1109\/ASRU.2009.5372886"},{"key":"ref55","doi-asserted-by":"crossref","first-page":"2067","DOI":"10.1109\/TPAMI.2008.26","article-title":"Emotion recognition based on physiological changes in music listening","volume":"30","author":"kim","year":"2008","journal-title":"IEEE Trans on Pattern Analysis & Machine Intelligence"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1109\/TITB.2009.2038481"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR.2010.1031"},{"key":"ref52","first-page":"1","article-title":"Emotion-aware technologies for consumer electronics","author":"gu","year":"0","journal-title":"Proc IEEE Int Symp on Consumer Electronics"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1007\/11821830_8"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1007\/BF02686918"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1037\/h0077714"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/TASL.2010.2052246"},{"key":"ref5","doi-asserted-by":"crossref","DOI":"10.1093\/oso\/9780195130072.001.0001","author":"scherer","year":"2001","journal-title":"Appraisal Processes in Emotion Theory Methods Research"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1111\/j.1467-9280.2007.02024.x"},{"key":"ref7","first-page":"5138","article-title":"Features selection for primitives estimation on emotional speech","author":"espinosa","year":"0","journal-title":"Proc IEEE ICASSP"},{"key":"ref49","article-title":"Special issue on naturalistic affect resources for system building and evaluation","author":"schuller","year":"2011","journal-title":"IEEE Tran on Affective Computing"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/ICME.2010.5583006"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/ACII.2009.5349594"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/ACII.2009.5349539"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1109\/ICME.2008.4607572"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-15184-2_14"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijhcs.2009.03.005"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN.2009.5178854"},{"key":"ref44","article-title":"The cognitive psychophysiology of Emotion: Anxiety and the anxiety disorders","author":"lang","year":"1985"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/TRO.2007.904899"},{"key":"ref73","article-title":"Audio-visual emotion recognition using an emotion recognition space concept","author":"kanluan","year":"0","journal-title":"Proc European Signal Processing Conf"},{"key":"ref72","doi-asserted-by":"publisher","DOI":"10.1145\/1180995.1181029"},{"key":"ref71","doi-asserted-by":"publisher","DOI":"10.1109\/IEMBS.2007.4352270"},{"key":"ref70","doi-asserted-by":"publisher","DOI":"10.1109\/TBME.2009.2035926"},{"key":"ref76","doi-asserted-by":"publisher","DOI":"10.1111\/j.1469-8986.1997.tb02140.x"},{"key":"ref77","first-page":"39","article-title":"Ultra short term analysis of heart rate variability for monitoring mental stress in mobile settings","author":"salahuddin","year":"0","journal-title":"Proc IEEE Int Conf of EMBS"},{"key":"ref74","doi-asserted-by":"publisher","DOI":"10.1109\/FG.2011.5771417"},{"key":"ref75","doi-asserted-by":"publisher","DOI":"10.1109\/ACII.2009.5349552"},{"key":"ref78","first-page":"219","article-title":"Emotion recognition from electromyography and skin conductance","author":"nakasone","year":"0","journal-title":"Proc 3rd Int Workshop on Biosignal Interpretation"},{"key":"ref79","doi-asserted-by":"publisher","DOI":"10.1145\/1647314.1647321"},{"key":"ref60","first-page":"2794","article-title":"The interspeech 2010 paralinguistic challenge","author":"schuller","year":"0","journal-title":"Proc INTERSPEECH"},{"key":"ref62","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2010.5543833"},{"key":"ref61","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR.2010.900"},{"key":"ref63","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2008.4563173"},{"key":"ref64","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2005.03.004"},{"key":"ref65","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-15892-6_39"},{"key":"ref66","doi-asserted-by":"publisher","DOI":"10.1109\/ACII.2009.5349544"},{"key":"ref67","doi-asserted-by":"publisher","DOI":"10.1109\/FG.2011.5771396"},{"key":"ref68","first-page":"43","article-title":"Automatic segmentation of spontaneous data using dimensional labels from multiple coders","author":"nicolaou","year":"0","journal-title":"Proc Int Workshop Multimodal Corpora"},{"key":"ref69","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2008.2004911"},{"key":"ref2","author":"ekman","year":"1975","journal-title":"Unmasking the Face A Guide to Recognizing Emotions From Facial Clues"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1016\/j.concog.2008.03.019"},{"key":"ref109","doi-asserted-by":"publisher","DOI":"10.1109\/ROMAN.2010.5598640"},{"key":"ref95","doi-asserted-by":"publisher","DOI":"10.1109\/MMUL.2003.1237553"},{"key":"ref108","doi-asserted-by":"publisher","DOI":"10.1109\/ROMAN.2010.5598649"},{"key":"ref94","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-540-44972-0_17"},{"key":"ref107","doi-asserted-by":"crossref","first-page":"67","DOI":"10.1109\/MSP.2009.932562","article-title":"Sensitive talking heads","volume":"26","author":"huang","year":"0","journal-title":"IEEE Signal Processing Magazine"},{"key":"ref93","doi-asserted-by":"publisher","DOI":"10.1007\/s10055-005-0153-5"},{"key":"ref106","doi-asserted-by":"publisher","DOI":"10.1109\/TITB.2010.2042608"},{"key":"ref92","doi-asserted-by":"publisher","DOI":"10.1111\/1467-8659.t01-1-00645"},{"key":"ref105","first-page":"1459","article-title":"opensmile-the munich versatile and fast open-source audio feature extractor","author":"eyben","year":"0","journal-title":"Proc ACM Multimedia"},{"key":"ref91","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2008.2001355"},{"key":"ref104","doi-asserted-by":"publisher","DOI":"10.1155\/2010\/319406"},{"key":"ref90","first-page":"561","article-title":"Emotional speech synthesis:a review","volume":"1","author":"schroder","year":"0","journal-title":"Proc EUROSPEECH"},{"key":"ref103","first-page":"1","article-title":"Facial feedback signals for ecas","author":"bevacqua","year":"0","journal-title":"Proc Artificial & Ambient Intelligence"},{"key":"ref102","doi-asserted-by":"publisher","DOI":"10.1098\/rstb.2009.0186"},{"key":"ref111","doi-asserted-by":"publisher","DOI":"10.1109\/MDM.2009.78"},{"key":"ref112","doi-asserted-by":"publisher","DOI":"10.1109\/BSN.2009.21"},{"key":"ref110","doi-asserted-by":"publisher","DOI":"10.1109\/ICVR.2009.5174225"},{"key":"ref98","doi-asserted-by":"publisher","DOI":"10.1109\/ICNC.2010.5584720"},{"key":"ref99","doi-asserted-by":"publisher","DOI":"10.1109\/ACII.2009.5349476"},{"key":"ref96","doi-asserted-by":"publisher","DOI":"10.1109\/TASL.2006.885910"},{"key":"ref97","doi-asserted-by":"publisher","DOI":"10.1109\/ACII.2009.5349579"},{"key":"ref10","author":"frijda","year":"1986","journal-title":"The Emotions"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2005.03.001"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1017\/CBO9780511571299"},{"key":"ref13","first-page":"39","article-title":"Integrating the occ model of emotions in embodied characters","author":"bartneck","year":"2002"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.4018\/jse.2010101605"},{"key":"ref15","article-title":"Affective agents: Effects of agent affect on arousal, attention, liking and learning","author":"dietz","year":"0","journal-title":"Proc Cognitive Technology"},{"key":"ref118","doi-asserted-by":"publisher","DOI":"10.1007\/11558651_18"},{"key":"ref82","doi-asserted-by":"publisher","DOI":"10.1109\/ICME.2006.262725"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/ICSMC.2007.4413638"},{"key":"ref117","doi-asserted-by":"publisher","DOI":"10.1109\/ACII.2009.5349526"},{"key":"ref81","first-page":"79","article-title":"A corpus-based approach to iahem\/i expressive speech synthesis","author":"eide","year":"0","journal-title":"Proc IEEE Speech Synthesis Workshop"},{"key":"ref17","first-page":"36","article-title":"Emotion recognition using bio-sensors: First steps towards an automatic system","author":"haag","year":"2004","journal-title":"LNCS 3068"},{"key":"ref84","doi-asserted-by":"publisher","DOI":"10.1109\/ICSPS.2010.5555658"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/TNSRE.2006.875544"},{"key":"ref119","year":"2010"},{"key":"ref83","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2007.367300"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1126\/science.7146906"},{"key":"ref114","doi-asserted-by":"publisher","DOI":"10.1155\/2010\/263593"},{"key":"ref113","doi-asserted-by":"publisher","DOI":"10.1109\/IA.2009.4927496"},{"key":"ref116","doi-asserted-by":"publisher","DOI":"10.1109\/ICME.2009.5202772"},{"key":"ref80","article-title":"Multilingual voice creation toolkit for the mary tts platform","author":"pammi","year":"0","journal-title":"Proc LREC"},{"key":"ref115","first-page":"566","article-title":"An improved valence-arousal emotion space for video affective content representation and recognition","author":"sun","year":"0","journal-title":"Proc IEEE ICME"},{"key":"ref120","doi-asserted-by":"publisher","DOI":"10.1109\/MIS.2006.83"},{"key":"ref89","doi-asserted-by":"publisher","DOI":"10.1002\/cav.32"},{"key":"ref121","article-title":"Special issue on ethics & affective computing","author":"beavers","year":"2011","journal-title":"IEEE Trans on Affective Computing"},{"key":"ref85","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2008.4518688"},{"key":"ref86","first-page":"1145","article-title":"Prosody conversion from neutral speech to emotional speech","volume":"14","author":"tao","year":"0","journal-title":"IEEE Trans On Audio Speech & Language Processing"},{"key":"ref87","doi-asserted-by":"publisher","DOI":"10.1145\/1543834.1543972"},{"key":"ref88","article-title":"Emotional speech synthesis: Applications, history and possible future","author":"burkhardt","year":"0","journal-title":"Proc ESSV"}],"event":{"name":"Gesture Recognition (FG 2011)","location":"Santa Barbara, CA, USA","start":{"date-parts":[[2011,3,21]]},"end":{"date-parts":[[2011,3,25]]}},"container-title":["Face and Gesture 2011"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx5\/5765597\/5771322\/05771357.pdf?arnumber=5771357","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,4,7]],"date-time":"2024-04-07T01:41:54Z","timestamp":1712454114000},"score":1,"resource":{"primary":{"URL":"http:\/\/ieeexplore.ieee.org\/document\/5771357\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2011,3]]},"references-count":121,"URL":"https:\/\/doi.org\/10.1109\/fg.2011.5771357","relation":{},"subject":[],"published":{"date-parts":[[2011,3]]}}}