{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,15]],"date-time":"2026-01-15T01:20:52Z","timestamp":1768440052320,"version":"3.49.0"},"reference-count":54,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"3","license":[{"start":{"date-parts":[[2017,7,1]],"date-time":"2017-07-01T00:00:00Z","timestamp":1498867200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"}],"funder":[{"name":"University of Rome Tor Vergata"},{"name":"European Union\u2019s ERC","award":["338164"],"award-info":[{"award-number":["338164"]}]},{"name":"iHEARu"},{"name":"Horizon 2020 Programme"},{"name":"Innovation Action (IA)","award":["#644632"],"award-info":[{"award-number":["#644632"]}]},{"name":"MixedEmotions","award":["#645094"],"award-info":[{"award-number":["#645094"]}]},{"name":"SEWA","award":["IA #645378"],"award-info":[{"award-number":["IA #645378"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Affective Comput."],"published-print":{"date-parts":[[2017,7,1]]},"DOI":"10.1109\/taffc.2016.2531664","type":"journal-article","created":{"date-parts":[[2016,2,18]],"date-time":"2016-02-18T21:52:52Z","timestamp":1455832372000},"page":"314-327","source":"Crossref","is-referenced-by-count":38,"title":["Continuous Estimation of Emotions in Speech by Dynamic Cooperative Speaker Models"],"prefix":"10.1109","volume":"8","author":[{"given":"Arianna","family":"Mencattini","sequence":"first","affiliation":[]},{"given":"Eugenio","family":"Martinelli","sequence":"additional","affiliation":[]},{"given":"Fabien","family":"Ringeval","sequence":"additional","affiliation":[]},{"given":"Bjorn","family":"Schuller","sequence":"additional","affiliation":[]},{"given":"Corrado Di","family":"Natale","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1145\/2682899"},{"key":"ref38","first-page":"3","article-title":"AV+EC 2015 &#x2013; The first affect recognition challenge bridging across audio, video, and physiological data","author":"ringeval","year":"0","journal-title":"Proc of International Audio\/Visual Emotion Challenge and Workshop"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.3389\/fpsyg.2013.00292"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2011.12.005"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/ICME.2010.5583101"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/ACII.2013.21"},{"key":"ref37","first-page":"162","article-title":"Multimodal fusion framework: A multiresolution approach for emotion classification and recognition from physiological signals","volume":"102","author":"verma","year":"2013","journal-title":"NeuroImage Special Issue Multimodal Data Fusion"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/ICOT.2014.6956642"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1145\/2379776.2379786"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2016.7472666"},{"key":"ref28","first-page":"43","article-title":"Automatic segmentation of spontaneous data using dimensional labels from multiple coders","author":"nicolaou","year":"0","journal-title":"Proc Int Workshop Multimodal Corpora Adv Capturing Coding Anal"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1016\/j.patrec.2014.11.007"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1145\/2388676.2388783"},{"key":"ref2","doi-asserted-by":"crossref","first-page":"228","DOI":"10.1016\/j.bspc.2014.08.006","article-title":"Prediction of major depression in adolescents using an optimized multi-channel weighted speech classification system","volume":"14","author":"ooi","year":"2014","journal-title":"Biomed Signal Process Control"},{"key":"ref1","first-page":"141","article-title":"From joyous to clinically depressed: Mood detection using spontaneous speech","author":"alghowinem","year":"0","journal-title":"Proc Int Florida Artif Intell Res Soc Conf"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2007.367262"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1007\/s12193-009-0032-6"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1007\/s12559-009-9016-9"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1037\/h0026750"},{"key":"ref23","article-title":"Appraisal processes in emotion: Theory, methods, research","author":"scherer","year":"2001","journal-title":"Series in Affective Science"},{"key":"ref26","first-page":"29","article-title":"A multi-task approach to continuous five-dimensional affect sensing in natural speech","volume":"2","author":"eyben","year":"2012","journal-title":"ACM Trans Interactive Intell Syst (TiiS) - Special Issue Affective Interaction Natural Environ"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2014.2334294"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.2307\/2532051"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/TASL.2010.2052246"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1016\/j.proeng.2015.08.716"},{"key":"ref53","first-page":"1977","article-title":"Face reading from speech &#x2013; predicting facial action units from audio cues","author":"ringeval","year":"0","journal-title":"Proc INTERSPEECH 2015 16th Annu Conf Int Speech Commun Assoc"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1145\/1961189.1961199"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2012.06.016"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1016\/j.specom.2006.04.003"},{"key":"ref40","doi-asserted-by":"crossref","first-page":"83","DOI":"10.1016\/j.snb.2014.03.070","article-title":"Cooperative classifiers for reconfigurable sensor arrays","volume":"199","year":"2014","journal-title":"Sensors Actuators B Chem"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2010.09.020"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2014.03.019"},{"key":"ref14","first-page":"71","article-title":"The automatic recognition of emotions in speech","author":"batliner","year":"2010","journal-title":"Emotion-Oriented Systems The Humaine Handbook (Cognitive Technologies)"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1016\/S0167-6393(02)00071-7"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1017\/S0954579405050340"},{"key":"ref17","doi-asserted-by":"crossref","first-page":"488","DOI":"10.1007\/978-3-540-74889-2_43","article-title":"The HUMAINE database: Addressing the collection and annotation of naturalistic and induced emotional data","volume":"4738","author":"douglas-cowie","year":"0","journal-title":"Int Conf on Affective Computing and Intelligent Interaction"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/T-AFFC.2011.34"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/FG.2013.6553805"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/TASL.2010.2090147"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1515\/9781614515159.207"},{"key":"ref6","first-page":"246","article-title":"A taxonomy of applications that utilize emotional awareness","author":"batliner","year":"0","journal-title":"Proc 5th Slovenian 1st Int Language Technol Conf"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijhcs.2009.08.007"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1145\/2370216.2370270"},{"key":"ref7","doi-asserted-by":"crossref","first-page":"363","DOI":"10.1166\/asl.2012.3169","article-title":"Research on e-learning system using speech emotion recognition","volume":"5","author":"luo","year":"2012","journal-title":"Adv Sci Lett"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1016\/0169-7439(93)85002-X"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1016\/j.specom.2008.03.012"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1002\/9781118706664"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1037\/0022-3514.70.3.614"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2005.159"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btm344"},{"key":"ref42","first-page":"148","article-title":"The INTERSPEECH 2013 computational paralinguistics challenge: Social signals, conflict, emotion, autism","author":"schuller","year":"0","journal-title":"Proc Annu Conf Int Speech Commun Assoc"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1109\/ICME.2008.4607572"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-27299-3"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1145\/2502081.2502224"}],"container-title":["IEEE Transactions on Affective Computing"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/5165369\/8023989\/07412670.pdf?arnumber=7412670","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,1,12]],"date-time":"2022-01-12T16:40:54Z","timestamp":1642005654000},"score":1,"resource":{"primary":{"URL":"http:\/\/ieeexplore.ieee.org\/document\/7412670\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2017,7,1]]},"references-count":54,"journal-issue":{"issue":"3"},"URL":"https:\/\/doi.org\/10.1109\/taffc.2016.2531664","relation":{},"ISSN":["1949-3045"],"issn-type":[{"value":"1949-3045","type":"print"}],"subject":[],"published":{"date-parts":[[2017,7,1]]}}}