{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,3]],"date-time":"2026-03-03T16:04:19Z","timestamp":1772553859197,"version":"3.50.1"},"reference-count":66,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"1","license":[{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Affective Comput."],"published-print":{"date-parts":[[2019,1,1]]},"DOI":"10.1109\/taffc.2017.2684799","type":"journal-article","created":{"date-parts":[[2017,3,20]],"date-time":"2017-03-20T18:13:38Z","timestamp":1490033618000},"page":"85-99","source":"Crossref","is-referenced-by-count":54,"title":["Cross-Corpus Acoustic Emotion Recognition with Multi-Task Learning: Seeking Common Ground While Preserving Differences"],"prefix":"10.1109","volume":"10","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1598-2660","authenticated-orcid":false,"given":"Biqiao","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Emily Mower","family":"Provost","sequence":"additional","affiliation":[]},{"given":"Georg","family":"Essl","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","first-page":"1328","article-title":"Predicting emotion perception across domains: A study of singing and speaking","author":"zhang","year":"2015","journal-title":"Proc 29th AAAI Conf Artif Intell"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN.2014.6889814"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1037\/0033-2909.129.5.770"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1145\/2168752.2168754"},{"key":"ref31","first-page":"255","article-title":"Music emotion recognition: A state of the art review","author":"kim","year":"2010","journal-title":"Proc 11th Int Soc Music Inf Retrieval Conf"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1016\/j.specom.2011.01.011"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1121\/1.4799460"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1016\/j.csl.2013.10.002"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1525\/mp.2011.28.3.247"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1525\/mp.2006.23.4.319"},{"key":"ref60","first-page":"1871","article-title":"LIBLINEAR: A library for large linear classification","volume":"9","author":"fan","year":"2008","journal-title":"J Mach Learn Res"},{"key":"ref62","first-page":"1475","article-title":"Learning multiple tasks using shared hypotheses","author":"crammer","year":"2012","journal-title":"Proc 25th Int Conf Neural Inf Process Syst"},{"key":"ref61","first-page":"46","article-title":"Learning multiple tasks using manifold regularization","author":"agarwal","year":"2010","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref63","first-page":"148","article-title":"The Interspeech 2013 computational paralinguistics challenge: Social signals, conflict, emotion, autism","author":"schuller","year":"2013","journal-title":"Proc Annu Conf Int Speech Commun Assoc"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1016\/j.specom.2016.07.010"},{"key":"ref64","doi-asserted-by":"publisher","DOI":"10.1145\/1873951.1874246"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2016.7472149"},{"key":"ref65","doi-asserted-by":"publisher","DOI":"10.1109\/ACII.2015.7344678"},{"key":"ref66","doi-asserted-by":"publisher","DOI":"10.1109\/ASRU.2009.5372886"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/ASRU.2011.6163986"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-540-85099-1_1"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.4414\/smw.2013.13786"},{"key":"ref20","article-title":"Action-affect classification and morphing using multi-task representation learning","author":"shields","year":"2016","journal-title":"arXiv preprint arXiv 1603 02895"},{"key":"ref22","first-page":"3","article-title":"Automatic classification of emotions in speech using multi-corpora approaches","author":"shami","year":"2006","journal-title":"Proc 2nd Annu IEEE Benelux\/DSP Valley Signal Process Symp"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2016.7472790"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1016\/j.csl.2012.11.003"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/ACII.2015.7344562"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2015.7178934"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1186\/s13636-014-0049-y"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1109\/T-AFFC.2011.20"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/JPROC.2012.2236291"},{"key":"ref59","first-page":"521","article-title":"Learning with whom to share in multi-task feature learning","author":"kang","year":"2011","journal-title":"Proc 28th Int Conf Mach Learn"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.1007\/s10994-007-5040-8"},{"key":"ref57","first-page":"41","article-title":"Multi-task feature learning","author":"argyriou","year":"2007","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1080\/17470218.2014.971034"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1145\/2388676.2388776"},{"key":"ref54","doi-asserted-by":"crossref","first-page":"415","DOI":"10.1007\/978-3-642-24571-8_53","article-title":"AVEC 2011&#x2013;the first international audio\/visual emotion challenge","author":"schuller","year":"2011","journal-title":"Proc 4th Int Conf Affective Comput Intell Interaction"},{"key":"ref53","article-title":"Ravdess: The Ryerson audio-visual database of emotional speech and song","author":"livingstone","year":"2012","journal-title":"Annual Meeting of the Canadian Society for Brain Behaviour and Cognitive Science"},{"key":"ref52","article-title":"The eNTERFACE&#x2019;05 audio-visual emotion database","author":"martin","year":"2006","journal-title":"Proceedings of the 22nd International Conference on Data Engineering Workshops"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1587\/transinf.2014EDL8038"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/ACII.2015.7344563"},{"key":"ref40","first-page":"58","article-title":"Towards measuring similarity between emotional corpora","author":"brendel","year":"2010","journal-title":"Proc of a Satellite Workshop of LREC"},{"key":"ref12","doi-asserted-by":"crossref","first-page":"27","DOI":"10.1145\/2647868.2654934","article-title":"Say cheese versus smile: Reducing speech-related variability for facial emotion recognition","author":"kim","year":"2014","journal-title":"Proc 22nd ACM Int Conf Multimedia"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR.2006.489"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/TASL.2010.2076804"},{"key":"ref15","first-page":"341","article-title":"Automatic speech classification to five emotional states based on gender information","author":"ververidis","year":"2004","journal-title":"Proc 12th Eur Signal Process Conf"},{"key":"ref16","doi-asserted-by":"crossref","first-page":"293","DOI":"10.1109\/TSA.2004.838534","article-title":"Toward detecting emotions in spoken dialogs","volume":"13","author":"lee","year":"2005","journal-title":"IEEE Trans Speech Audio Process"},{"key":"ref17","first-page":"1123","article-title":"Improving automatic emotion recognition from speech via gender differentiation","author":"vogt","year":"2006","journal-title":"Proc Lang Resources Eval Conf"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/FG.2013.6553779"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.430"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1007\/s10772-011-9125-1"},{"key":"ref3","author":"reeves","year":"1996","journal-title":"How people treat computers television and new media like real people and places"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/T-AFFC.2010.8"},{"key":"ref5","doi-asserted-by":"crossref","first-page":"75","DOI":"10.1007\/978-3-540-85099-1_7","article-title":"Automatic recognition of emotions from speech: A review of the literature and recommendations for practical realisation","author":"vogt","year":"2008","journal-title":"Affect and Emotion in Human-Computer Interaction"},{"key":"ref8","doi-asserted-by":"crossref","first-page":"1553","DOI":"10.21437\/Interspeech.2011-468","article-title":"Using multiple databases for training in emotion recognition: To unite or to vote?","author":"schuller","year":"2011","journal-title":"Proc INTERSPEECH"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-15760-8_45"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1109\/ICME.2010.5583006"},{"key":"ref9","article-title":"Selecting training data for cross-corpus speech emotion recognition: Prototypicality versus generalization","author":"schuller","year":"2011","journal-title":"Proc Afeka-AVIOS Speech Process Conf"},{"key":"ref46","article-title":"Documentation of the Danish emotional speech database des","author":"engberg","year":"1996"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2005-446"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1109\/ICME.2008.4607572"},{"key":"ref47","first-page":"28","article-title":"Releasing a thoroughly annotated and processed spontaneous emotional database: the Fau Aibo emotion corpus","author":"batliner","year":"2008","journal-title":"Proc of a Satellite Workshop of LREC"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/ICME.2011.6012003"},{"key":"ref41","first-page":"141","article-title":"From joyous to clinically depressed: Mood detection using spontaneous speech","author":"alghowinem","year":"2012","journal-title":"Proc Int Florida Artif Intell Res Soc Conf"},{"key":"ref44","first-page":"1444","article-title":"Multilinear multitask learning","author":"romera-paredes","year":"2013","journal-title":"Proc 30th Int Conf Int Conf Mach Learn"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1007\/s10772-012-9170-4"}],"container-title":["IEEE Transactions on Affective Computing"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/5165369\/8662807\/07882650.pdf?arnumber=7882650","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,7,13]],"date-time":"2022-07-13T20:57:11Z","timestamp":1657745831000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/7882650\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,1,1]]},"references-count":66,"journal-issue":{"issue":"1"},"URL":"https:\/\/doi.org\/10.1109\/taffc.2017.2684799","relation":{},"ISSN":["1949-3045","2371-9850"],"issn-type":[{"value":"1949-3045","type":"electronic"},{"value":"2371-9850","type":"electronic"}],"subject":[],"published":{"date-parts":[[2019,1,1]]}}}