{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,15]],"date-time":"2026-04-15T17:57:10Z","timestamp":1776275830410,"version":"3.50.1"},"reference-count":20,"publisher":"IEEE","license":[{"start":{"date-parts":[[2022,5,23]],"date-time":"2022-05-23T00:00:00Z","timestamp":1653264000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,5,23]],"date-time":"2022-05-23T00:00:00Z","timestamp":1653264000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001691","name":"Japan Society for the Promotion of Science","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001691","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022,5,23]]},"DOI":"10.1109\/icassp43922.2022.9747129","type":"proceedings-article","created":{"date-parts":[[2022,4,27]],"date-time":"2022-04-27T19:50:34Z","timestamp":1651089034000},"page":"6427-6431","source":"Crossref","is-referenced-by-count":10,"title":["Domain-Invariant Feature Learning for Cross Corpus Speech Emotion Recognition"],"prefix":"10.1109","author":[{"given":"Yuan","family":"Gao","sequence":"first","affiliation":[{"name":"Tianjin University,Tianjin Key Laboratory of Cognitive Computing and Application, College of Intelligence and Computing,Tianjin,China"}]},{"given":"Shogo","family":"Okada","sequence":"additional","affiliation":[{"name":"Japan Advanced Institute of Science and Technology,Ishikawa,Japan"}]},{"given":"Longbiao","family":"Wang","sequence":"additional","affiliation":[{"name":"Tianjin University,Tianjin Key Laboratory of Cognitive Computing and Application, College of Intelligence and Computing,Tianjin,China"}]},{"given":"Jiaxing","family":"Liu","sequence":"additional","affiliation":[{"name":"Tianjin University,Tianjin Key Laboratory of Cognitive Computing and Application, College of Intelligence and Computing,Tianjin,China"}]},{"given":"Jianwu","family":"Dang","sequence":"additional","affiliation":[{"name":"Tianjin University,Tianjin Key Laboratory of Cognitive Computing and Application, College of Intelligence and Computing,Tianjin,China"}]}],"member":"263","reference":[{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2015.2503757"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2019-2753"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ASRU46091.2019.9003838"},{"key":"ref13","doi-asserted-by":"crossref","first-page":"1089","DOI":"10.21437\/Interspeech.2017-200","article-title":"Efficient emotion recognition from speech using deep learning on spectrograms","author":"satt","year":"2017","journal-title":"InterSpeech"},{"key":"ref14","first-page":"2096","article-title":"Domain-adversarial training of neural networks","volume":"17","author":"ganin","year":"2016","journal-title":"The Journal of Machine Learning Research"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46478-7_31"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1007\/s10579-008-9076-6"},{"key":"ref17","doi-asserted-by":"crossref","first-page":"67","DOI":"10.1109\/TAFFC.2016.2515617","article-title":"Msp-improv: An acted corpus of dyadic interactions to study emotion perception","volume":"8","author":"busso","year":"2016","journal-title":"IEEE Transactions on Affective Computing"},{"key":"ref18","author":"jackson","year":"2014","journal-title":"Surrey audio-visual expressed emotion (savee) database"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2005-446"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2018.08.035"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2018.8462677"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/TASLP.2018.2867099"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP40776.2020.9053192"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/T-AFFC.2010.8"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2018.8462162"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2017.02.013"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2010.09.020"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/TASL.2013.2255278"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2017.2684799"}],"event":{"name":"ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","location":"Singapore, Singapore","start":{"date-parts":[[2022,5,23]]},"end":{"date-parts":[[2022,5,27]]}},"container-title":["ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9745891\/9746004\/09747129.pdf?arnumber=9747129","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,8,22]],"date-time":"2022-08-22T20:10:30Z","timestamp":1661199030000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9747129\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,5,23]]},"references-count":20,"URL":"https:\/\/doi.org\/10.1109\/icassp43922.2022.9747129","relation":{},"subject":[],"published":{"date-parts":[[2022,5,23]]}}}