{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,27]],"date-time":"2026-02-27T16:00:28Z","timestamp":1772208028502,"version":"3.50.1"},"reference-count":20,"publisher":"Institute of Electronics, Information and Communications Engineers (IEICE)","issue":"2","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEICE Trans. Inf. &amp; Syst."],"published-print":{"date-parts":[[2020,2,1]]},"DOI":"10.1587\/transinf.2019edl8136","type":"journal-article","created":{"date-parts":[[2020,1,31]],"date-time":"2020-01-31T22:09:45Z","timestamp":1580508585000},"page":"459-463","source":"Crossref","is-referenced-by-count":23,"title":["Cross-Corpus Speech Emotion Recognition Based on Deep Domain-Adaptive Convolutional Neural Network"],"prefix":"10.1587","volume":"E103.D","author":[{"given":"Jiateng","family":"LIU","sequence":"first","affiliation":[{"name":"Key Laboratory of Child Development and Learning Science of Ministry of Education, Southeast University"}]},{"given":"Wenming","family":"ZHENG","sequence":"additional","affiliation":[{"name":"Key Laboratory of Child Development and Learning Science of Ministry of Education, Southeast University"}]},{"given":"Yuan","family":"ZONG","sequence":"additional","affiliation":[{"name":"Key Laboratory of Child Development and Learning Science of Ministry of Education, Southeast University"}]},{"given":"Cheng","family":"LU","sequence":"additional","affiliation":[{"name":"School of Information Science and Engineering, Southeast University"}]},{"given":"Chuangao","family":"TANG","sequence":"additional","affiliation":[{"name":"Key Laboratory of Child Development and Learning Science of Ministry of Education, Southeast University"}]}],"member":"532","reference":[{"key":"1","doi-asserted-by":"publisher","unstructured":"[1] M.E. Ayadi, M.S. Kamel, and F. Karray, \u201cSurvey on speech emotion recognition: Features, classification schemes, and databases,\u201d Pattern Recognition, vol.44, no.3, pp.572-587, 2011. 10.1016\/j.patcog.2010.09.020","DOI":"10.1016\/j.patcog.2010.09.020"},{"key":"2","doi-asserted-by":"publisher","unstructured":"[2] L. Sun, J. Chen, K. Xie, and T. Gu, \u201cDeep and shallow features fusion based on deep convolutional neural network for speech emotion recognition,\u201d International Journal of Speech Technology, vol.21, no.4, pp.931-940, 2018. 10.1007\/s10772-018-9551-4","DOI":"10.1007\/s10772-018-9551-4"},{"key":"3","doi-asserted-by":"publisher","unstructured":"[3] P. Song, W. Zheng, S. Ou, X. Zhang, Y. Jin, J. Liu, and Y. Yu, \u201cCross-corpus speech emotion recognition based on transfer non-negative matrix factorization,\u201d Speech Communication, vol.83, pp.34-41, 2016. 10.1016\/j.specom.2016.07.010","DOI":"10.1016\/j.specom.2016.07.010"},{"key":"4","doi-asserted-by":"publisher","unstructured":"[4] Y. Zong, W. Zheng, T. Zhang, and X. Huang, \u201cCross-corpus speech emotion recognition based on domain-adaptive least-squares regression,\u201d IEEE Signal Process. Lett., vol.23, no.5, pp.585-589, 2016. 10.1109\/lsp.2016.2537926","DOI":"10.1109\/LSP.2016.2537926"},{"key":"5","doi-asserted-by":"crossref","unstructured":"[5] A.M. Badshah, J. Ahmad, N. Rahim, and S.W. Baik, \u201cSpeech emotion recognition from spectrograms with deep convolutional neural network,\u201d 2017 international conference on platform technology and service (PlatCon), pp.1-5, IEEE, 2017. 10.1109\/platcon.2017.7883728","DOI":"10.1109\/PlatCon.2017.7883728"},{"key":"6","doi-asserted-by":"publisher","unstructured":"[6] T. Song, W. Zheng, C. Lu, Y. Zong, X. Zhang, and Z. Cui, \u201cMped: A multi-modal physiological emotion database for discrete emotion recognition,\u201d IEEE Access, vol.7, pp.12177-12191, 2019. 10.1109\/access.2019.2891579","DOI":"10.1109\/ACCESS.2019.2891579"},{"key":"7","doi-asserted-by":"publisher","unstructured":"[7] R.V. Shannon, F.-G. Zeng, V. Kamath, J. Wygonski, and M. Ekelid, \u201cSpeech recognition with primarily temporal cues,\u201d Science, vol.270, no.5234, pp.303-304, 1995. 10.1126\/science.270.5234.303","DOI":"10.1126\/science.270.5234.303"},{"key":"8","doi-asserted-by":"publisher","unstructured":"[8] Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner, \u201cGradient-based learning applied to document recognition,\u201d Proceedings of the IEEE, vol.86, no.11, pp.2278-2324, 1998. 10.1109\/5.726791","DOI":"10.1109\/5.726791"},{"key":"9","doi-asserted-by":"publisher","unstructured":"[9] A. Krizhevsky, I. Sutskever, and G.E. Hinton, \u201cImagenet classification with deep convolutional neural networks,\u201d Commun. ACM, vol.60, no.6, pp.84-90, 2017. 10.1145\/3065386","DOI":"10.1145\/3065386"},{"key":"10","doi-asserted-by":"crossref","unstructured":"[10] A. Gretton, K. Borgwardt, M. Rasch, B. Sch\u00f6lkopf, and A.J. Smola, \u201cA kernel method for the two-sample-problem,\u201d Advances in Neural Information Processing Systems, pp.513-520, 2007.","DOI":"10.7551\/mitpress\/7503.003.0069"},{"key":"11","unstructured":"[11] M. Long, H. Zhu, J. Wang, and M.I. Jordan, \u201cDeep transfer learning with joint adaptation networks,\u201d Proceedings of the 34th International Conference on Machine Learning, pp.2208-2217, 2017."},{"key":"12","doi-asserted-by":"publisher","unstructured":"[12] K. Yan, W. Zheng, T. Zhang, Y. Zong, C. Tang, C. Lu, and Z. Cui, \u201cCross-domain facial expression recognition based on transductive deep transfer learning,\u201d IEEE Access, vol.7, pp.108906-108915, 2019. 10.1109\/access.2019.2930359","DOI":"10.1109\/ACCESS.2019.2930359"},{"key":"13","doi-asserted-by":"crossref","unstructured":"[13] A. Smola, A. Gretton, L. Song, and B. Sch\u00f6lkopf, \u201cA hilbert space embedding for distributions,\u201d International Conference on Algorithmic Learning Theory, vol.4754, pp.13-31, Springer, 2007. 10.1007\/978-3-540-75225-7_5","DOI":"10.1007\/978-3-540-75225-7_5"},{"key":"14","doi-asserted-by":"crossref","unstructured":"[14] F. Burkhardt, A. Paeschke, M. Rolfes, W.F. Sendlmeier, and B. Weiss, \u201cA database of german emotional speech,\u201d Ninth European Conference on Speech Communication and Technology, 2005.","DOI":"10.21437\/Interspeech.2005-446"},{"key":"15","doi-asserted-by":"crossref","unstructured":"[15] O. Martin, I. Kotsia, B. Macq, and I. Pitas, \u201cThe enterface&apos;05 audio-visual emotion database,\u201d 22nd International Conference on Data Engineering Workshops (ICDEW&apos;06), p.8, IEEE, 2006. 10.1109\/icdew.2006.145","DOI":"10.1109\/ICDEW.2006.145"},{"key":"16","unstructured":"[16] J. Tao, F. Liu, M. Zhang, and H. Jia, \u201cDesign of speech corpus for mandarin text to speech,\u201d The Blizzard Challenge 2008 Workshop, 2008."},{"key":"17","doi-asserted-by":"publisher","unstructured":"[17] S.J. Pan, I.W. Tsang, J.T. Kwok, and Q. Yang, \u201cDomain adaptation via transfer component analysis,\u201d IEEE Trans. Neural Netw., vol.22, no.2, pp.199-210, 2010. 10.1109\/tnn.2010.2091281","DOI":"10.1109\/TNN.2010.2091281"},{"key":"18","doi-asserted-by":"publisher","unstructured":"[18] M. Long, J. Wang, J. Sun, and P.S. Yu, \u201cDomain invariant transfer kernel learning,\u201d IEEE Trans. Knowl. Data Eng., vol.27, no.6, pp.1519-1532, 2015. 10.1109\/tkde.2014.2373376","DOI":"10.1109\/TKDE.2014.2373376"},{"key":"19","doi-asserted-by":"crossref","unstructured":"[19] B. Schuller, S. Steidl, and A. Batliner, \u201cThe interspeech 2009 emotion challenge,\u201d Tenth Annual Conference of the International Speech Communication Association, 2009.","DOI":"10.21437\/Interspeech.2009-103"},{"key":"20","doi-asserted-by":"crossref","unstructured":"[20] F. Eyben, M. W\u00f6llmer, and B. Schuller, \u201cOpensmile: the munich versatile and fast open-source audio feature extractor,\u201d Proceedings of the 18th ACM international conference on Multimedia, pp.1459-1462, ACM, 2010. 10.1145\/1873951.1874246","DOI":"10.1145\/1873951.1874246"}],"container-title":["IEICE Transactions on Information and Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.jstage.jst.go.jp\/article\/transinf\/E103.D\/2\/E103.D_2019EDL8136\/_pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,9,26]],"date-time":"2023-09-26T00:10:35Z","timestamp":1695687035000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.jstage.jst.go.jp\/article\/transinf\/E103.D\/2\/E103.D_2019EDL8136\/_article"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,2,1]]},"references-count":20,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2020]]}},"URL":"https:\/\/doi.org\/10.1587\/transinf.2019edl8136","relation":{},"ISSN":["0916-8532","1745-1361"],"issn-type":[{"value":"0916-8532","type":"print"},{"value":"1745-1361","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,2,1]]}}}