{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,10,29]],"date-time":"2024-10-29T09:25:12Z","timestamp":1730193912642,"version":"3.28.0"},"reference-count":20,"publisher":"IEEE","license":[{"start":{"date-parts":[[2019,9,1]],"date-time":"2019-09-01T00:00:00Z","timestamp":1567296000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2019,9,1]],"date-time":"2019-09-01T00:00:00Z","timestamp":1567296000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2019,9,1]],"date-time":"2019-09-01T00:00:00Z","timestamp":1567296000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019,9]]},"DOI":"10.1109\/acii.2019.8925496","type":"proceedings-article","created":{"date-parts":[[2019,12,27]],"date-time":"2019-12-27T13:44:34Z","timestamp":1577454274000},"page":"83-88","source":"Crossref","is-referenced-by-count":2,"title":["A Dual-Complementary Acoustic Embedding Network Learned from Raw Waveform for Speech Emotion Recognition"],"prefix":"10.1109","author":[{"given":"Tzu-Yun","family":"Huang","sequence":"first","affiliation":[]},{"given":"Jeng-Lin","family":"Li","sequence":"additional","affiliation":[]},{"given":"Chun-Min","family":"Chang","sequence":"additional","affiliation":[]},{"given":"Chi-Chun","family":"Lee","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2017.7952552"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/ACII.2015.7344669"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2018-1353"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2018-2397"},{"key":"ref14","article-title":"Reusing neural speech representations for auditory emotion recognition","volume":"1","author":"lakomkin","year":"2017","journal-title":"Proceedings of the Eighth International Joint Conference on Natural Language Processing"},{"key":"ref15","doi-asserted-by":"crossref","first-page":"190","DOI":"10.1109\/TAFFC.2015.2457417","article-title":"The geneva minimalistic acoustic parameter set (gemaps) for voice research and affective computing","volume":"7","author":"florian","year":"2016","journal-title":"IEEE Transactions on Affective Computing"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1007\/s10579-008-9076-6"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2017.02.013"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1145\/2502081.2502224"},{"key":"ref19","doi-asserted-by":"crossref","first-page":"1301","DOI":"10.1109\/JSTSP.2017.2764438","article-title":"End-to-end multimodal emotion recognition using deep neural networks","volume":"11","author":"panagiotis","year":"2017","journal-title":"IEEE Journal of Selected Topics in Signal Processing"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1023\/A:1013215010749"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/TCE.2009.5278031"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/MC.2004.1297238"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1111\/j.1559-1816.2007.00272.x"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.3813\/AAA.919214"},{"journal-title":"Attentive Convolutional Neural Network based Speech Emotion Recognition A Study on the Impact of Input Features Signal Length and Acted Speech","year":"2017","author":"neumann","key":"ref7"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1016\/S0167-6393(03)00099-2"},{"key":"ref1","first-page":"i","article-title":"Hidden markov model-based speech emotion recognition","volume":"1","author":"schuller","year":"2003","journal-title":"Multimedia and Expo"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2017.7952655"},{"key":"ref20","article-title":"The opensmile book - opensmile: The munich versatile and fast open-source audio feature extractor","author":"eyben","year":"2010","journal-title":"ACM Multimedia"}],"event":{"name":"2019 8th International Conference on Affective Computing and Intelligent Interaction (ACII)","start":{"date-parts":[[2019,9,3]]},"location":"Cambridge, United Kingdom","end":{"date-parts":[[2019,9,6]]}},"container-title":["2019 8th International Conference on Affective Computing and Intelligent Interaction (ACII)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/8911251\/8925431\/08925496.pdf?arnumber=8925496","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,7,19]],"date-time":"2022-07-19T20:26:36Z","timestamp":1658262396000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8925496\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,9]]},"references-count":20,"URL":"https:\/\/doi.org\/10.1109\/acii.2019.8925496","relation":{},"subject":[],"published":{"date-parts":[[2019,9]]}}}