{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,12]],"date-time":"2025-10-12T04:57:27Z","timestamp":1760245047468,"version":"3.28.0"},"reference-count":25,"publisher":"IEEE","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2018,7]]},"DOI":"10.1109\/icme.2018.8486564","type":"proceedings-article","created":{"date-parts":[[2018,10,11]],"date-time":"2018-10-11T18:34:18Z","timestamp":1539282858000},"page":"1-6","source":"Crossref","is-referenced-by-count":11,"title":["Auditory-Inspired End-to-End Speech Emotion Recognition Using 3D Convolutional Recurrent Neural Networks Based on Spectral-Temporal Representation"],"prefix":"10.1109","author":[{"given":"Zhichao","family":"Peng","sequence":"first","affiliation":[]},{"given":"Zhi","family":"Zhu","sequence":"additional","affiliation":[]},{"given":"Masashi","family":"Unoki","sequence":"additional","affiliation":[]},{"given":"Jianwu","family":"Dang","sequence":"additional","affiliation":[]},{"given":"Masato","family":"Akagi","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref10","doi-asserted-by":"crossref","first-page":"1926","DOI":"10.1109\/TASLP.2015.2456420","article-title":"An auditory inspired amplitude modulation filter bank for robust feature extraction in automatic speech recognition","volume":"23 11","author":"moritz","year":"2015","journal-title":"IEEE\/ACM Transactions on Audio Speech and Language Processing"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1016\/j.specom.2010.08.013"},{"key":"ref12","first-page":"262","article-title":"Modulation Spectral Features for Predicting Vocal Emotion Recognition by Simulated Cochlear Implants","author":"zhu","year":"2016","journal-title":"Proceedings of Interspeech"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1016\/j.neuron.2011.06.032"},{"key":"ref14","first-page":"3361","article-title":"Convolutional networks for images, speech, and time series","volume":"10","author":"lecun","year":"1995","journal-title":"The Handbook of Brain Theory and Neural Networks"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1523\/JNEUROSCI.20-06-02315.2000"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2014.2360798"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/APSIPA.2016.7820699"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN.2016.7727636"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.3109\/00206097209072581"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1111\/j.1748-1716.1971.tb04931.x"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.1982.1171644"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1121\/1.1945807"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1121\/1.414456"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/TASL.2006.874669"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1121\/1.420344"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1250\/ast.22.426"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1007\/s10772-011-9125-1"},{"journal-title":"Attentive Convolutional Neural Network based Speech Emotion Recognition A Study on the Impact of Input Features Signal Length and Acted Speech","year":"2017","author":"neumann","key":"ref20"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1007\/s10579-008-9076-6"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2016.7472669"},{"journal-title":"Emotion Recognition From Speech With Recurrent Neural Networks","year":"2017","author":"chernykh","key":"ref24"},{"key":"ref23","first-page":"223","article-title":"Speech emotion recognition using deep neural network and extreme learning machine","author":"han","year":"2014","journal-title":"Proceedings of Interspeech"},{"journal-title":"Learning representations of affect from speech","year":"2015","author":"ghosh","key":"ref25"}],"event":{"name":"2018 IEEE International Conference on Multimedia and Expo (ICME)","start":{"date-parts":[[2018,7,23]]},"location":"San Diego, CA","end":{"date-parts":[[2018,7,27]]}},"container-title":["2018 IEEE International Conference on Multimedia and Expo (ICME)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/8472825\/8486434\/08486564.pdf?arnumber=8486564","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,1,26]],"date-time":"2022-01-26T12:13:47Z","timestamp":1643199227000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8486564\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018,7]]},"references-count":25,"URL":"https:\/\/doi.org\/10.1109\/icme.2018.8486564","relation":{},"subject":[],"published":{"date-parts":[[2018,7]]}}}