{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,6]],"date-time":"2024-09-06T22:04:27Z","timestamp":1725660267596},"reference-count":9,"publisher":"IEEE","license":[{"start":{"date-parts":[[2019,10,1]],"date-time":"2019-10-01T00:00:00Z","timestamp":1569888000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2019,10,1]],"date-time":"2019-10-01T00:00:00Z","timestamp":1569888000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2019,10,1]],"date-time":"2019-10-01T00:00:00Z","timestamp":1569888000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019,10]]},"DOI":"10.1109\/gcce46687.2019.9015558","type":"proceedings-article","created":{"date-parts":[[2020,2,28]],"date-time":"2020-02-28T05:21:46Z","timestamp":1582867306000},"page":"221-224","source":"Crossref","is-referenced-by-count":5,"title":["Emotion Estimation by Joint Facial Expression and Speech Tonality Using Evolutionary Deep Learning Structures"],"prefix":"10.1109","author":[{"given":"Chih-Che","family":"Chung","sequence":"first","affiliation":[]},{"given":"Wei-Ting","family":"Lin","sequence":"additional","affiliation":[]},{"given":"Rong","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Kai-Wen","family":"Liang","sequence":"additional","affiliation":[]},{"given":"Pao-Chi","family":"Chang","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"doi-asserted-by":"publisher","key":"ref4","DOI":"10.1007\/3-540-58484-6_252"},{"doi-asserted-by":"publisher","key":"ref3","DOI":"10.1109\/41.538609"},{"key":"ref6","article-title":"Feature Fusion Algorithm for Multimodal Emotion Recognition from Speech and Facial Expression Signal","author":"zhiyan","year":"2016","journal-title":"MATEC Web of Conferences"},{"year":"2017","author":"arriaga","journal-title":"Real-time Convolutional Neural Networks for Emotion and Gender Classification","key":"ref5"},{"key":"ref8","article-title":"DialogueRNN: An Attentive RNN for Emotion Estimation in Conversations","author":"majumder","year":"2018","journal-title":"CoRR(abs\/1811 00405)"},{"key":"ref7","article-title":"ICON: Interactive Conversational Memory Network for Multimodal Emotion Estimation","author":"hazarika","year":"2018","journal-title":"Conference on Empirical Methods in Natural Language Processing"},{"doi-asserted-by":"publisher","key":"ref2","DOI":"10.1007\/s10579-008-9076-6"},{"key":"ref9","article-title":"Multi-Modal Emotion recognition on IEMOCAP Dataset using Deep Learning","author":"tripathi","year":"2018","journal-title":"CoRR(abs\/1804 05788)"},{"year":"0","author":"goodfellow","journal-title":"Challenges in representation learning A report on three machine learning contests","key":"ref1"}],"event":{"name":"2019 IEEE 8th Global Conference on Consumer Electronics (GCCE)","start":{"date-parts":[[2019,10,15]]},"location":"Osaka, Japan","end":{"date-parts":[[2019,10,18]]}},"container-title":["2019 IEEE 8th Global Conference on Consumer Electronics (GCCE)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/8977397\/9014630\/09015558.pdf?arnumber=9015558","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,7,17]],"date-time":"2022-07-17T17:51:13Z","timestamp":1658080273000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9015558\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,10]]},"references-count":9,"URL":"https:\/\/doi.org\/10.1109\/gcce46687.2019.9015558","relation":{},"subject":[],"published":{"date-parts":[[2019,10]]}}}