{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,29]],"date-time":"2025-11-29T07:58:55Z","timestamp":1764403135335},"reference-count":26,"publisher":"IEEE","license":[{"start":{"date-parts":[[2021,9,28]],"date-time":"2021-09-28T00:00:00Z","timestamp":1632787200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2021,9,28]],"date-time":"2021-09-28T00:00:00Z","timestamp":1632787200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2021,9,28]],"date-time":"2021-09-28T00:00:00Z","timestamp":1632787200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021,9,28]]},"DOI":"10.1109\/aciiw52867.2021.9666308","type":"proceedings-article","created":{"date-parts":[[2022,1,10]],"date-time":"2022-01-10T16:14:00Z","timestamp":1641831240000},"page":"1-5","source":"Crossref","is-referenced-by-count":3,"title":["Emotion Recognition In Emergency Call Centers: The challenge of real-life emotions"],"prefix":"10.1109","author":[{"given":"Theo","family":"Deschamps-Berger","sequence":"first","affiliation":[]}],"member":"263","reference":[{"key":"ref10","first-page":"21","author":"etienne","year":"2018","journal-title":"CNN+LSTM Architecture for Speech Emotion Recognition with Data Augmentation"},{"journal-title":"A structured self-attentive sentence embedding","year":"2017","author":"lin","key":"ref11"},{"journal-title":"Attention is all you need","year":"2017","author":"vaswani","key":"ref12"},{"journal-title":"Deep Architecture Enhancing Robustness to Noise Adversarial At-tacks and Cross-corpus Setting for Speech Emotion Recognition","year":"2020","author":"latif","key":"ref13"},{"journal-title":"Zero-shot learning - a comprehensive evaluation of the good the bad and the ugly","year":"2020","author":"xian","key":"ref14"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2019-2406"},{"key":"ref16","article-title":"Modality-Transferable Emotion Embeddings for Low-Resource Multimodal Emotion Recognition","author":"dai","year":"2020","journal-title":"ACL\/IJCNLP"},{"journal-title":"Bert Pretraining of deep bidirectional transformers for language understanding","year":"2019","author":"devlin","key":"ref17"},{"key":"ref18","article-title":"Improving Language Understanding by Generative Pre-Training","author":"radford","year":"2018","journal-title":"Undefined"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2020-3156"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2005.03.001"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.3389\/fpsyg.2019.01606"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1007\/s12369-015-0297-8"},{"key":"ref5","first-page":"iii","author":"devillers","year":"2003","journal-title":"Emotion detection in task-oriented spoken dialogues"},{"key":"ref8","first-page":"1459","article-title":"openSMILE - The Munich Versatile and Fast Open-Source Audio Feature Extractor","author":"eyben","year":"2010","journal-title":"MM'10 - Proceedings of the ACM Multimedia 2010 International Conference"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-540-74889-2_39"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2005.03.007"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2015.2457417"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1511\/2001.4.344"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2020-1653"},{"key":"ref22","article-title":"Speech Recognition Technology: Improving Speed and Accuracy of Emergency Medical Services Documentation to Protect Patients","author":"tran","year":"2018","journal-title":"Auctus The Journal of Undergraduate Research and Creative Scholarship"},{"journal-title":"Learning alignment for multimodal emotion recognition from speech","year":"2020","author":"xu","key":"ref21"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1186\/s13049-021-00862-w"},{"journal-title":"AI Singapore develops speech recognition technology for emergency calls","year":"0","key":"ref23"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1037\/0022-3514.79.4.644"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1007\/s10579-008-9076-6"}],"event":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","start":{"date-parts":[[2021,9,28]]},"location":"Nara, Japan","end":{"date-parts":[[2021,10,1]]}},"container-title":["2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9666213\/9666196\/09666308.pdf?arnumber=9666308","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,10]],"date-time":"2022-05-10T12:56:52Z","timestamp":1652187412000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9666308\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,9,28]]},"references-count":26,"URL":"https:\/\/doi.org\/10.1109\/aciiw52867.2021.9666308","relation":{},"subject":[],"published":{"date-parts":[[2021,9,28]]}}}