{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T14:29:56Z","timestamp":1742912996632,"version":"3.40.3"},"publisher-location":"Cham","reference-count":20,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030664701"},{"type":"electronic","value":"9783030664718"}],"license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-3-030-66471-8_28","type":"book-chapter","created":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T22:02:50Z","timestamp":1609538570000},"page":"365-379","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Deep Neural Networks for Emotion Recognition"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3651-7629","authenticated-orcid":false,"given":"Eugene Yu.","family":"Shchetinin","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1856-4643","authenticated-orcid":false,"given":"Leonid A.","family":"Sevastianov","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0877-7063","authenticated-orcid":false,"given":"Dmitry S.","family":"Kulyabov","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6478-4041","authenticated-orcid":false,"given":"Edik A.","family":"Ayrjan","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1000-9650","authenticated-orcid":false,"given":"Anastasia V.","family":"Demidova","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,1,2]]},"reference":[{"key":"28_CR1","unstructured":"Deep learning library (2020). https:\/\/pytorch.org\/. Accessed 4 Oct 2020"},{"key":"28_CR2","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. CoRR abs\/1512.03385 (2015). http:\/\/arxiv.org\/abs\/1512.03385"},{"key":"28_CR3","unstructured":"Hochreiter, S., Bengio, Y., Frasconi, P., Schmidhuber, J.: Gradient flow in recurrent nets: the difficulty of learning long-term dependencies. In: Kremer, S.C., Kolen, J.F. (eds.) A Field Guide to Dynamical Recurrent Neural Networks. IEEE Press (2001)"},{"key":"28_CR4","doi-asserted-by":"crossref","unstructured":"Ishi, C., Ishiguro, H., Hagita, N.: Using prosodic and voice quality features for paralinguistic information extraction. In: Proceedings of the Speech Prosody 2006, pp. 883\u2013886, Dresden (2006)","DOI":"10.21437\/SpeechProsody.2006-189"},{"issue":"4","key":"28_CR5","first-page":"581","volume":"16","author":"AA Karpov","year":"2016","unstructured":"Karpov, A.A., Kaya, H., Salakh, A.A.: Actual problems and achievements of paralinguistic speech analysis. Nauchno-tekhnicheskiy vestnik informatsionnykh tekhnologiy, mekhaniki i optiki 16(4), 581\u2013592 (2016). (in Russian)","journal-title":"Nauchno-tekhnicheskiy vestnik informatsionnykh tekhnologiy, mekhaniki i optiki"},{"key":"28_CR6","unstructured":"Kennedy, L., Ellis, D.: Pitch-based emphasis detection for characterization of meeting recordings. In: Proceedings of the ASRU, pp. 243\u2013248, Virgin Islands (2003)"},{"key":"28_CR7","doi-asserted-by":"crossref","unstructured":"Kockmann, M., Burget, L., Cernock, J.: Brno university of technology system for interspeech 2010 paralinguistic challenge, pp. 2822\u20132825 (2010)","DOI":"10.21437\/Interspeech.2010-746"},{"key":"28_CR8","unstructured":"Kurkov, N.A., Shchetinin, E.Y.: Emotion classification by voice using the blstm neural network. In: Information and Telecommunication Technologies and Mathematical Modeling of High-Tech Systems, pp. 461\u2013464 (2019)"},{"key":"28_CR9","unstructured":"Lee, C., Narayanan, S., Pieraccini, R.: Recognition of negative emotions from the speech signal, pp. 240\u2013243 (2001)"},{"key":"28_CR10","doi-asserted-by":"publisher","first-page":"999","DOI":"10.1109\/ICME.2007.4284821","volume-title":"Speech Emotion Recognition Using an Enhanced Co-training Algorithm","author":"J Liu","year":"2007","unstructured":"Liu, J., Chen, C., Bu, J., You, M., Tao, J.: Speech Emotion Recognition Using an Enhanced Co-training Algorithm, pp. 999\u20131002. Springer, Heidelberg (2007). https:\/\/doi.org\/10.1109\/ICME.2007.4284821"},{"issue":"5","key":"28_CR11","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1371\/journal.pone.0196391","volume":"13","author":"SR Livingstone","year":"2018","unstructured":"Livingstone, S.R., Russo, F.A.: The Ryerson audio-visual database of emotional speech and song (Ravdess): a dynamic, multimodal set of facial and vocal expressions in North American English. PLoS ONE 13(5), 1\u201335 (2018). https:\/\/doi.org\/10.1371\/journal.pone.0196391","journal-title":"PLoS ONE"},{"key":"28_CR12","series-title":"Studies in Computational Intelligence","doi-asserted-by":"publisher","first-page":"117","DOI":"10.1007\/978-3-319-66604-4_18","volume-title":"Advances in Neural Computation, Machine Learning, and Cognitive Research","author":"AS Popova","year":"2018","unstructured":"Popova, A.S., Rassadin, A.G., Ponomarenko, A.A.: Emotion recognition in sound. In: Kryzhanovsky, B., Dunin-Barkowski, W., Redko, V. (eds.) NEUROINFORMATICS 2017. SCI, vol. 736, pp. 117\u2013124. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-319-66604-4_18"},{"key":"28_CR13","volume-title":"Fundamental of Speech Recognition","author":"L Rabiner","year":"1993","unstructured":"Rabiner, L., Juang, B.: Fundamental of Speech Recognition. Prentice-Hall, Englewood Cliffs (1993)"},{"issue":"4","key":"28_CR14","doi-asserted-by":"publisher","first-page":"1264","DOI":"10.1109\/MSP.2012.2192211","volume":"29","author":"B Schuller","year":"2012","unstructured":"Schuller, B.: The computational paralinguistics challenge. IEEE Signal Process. Mag. 29(4), 1264\u20131281 (2012)","journal-title":"IEEE Signal Process. Mag."},{"key":"28_CR15","doi-asserted-by":"publisher","DOI":"10.1002\/9781118706664","volume-title":"Computational Paralinguistics: Emotion Affect and Personality in Speech and Language Processing","author":"B Schuller","year":"2013","unstructured":"Schuller, B., Batliner, A.: Computational Paralinguistics: Emotion Affect and Personality in Speech and Language Processing. Wiley, New York (2013)"},{"key":"28_CR16","doi-asserted-by":"publisher","first-page":"2673","DOI":"10.1109\/78.650093","volume":"45","author":"M Schuster","year":"1997","unstructured":"Schuster, M., Paliwal, K.: Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 45, 2673\u20132681 (1997). https:\/\/doi.org\/10.1109\/78.650093","journal-title":"IEEE Trans. Signal Process."},{"issue":"1","key":"28_CR17","doi-asserted-by":"publisher","first-page":"67","DOI":"10.14357\/19922264200109","volume":"14","author":"LA Sevastyanov","year":"2020","unstructured":"Sevastyanov, L.A., Shchetinin, E.Y.: On methods of increasing the accuracy of multiclass classification based on unbalanced data. Inf. Appl. 14(1), 67\u201374 (2020). https:\/\/doi.org\/10.14357\/19922264200109","journal-title":"Inf. Appl."},{"issue":"3","key":"28_CR18","doi-asserted-by":"publisher","first-page":"45","DOI":"10.18311\/gjeis\/2017\/15973","volume":"9","author":"N Singh","year":"2017","unstructured":"Singh, N., Agrawal, A., Khan, R.A.: Automatic speaker recognition: current approaches and progress in last six decades. Global J. Enterp. Inf. Syst. 9(3), 45\u201352 (2017). https:\/\/doi.org\/10.18311\/gjeis\/2017\/15973","journal-title":"Global J. Enterp. Inf. Syst."},{"key":"28_CR19","volume-title":"Automatic Classification of Emotion-Related User States in Spontaneous Children\u2019s Speech","author":"S Steidl","year":"2009","unstructured":"Steidl, S.: Automatic Classification of Emotion-Related User States in Spontaneous Children\u2019s Speech. Logos Verlag, Berlin (2009)"},{"key":"28_CR20","unstructured":"Sterling, G., Prikhodko, P.: Deep learning in the problem of recognizing emotions from speech. In: Proceedings of the Conference Information Technologies and Systems 2016 IITP RAS. pp. 451\u2013456 (2016). (in Russian)"}],"container-title":["Lecture Notes in Computer Science","Distributed Computer and Communication Networks"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-66471-8_28","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,8,20]],"date-time":"2024-08-20T14:53:20Z","timestamp":1724165600000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-030-66471-8_28"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"ISBN":["9783030664701","9783030664718"],"references-count":20,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-66471-8_28","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2020]]},"assertion":[{"value":"2 January 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"DCCN","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Distributed Computer and Communication Networks","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Moscow","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Russia","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"14 September 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 September 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"dccn2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/dccn.ru\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"uConfy","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"167","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"97","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"58% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"7","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}