{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,14]],"date-time":"2026-02-14T03:13:51Z","timestamp":1771038831727,"version":"3.50.1"},"reference-count":229,"publisher":"Springer Science and Business Media LLC","issue":"19","license":[{"start":{"date-parts":[[2023,2,22]],"date-time":"2023-02-22T00:00:00Z","timestamp":1677024000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,2,22]],"date-time":"2023-02-22T00:00:00Z","timestamp":1677024000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"published-print":{"date-parts":[[2023,8]]},"DOI":"10.1007\/s11042-023-14656-y","type":"journal-article","created":{"date-parts":[[2023,2,22]],"date-time":"2023-02-22T10:03:38Z","timestamp":1677060218000},"page":"29307-29351","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":18,"title":["Trends in speech emotion recognition: a comprehensive survey"],"prefix":"10.1007","volume":"82","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-3542-1214","authenticated-orcid":false,"given":"Kamaldeep","family":"Kaur","sequence":"first","affiliation":[]},{"given":"Parminder","family":"Singh","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,2,22]]},"reference":[{"issue":"May","key":"14656_CR1","doi-asserted-by":"publisher","first-page":"19","DOI":"10.1016\/j.specom.2020.04.005","volume":"122","author":"L Abdel-Hamid","year":"2020","unstructured":"Abdel-Hamid L (2020) Egyptian Arabic speech emotion recognition using prosodic, spectral and wavelet features. Speech Commun 122(May):19\u201330. https:\/\/doi.org\/10.1016\/j.specom.2020.04.005","journal-title":"Speech Commun"},{"key":"14656_CR2","doi-asserted-by":"publisher","unstructured":"Abdelwahab M, Busso C (2019) \u201cActive Learning for Speech Emotion Recognition Using Deep Neural Network,\u201d 2019 8th International Conference on Affective Computing and Intelligent Interaction, ACII 2019, pp. 441\u2013447, https:\/\/doi.org\/10.1109\/ACII.2019.8925524.","DOI":"10.1109\/ACII.2019.8925524"},{"issue":"4","key":"14656_CR3","doi-asserted-by":"publisher","first-page":"433","DOI":"10.1002\/wics.101","volume":"2","author":"H Abdi","year":"2010","unstructured":"Abdi H, Williams LJ (2010) Principal component analysis. WIREs Comput Stat 2(4):433\u2013459. https:\/\/doi.org\/10.1002\/wics.101","journal-title":"WIREs Comput Stat"},{"key":"14656_CR4","doi-asserted-by":"publisher","unstructured":"Agrawal SS (2011) \u201cEmotions in Hindi speech- Analysis, perception and recognition,\u201d 2011 International Conference on Speech Database and Assessments, Oriental COCOSDA 2011 - Proceedings, pp. 7\u201313, https:\/\/doi.org\/10.1109\/ICSDA.2011.6085972.","DOI":"10.1109\/ICSDA.2011.6085972"},{"key":"14656_CR5","doi-asserted-by":"publisher","first-page":"56","DOI":"10.1016\/j.specom.2019.12.001","volume":"116","author":"MB Ak\u00e7ay","year":"2020","unstructured":"Ak\u00e7ay MB, O\u011fuz K (2020) Speech emotion recognition: emotional models, databases, features, preprocessing methods, supporting modalities, and classifiers. Speech Comm 116:56\u201376. https:\/\/doi.org\/10.1016\/j.specom.2019.12.001","journal-title":"Speech Comm"},{"key":"14656_CR6","doi-asserted-by":"publisher","unstructured":"Albawi S, Abed Mohammed T, Alzawi S (2017) Understanding of a convolutional neural network. In:\u00a02017 IEEE International Conference on Engineering and Technology (ICET). https:\/\/doi.org\/10.1109\/ICEngTechnol.2017.8308186","DOI":"10.1109\/ICEngTechnol.2017.8308186"},{"issue":"1","key":"14656_CR7","doi-asserted-by":"publisher","first-page":"43","DOI":"10.1109\/TAFFC.2015.2503757","volume":"8","author":"EM Albornoz","year":"2017","unstructured":"Albornoz EM, Milone DH (2017) Emotion recognition in never-seen languages using a novel ensemble method with emotion profiles. IEEE Trans Affect Comput 8(1):43\u201353. https:\/\/doi.org\/10.1109\/TAFFC.2015.2503757","journal-title":"IEEE Trans Affect Comput"},{"issue":"3","key":"14656_CR8","doi-asserted-by":"publisher","first-page":"556","DOI":"10.1016\/j.csl.2010.10.001","volume":"25","author":"EM Albornoz","year":"2011","unstructured":"Albornoz EM, Milone DH, Rufiner HL (2011) Spoken emotion recognition using hierarchical classifiers. Comput Speech Lang 25(3):556\u2013570. https:\/\/doi.org\/10.1016\/j.csl.2010.10.001","journal-title":"Comput Speech Lang"},{"issue":"2","key":"14656_CR9","doi-asserted-by":"publisher","first-page":"155","DOI":"10.1007\/s10462-012-9368-5","volume":"43","author":"CN Anagnostopoulos","year":"2012","unstructured":"Anagnostopoulos CN, Iliou T, Giannoukos I (2012) Features and classifiers for emotion recognition from speech: a survey from 2000 to 2011. Artif Intell Rev 43(2):155\u2013177. https:\/\/doi.org\/10.1007\/s10462-012-9368-5","journal-title":"Artif Intell Rev"},{"key":"14656_CR10","unstructured":"Balakrishnama S, Ganapathiraju A (1998) \u201cLinear Discriminant Analysis\u2014A Brief Tutorial,\u201d accessed on 10.09.2021"},{"key":"14656_CR11","doi-asserted-by":"publisher","unstructured":"Bansal S, Dev A (2013) \u201cEmotional hindi speech database,\u201d 2013 International Conference Oriental COCOSDA Held Jointly with 2013 Conference on Asian Spoken Language Research and Evaluation, O-COCOSDA\/CASLRE 2013, pp. 1\u20134, https:\/\/doi.org\/10.1109\/ICSDA.2013.6709867.","DOI":"10.1109\/ICSDA.2013.6709867"},{"key":"14656_CR12","first-page":"1865","volume":"03","author":"S Bansal","year":"2015","unstructured":"Bansal S, Dev A (2015) Emotional Hindi speech: Feature extraction and classification. 2015 2nd International Conference on Computing for Sustainable Global Development (INDIACom) 03:1865\u20131868","journal-title":"2015 2nd International Conference on Computing for Sustainable Global Development (INDIACom)"},{"issue":"2","key":"14656_CR13","doi-asserted-by":"publisher","first-page":"422","DOI":"10.1109\/78.348125","volume":"43","author":"F Beaufays","year":"1995","unstructured":"Beaufays F (1995) Transform-domain adaptive filters: an analytical approach. IEEE Trans Signal Process 43(2):422\u2013431. https:\/\/doi.org\/10.1109\/78.348125","journal-title":"IEEE Trans Signal Process"},{"key":"14656_CR14","doi-asserted-by":"publisher","first-page":"297","DOI":"10.1007\/978-981-10-8240-5","volume":"475","author":"S Bhattacharyya","year":"2018","unstructured":"Bhattacharyya S et al (2018) Speech Background Noise Removal Using Different Linear Filtering Techniques. Lect Notes Electr Eng 475:297\u2013307. https:\/\/doi.org\/10.1007\/978-981-10-8240-5","journal-title":"Lect Notes Electr Eng"},{"key":"14656_CR15","first-page":"341","volume":"5","author":"P Boersma","year":"2001","unstructured":"Boersma P, Weenink D (2001) PRAAT, a system for doing phonetics by computer. Glot Int 5:341\u2013345","journal-title":"Glot Int"},{"key":"14656_CR16","unstructured":"Boggs K, Liam (2017) Performance measures for machine learning, accessed on 11.08.2021"},{"issue":"4","key":"14656_CR17","doi-asserted-by":"publisher","first-page":"429","DOI":"10.1109\/89.848224","volume":"8","author":"SE Bou-Ghazale","year":"2000","unstructured":"Bou-Ghazale SE, Hansen JHL (Jul. 2000) A comparative study of traditional and newly proposed features for recognition of speech under stress. IEEE Transact Speech Aud Process 8(4):429\u2013442. https:\/\/doi.org\/10.1109\/89.848224","journal-title":"IEEE Transact Speech Aud Process"},{"key":"14656_CR18","unstructured":"Brookes M (1997) Voicebox: Speech processing toolbox for matlab. Imperial College, London. http:\/\/www.ee.ic.ac.uk\/hp\/staff\/dmb\/voicebox\/voicebox.html. Accessed 06.09.2021"},{"issue":"2","key":"14656_CR19","doi-asserted-by":"publisher","first-page":"121","DOI":"10.1023\/A:1009715923555","volume":"2","author":"CJC Burges","year":"1998","unstructured":"Burges CJC (1998) A tutorial on support vector machines for pattern recognition. Data Min Knowl Disc 2(2):121\u2013167. https:\/\/doi.org\/10.1023\/A:1009715923555","journal-title":"Data Min Knowl Disc"},{"key":"14656_CR20","doi-asserted-by":"crossref","unstructured":"Burkhardt F, Paeschke A, Rolfes M, Sendlmeier W, Weiss B (2005) \u201cA database of German emotional speech,\u201d in 9th European Conference on Speech Communication and Technology, , vol. 5, pp. 1517\u20131520","DOI":"10.21437\/Interspeech.2005-446"},{"key":"14656_CR21","doi-asserted-by":"publisher","first-page":"335","DOI":"10.1007\/s10579-008-9076-6","volume":"42","author":"C Busso","year":"2008","unstructured":"Busso C, Bulut M, Lee CC, Kazemzadeh A, Mower E, Kim S, Chang JN, Lee S, Narayanan SS (2008) IEMOCAP: interactive emotional dyadic motion capture database. Lang Resour Eval 42:335\u2013359. https:\/\/doi.org\/10.1007\/s10579-008-9076-6","journal-title":"Lang Resour Eval"},{"key":"14656_CR22","doi-asserted-by":"publisher","unstructured":"Busso C, Metallinou A, Narayanan SS (2011) \u201cIterative feature normalization for emotional speech detection,\u201d in 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 5692\u20135695, https:\/\/doi.org\/10.1109\/ICASSP.2011.5947652.","DOI":"10.1109\/ICASSP.2011.5947652"},{"key":"14656_CR23","unstructured":"C. academic of science Institute of automation (2005) CASIA-Chinese emotional speech corpus, Chin Linguist Data Consortium (CLDC). http:\/\/shachi.org\/resources\/27. Accessed 17 Oct\u00a02021"},{"issue":"1","key":"14656_CR24","doi-asserted-by":"publisher","first-page":"186","DOI":"10.1016\/j.csl.2014.01.003","volume":"29","author":"H Cao","year":"2015","unstructured":"Cao H, Verma R, Nenkova A (2015) Speaker-sensitive emotion recognition via ranking: studies on acted and spontaneous speech. Comput Speech Lang 29(1):186\u2013202. https:\/\/doi.org\/10.1016\/j.csl.2014.01.003","journal-title":"Comput Speech Lang"},{"issue":"9","key":"14656_CR25","doi-asserted-by":"publisher","first-page":"693","DOI":"10.1016\/j.specom.2010.04.002","volume":"52","author":"S Chakroborty","year":"2010","unstructured":"Chakroborty S, Saha G (2010) Feature selection using singular value decomposition and QR factorization with column pivoting for text-independent speaker identification. Speech Commun 52(9):693\u2013709. https:\/\/doi.org\/10.1016\/j.specom.2010.04.002","journal-title":"Speech Commun"},{"key":"14656_CR26","doi-asserted-by":"publisher","unstructured":"Chandrasekar P, Chapaneri S, Jayaswal D (2014) \u201cAutomatic speech emotion recognition: A survey,\u201d in 2014 International Conference on Circuits, Systems, Communication and Information Technology Applications, CSCITA 2014, pp. 341\u2013346, https:\/\/doi.org\/10.1109\/CSCITA.2014.6839284.","DOI":"10.1109\/CSCITA.2014.6839284"},{"key":"14656_CR27","doi-asserted-by":"publisher","unstructured":"Chen X, Jeong JC (2007) \u201cEnhanced recursive feature elimination,\u201d in Sixth International Conference on Machine Learning and Applications (ICMLA 2007), pp. 429\u2013435, https:\/\/doi.org\/10.1109\/ICMLA.2007.35.","DOI":"10.1109\/ICMLA.2007.35"},{"key":"14656_CR28","doi-asserted-by":"publisher","unstructured":"Chen Y, Xie J (2012) \u201cEmotional speech recognition based on SVM with GMM supervector,\u201d Journal of Electronics (China), vol. 29, https:\/\/doi.org\/10.1007\/s11767-012-0871-2.","DOI":"10.1007\/s11767-012-0871-2"},{"key":"14656_CR29","doi-asserted-by":"crossref","unstructured":"Chen C, You M, Song M, Bu J, Liu J (2006) \u201cAn Enhanced Speech Emotion Recognition System Based on Discourse Information BT - Computational Science \u2013 ICCS 2006,\u201d in ICCS, pp. 449\u2013456","DOI":"10.1007\/11758501_62"},{"key":"14656_CR30","doi-asserted-by":"publisher","unstructured":"Chen B, Yin Q, Guo P (2014) \u201cA study of deep belief network based Chinese speech emotion recognition,\u201d Proceedings - 2014 10th International Conference on Computational Intelligence and Security, CIS 2014, pp. 180\u2013184, https:\/\/doi.org\/10.1109\/CIS.2014.148.","DOI":"10.1109\/CIS.2014.148"},{"issue":"10","key":"14656_CR31","doi-asserted-by":"publisher","first-page":"1440","DOI":"10.1109\/LSP.2018.2860246","volume":"25","author":"M Chen","year":"2018","unstructured":"Chen M, He X, Yang J, Zhang H (2018) 3-D convolutional recurrent neural networks with attention model for speech emotion recognition. IEEE Signal Process Lett 25(10):1440\u20131444. https:\/\/doi.org\/10.1109\/LSP.2018.2860246","journal-title":"IEEE Signal Process Lett"},{"key":"14656_CR32","doi-asserted-by":"publisher","unstructured":"Chen Z, Jiang F, Cheng Y, Gu X, Liu W, Peng J (2018) \u201cXGBoost Classifier for DDoS Attack Detection and Analysis in SDN-Based Cloud,\u201d in 2018 IEEE international conference on big data and smart computing (BigComp), pp. 251\u2013256, https:\/\/doi.org\/10.1109\/BigComp.2018.00044.","DOI":"10.1109\/BigComp.2018.00044"},{"key":"14656_CR33","doi-asserted-by":"publisher","unstructured":"Chenchen Huang DF, Gong W, Wenlong F (2014) A research of speech emotion recognition based on deep belief network and SVM. Math Problems Eng, Article ID 749604. https:\/\/doi.org\/10.1155\/2014\/749604","DOI":"10.1155\/2014\/749604"},{"key":"14656_CR34","doi-asserted-by":"publisher","unstructured":"Chiu S, Tavella D (2008) Introduction to data mining. Data Min Market Intel Optimal Market Returns:137\u2013192. https:\/\/doi.org\/10.1016\/b978-0-7506-8234-3.00007-1","DOI":"10.1016\/b978-0-7506-8234-3.00007-1"},{"key":"14656_CR35","doi-asserted-by":"publisher","unstructured":"Choudhury AR, Ghosh A, Pandey R, Barman S (2018) \u201cEmotion recognition from speech signals using excitation source and spectral features,\u201d Proceedings of 2018 IEEE Applied Signal Processing Conference, ASPCON 2018, pp. 257\u2013261, https:\/\/doi.org\/10.1109\/ASPCON.2018.8748626.","DOI":"10.1109\/ASPCON.2018.8748626"},{"issue":"6","key":"14656_CR36","doi-asserted-by":"publisher","first-page":"487","DOI":"10.1016\/j.specom.2008.03.012","volume":"50","author":"C Clavel","year":"2008","unstructured":"Clavel C, Vasilescu I, Devillers L, Richard G, Ehrette T (2008) Fear-type emotion recognition for future audio-based surveillance systems. Speech Commun 50(6):487\u2013503. https:\/\/doi.org\/10.1016\/j.specom.2008.03.012","journal-title":"Speech Commun"},{"issue":"January","key":"14656_CR37","doi-asserted-by":"publisher","first-page":"35","DOI":"10.1016\/j.bica.2018.01.002","volume":"23","author":"RV Darekar","year":"2018","unstructured":"Darekar RV, Dhande AP (2018) Emotion recognition from Marathi speech database using adaptive artificial neural network. Biol Inspired Cogn Architect 23(January):35\u201342. https:\/\/doi.org\/10.1016\/j.bica.2018.01.002","journal-title":"Biol Inspired Cogn Architect"},{"key":"14656_CR38","doi-asserted-by":"publisher","unstructured":"Dellaert F, Polzin T, Waibel A (1996) \u201cRecognizing emotion in speech,\u201d in Proceeding of Fourth International Conference on Spoken Language Processing. ICSLP \u201896, vol. 3, pp. 1970\u20131973, https:\/\/doi.org\/10.1109\/ICSLP.1996.608022.","DOI":"10.1109\/ICSLP.1996.608022"},{"key":"14656_CR39","first-page":"34","volume-title":"Real-Life Emotion Recognition in Speech BT - Speaker Classification II: Selected Projects, C. M\u00fcller, Ed.","author":"L Devillers","year":"2007","unstructured":"Devillers L, Vidrascu L (2007) Real-Life Emotion Recognition in Speech BT - Speaker Classification II: Selected Projects, C. M\u00fcller, Ed. Springer Berlin Heidelberg, Berlin, Heidelberg, pp 34\u201342"},{"key":"14656_CR40","doi-asserted-by":"publisher","first-page":"200953","DOI":"10.1109\/ACCESS.2020.3035531","volume":"8","author":"A Dey","year":"2020","unstructured":"Dey A, Chattopadhyay S, Singh PK, Ahmadian A, Ferrara M, Sarkar R (2020) A hybrid Meta-heuristic feature selection method using Golden ratio and equilibrium optimization algorithms for speech emotion recognition. IEEE Access 8:200953\u2013200970. https:\/\/doi.org\/10.1109\/ACCESS.2020.3035531","journal-title":"IEEE Access"},{"key":"14656_CR41","unstructured":"Dhall A, Goecke R, Gedeon T (2011) Acted facial expressions in the wild database. Tech Rep, no, [Online]. Available: http:\/\/cs.anu.edu.au\/techreports\/. Accessed 27 Oct\u00a02021"},{"issue":"4","key":"14656_CR42","first-page":"462","volume":"19","author":"PEHRO Duda","year":"1973","unstructured":"Duda PEHRO, Hart PE, Duda RO (1973) Pattern classification and scene analysis. Leonardo 19(4):462\u2013463","journal-title":"Leonardo"},{"key":"14656_CR43","first-page":"182","volume":"39","author":"K Dupuis","year":"2011","unstructured":"Dupuis K, Pichora-Fuller M (2011) Recognition of emotional speech for younger and older talkers: Behavioural findings from the Toronto emotional speech set. Can Acoust Acoustique Canadienne 39:182\u2013183","journal-title":"Can Acoust Acoustique Canadienne"},{"issue":"3","key":"14656_CR44","doi-asserted-by":"publisher","first-page":"169","DOI":"10.1080\/02699939208411068","volume":"6","author":"P Ekman","year":"1992","unstructured":"Ekman P (1992) An argument for basic emotions. Cognit Emot 6(3):169\u2013200","journal-title":"Cognit Emot"},{"issue":"3","key":"14656_CR45","doi-asserted-by":"publisher","first-page":"572","DOI":"10.1016\/j.patcog.2010.09.020","volume":"44","author":"M El Ayadi","year":"2011","unstructured":"El Ayadi M, Kamel MS, Karray F (2011) Survey on speech emotion recognition: features, classification schemes, and databases. Pattern Recogn 44(3):572\u2013587. https:\/\/doi.org\/10.1016\/j.patcog.2010.09.020","journal-title":"Pattern Recogn"},{"key":"14656_CR46","doi-asserted-by":"crossref","unstructured":"Engberg IS, Hansen AV, Andersen O, Dalsgaard P (1997) \u201cDesign, recording and verification of a danish emotional speech database,\u201d in 5th European Conference on Speech Communication and Technology, Rhodes, Greece, pp. 1\u20134","DOI":"10.21437\/Eurospeech.1997-482"},{"key":"14656_CR47","doi-asserted-by":"publisher","unstructured":"Er MB (2020) \u201cA Novel Approach for Classification of Speech Emotions Based on Deep and Acoustic Features,\u201d IEEE Access, vol. 8, https:\/\/doi.org\/10.1109\/ACCESS.2020.3043201.","DOI":"10.1109\/ACCESS.2020.3043201"},{"key":"14656_CR48","unstructured":"Essentia Toolkit (n.d.) https:\/\/essentia.upf.edu. Accessed 16 Nov\u00a02021"},{"key":"14656_CR49","unstructured":"Eyben F (n.d.) Eight emotional speech databases. https:\/\/mediatum.ub.tum.de\/. Accessed 18 Nov\u00a02021"},{"issue":"4","key":"14656_CR50","doi-asserted-by":"publisher","first-page":"4","DOI":"10.1145\/2729095.2729097","volume":"6","author":"F Eyben","year":"2015","unstructured":"Eyben F, Schuller B (2015) OpenSMILE: the Munich open-source large-scale multimedia feature extractor. SIG Multimed Rec 6(4):4\u201313. https:\/\/doi.org\/10.1145\/2729095.2729097","journal-title":"SIG Multimed Rec"},{"key":"14656_CR51","doi-asserted-by":"publisher","unstructured":"Eyben F, W\u00f6llmer M, Schuller B (2009) \u201cOpenEAR \u2014 Introducing the munich open-source emotion and affect recognition toolkit,\u201d in 2009 3rd International Conference on Affective Computing and Intelligent Interaction and Workshops, pp. 1\u20136, https:\/\/doi.org\/10.1109\/ACII.2009.5349350.","DOI":"10.1109\/ACII.2009.5349350"},{"key":"14656_CR52","doi-asserted-by":"publisher","unstructured":"Eyben F, W\u00f6llmer M, Schuller B (2010) \u201cOpensmile: The Munich Versatile and Fast Open-Source Audio Feature Extractor,\u201d in Proceedings of the 18th ACM International Conference on Multimedia, pp. 1459\u20131462, https:\/\/doi.org\/10.1145\/1873951.1874246.","DOI":"10.1145\/1873951.1874246"},{"issue":"2","key":"14656_CR53","doi-asserted-by":"publisher","first-page":"190","DOI":"10.1109\/TAFFC.2015.2457417","volume":"7","author":"F Eyben","year":"2016","unstructured":"Eyben F, Scherer KR, Schuller BW, Sundberg J, Andre E, Busso C, Devillers LY, Epps J, Laukka P, Narayanan SS, Truong KP (2016) The Geneva minimalistic acoustic parameter set (GeMAPS) for voice research and affective computing. IEEE Trans Affect Comput 7(2):190\u2013202. https:\/\/doi.org\/10.1109\/TAFFC.2015.2457417","journal-title":"IEEE Trans Affect Comput"},{"key":"14656_CR54","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/access.2021.3054345","volume":"9","author":"M Ezz-Eldin","year":"2021","unstructured":"Ezz-Eldin M, Khalaf AAM, Hamed HFA, Hussein AI (2021) Efficient feature-aware hybrid model of deep learning architectures for speech emotion recognition. IEEE Access 9:1\u20131. https:\/\/doi.org\/10.1109\/access.2021.3054345","journal-title":"IEEE Access"},{"key":"14656_CR55","doi-asserted-by":"publisher","first-page":"105190","DOI":"10.1016\/j.knosys.2019.105190","volume":"191","author":"A Faramarzi","year":"2020","unstructured":"Faramarzi A, Heidarinejad M, Stephens B, Mirjalili S (2020) Equilibrium optimizer: A novel optimization algorithm. Knowl-Based Syst 191:105190. https:\/\/doi.org\/10.1016\/j.knosys.2019.105190","journal-title":"Knowl-Based Syst"},{"issue":"3","key":"14656_CR56","doi-asserted-by":"publisher","first-page":"553","DOI":"10.1007\/s10772-017-9426-0","volume":"20","author":"Z Farhoudi","year":"2017","unstructured":"Farhoudi Z, Setayeshi S, Rabiee A (2017) Using learning automata in brain emotional learning for speech emotion recognition. Int J Speech Technol 20(3):553\u2013562. https:\/\/doi.org\/10.1007\/s10772-017-9426-0","journal-title":"Int J Speech Technol"},{"key":"14656_CR57","doi-asserted-by":"publisher","first-page":"60","DOI":"10.1016\/j.neunet.2017.02.013","volume":"92","author":"HM Fayek","year":"2017","unstructured":"Fayek HM, Lech M, Cavedon L (2017) Evaluating deep learning architectures for speech emotion recognition. Neural Netw 92:60\u201368. https:\/\/doi.org\/10.1016\/j.neunet.2017.02.013","journal-title":"Neural Netw"},{"key":"14656_CR58","doi-asserted-by":"publisher","unstructured":"Ferdib-Al-Islam, L. Akter, and M. M. Islam (2021) \u201cHepatocellular Carcinoma Patient\u2019s Survival Prediction Using Oversampling and Machine Learning Techniques,\u201d Int Conf Robot Electr Signal Process Tech, pp. 445\u2013450, https:\/\/doi.org\/10.1109\/ICREST51555.2021.9331108.","DOI":"10.1109\/ICREST51555.2021.9331108"},{"issue":"1","key":"14656_CR59","doi-asserted-by":"publisher","first-page":"145","DOI":"10.1016\/S0167-6393(02)00080-8","volume":"40","author":"R Fernandez","year":"2003","unstructured":"Fernandez R, Picard RW (2003) Modeling drivers\u2019 speech under stress. Speech Commun 40(1):145\u2013159. https:\/\/doi.org\/10.1016\/S0167-6393(02)00080-8","journal-title":"Speech Commun"},{"issue":"1","key":"14656_CR60","doi-asserted-by":"publisher","first-page":"25","DOI":"10.1016\/j.patcog.2013.05.025","volume":"47","author":"A Fischer","year":"2014","unstructured":"Fischer A, Igel C (2014) Training restricted Boltzmann machines: An introduction. Pattern Recognit 47(1):25\u201339. https:\/\/doi.org\/10.1016\/j.patcog.2013.05.025","journal-title":"Pattern Recognit"},{"key":"14656_CR61","unstructured":"Fonti V (2017) Feature selection using LASSO. VU Amsterdam:1\u201326"},{"issue":"6","key":"14656_CR62","doi-asserted-by":"publisher","first-page":"671","DOI":"10.1109\/tpami.1983.4767461","volume":"5","author":"K Fukunaga","year":"1983","unstructured":"Fukunaga K, Mantock JM (1983) Nonparametric discriminant analysis. IEEE Trans Pattern Anal Mach Intell 5(6):671\u2013678. https:\/\/doi.org\/10.1109\/tpami.1983.4767461","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"12","key":"14656_CR63","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1371\/journal.pone.0144610","volume":"10","author":"T Giannakopoulos","year":"2015","unstructured":"Giannakopoulos T (2015) PyAudioAnalysis: An open-source python library for audio signal analysis. PLoS One 10(12):1\u201317. https:\/\/doi.org\/10.1371\/journal.pone.0144610","journal-title":"PLoS One"},{"key":"14656_CR64","doi-asserted-by":"publisher","unstructured":"Giannakopoulos T, Pikrakis A, Theodoridis S (2009) \u201cA dimensional approach to emotion recognition of speech from movies,\u201d in 2009 IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 65\u201368, https:\/\/doi.org\/10.1109\/ICASSP.2009.4959521.","DOI":"10.1109\/ICASSP.2009.4959521"},{"key":"14656_CR65","doi-asserted-by":"publisher","unstructured":"Gomes J, El-Sharkawy M (2015) \u201ci-Vector Algorithm with Gaussian Mixture Model for Efficient Speech Emotion Recognition,\u201d 2015 International Conference on Computational Science and Computational Intelligence (CSCI), pp. 476\u2013480, https:\/\/doi.org\/10.1109\/CSCI.2015.17.","DOI":"10.1109\/CSCI.2015.17"},{"key":"14656_CR66","doi-asserted-by":"publisher","unstructured":"Grimm M, Kroschel K, Narayanan S (2007) \u201cSupport Vector Regression for Automatic Recognition of Spontaneous Emotions in Speech,\u201d in 2007 IEEE International Conference on Acoustics, Speech and Signal Processing - ICASSP \u201807, vol. 4, pp. IV-1085-IV\u20131088, https:\/\/doi.org\/10.1109\/ICASSP.2007.367262.","DOI":"10.1109\/ICASSP.2007.367262"},{"key":"14656_CR67","doi-asserted-by":"crossref","unstructured":"Hansen JHL, Bou-Ghazale SE (1997) Getting started with SUSAS: a speech under simulated and actual stress database. https:\/\/catalog.ldc.upenn.edu\/LDC99S78. Accessed 28 Nov\u00a02021","DOI":"10.21437\/Eurospeech.1997-494"},{"key":"14656_CR68","doi-asserted-by":"publisher","unstructured":"Hifny Y, Ali A (2019) \u201cEfficient Arabic Emotion Recognition Using Deep Neural Networks,\u201d ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings, vol. 2019-May, pp. 6710\u20136714, https:\/\/doi.org\/10.1109\/ICASSP.2019.8683632.","DOI":"10.1109\/ICASSP.2019.8683632"},{"issue":"8","key":"14656_CR69","doi-asserted-by":"publisher","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","volume":"9","author":"S Hochreiter","year":"1997","unstructured":"Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735\u20131780. https:\/\/doi.org\/10.1162\/neco.1997.9.8.1735","journal-title":"Neural Comput"},{"key":"14656_CR70","doi-asserted-by":"publisher","first-page":"69","DOI":"10.1016\/j.inffus.2018.09.008","volume":"49","author":"MS Hossain","year":"2019","unstructured":"Hossain MS, Muhammad G (2019) Emotion recognition using deep learning approach from audio\u2013visual emotional big data. Inf Fus 49:69\u201378. https:\/\/doi.org\/10.1016\/j.inffus.2018.09.008","journal-title":"Inf Fus"},{"issue":"2","key":"14656_CR71","doi-asserted-by":"publisher","first-page":"01","DOI":"10.5121\/ijdkp.2015.5201","volume":"5","author":"M Hossin","year":"2015","unstructured":"Hossin M, Sulaiman MN (2015) A review on evaluation metrics for data classification evaluations. Int J Data Min Knowl Manag Process 5(2):01\u201311. https:\/\/doi.org\/10.5121\/ijdkp.2015.5201","journal-title":"Int J Data Min Knowl Manag Process"},{"key":"14656_CR72","unstructured":"Hozjan V, Kacic Z, Moreno A, Bonafonte A, Nogueiras A (2002) Interface databases: design and collection of a multilingual emotional speech database. http:\/\/www.lrec-conf.org\/proceedings\/lrec2002"},{"key":"14656_CR73","doi-asserted-by":"publisher","unstructured":"Huang R, Ma C (2006) \u201cToward A Speaker-Independent Real-Time Affect Detection System,\u201d in 18th International Conference on Pattern Recognition (ICPR\u201906), vol. 1, pp. 1204\u20131207, https:\/\/doi.org\/10.1109\/ICPR.2006.1127.","DOI":"10.1109\/ICPR.2006.1127"},{"key":"14656_CR74","unstructured":"Inger AVH, Engberg S (1996) Documentation of the danish emotional speech database DES. Aalborg. https:\/\/vbn.aau.dk\/en. Accessed 14 Aug\u00a02021"},{"key":"14656_CR75","doi-asserted-by":"publisher","unstructured":"Islam MM, Islam MR, Islam MS (2020) \u201cAn Efficient Human Computer Interaction through Hand Gesture Using Deep Convolutional Neural Network,\u201d SN Comput Sci, vol. 1, no. 4, https:\/\/doi.org\/10.1007\/s42979-020-00223-x.","DOI":"10.1007\/s42979-020-00223-x"},{"key":"14656_CR76","doi-asserted-by":"publisher","unstructured":"Islam MM, Islam MZ, Asraf A, Ding W (2020) \u201cDiagnosis of COVID-19 from X-rays using combined CNN-RNN architecture with transfer learning,\u201d medRxiv, https:\/\/doi.org\/10.1101\/2020.08.24.20181339.","DOI":"10.1101\/2020.08.24.20181339"},{"key":"14656_CR77","doi-asserted-by":"publisher","first-page":"94601","DOI":"10.1109\/ACCESS.2021.3091487","volume":"9","author":"MR Islam","year":"2021","unstructured":"Islam MR, Moni MA, Islam MM, Rashed-al-Mahfuz M, Islam MS, Hasan MK, Hossain MS, Ahmad M, Uddin S, Azad A, Alyami SA, Ahad MAR, Lio P (2021) Emotion recognition from EEG signal focusing on deep learning and shallow learning techniques. IEEE Access 9:94601\u201394624. https:\/\/doi.org\/10.1109\/ACCESS.2021.3091487","journal-title":"IEEE Access"},{"issue":"May","key":"14656_CR78","doi-asserted-by":"publisher","first-page":"104757","DOI":"10.1016\/j.compbiomed.2021.104757","volume":"136","author":"MR Islam","year":"2021","unstructured":"Islam MR et al (2021) EEG Channel Correlation Based Model for Emotion Recognition. Comput Biol Med 136(May):104757. https:\/\/doi.org\/10.1016\/j.compbiomed.2021.104757","journal-title":"Comput Biol Med"},{"key":"14656_CR79","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.bspc.2020.101894","volume":"59","author":"D Issa","year":"2020","unstructured":"Issa D, Fatih Demirci M, Yazici A (2020) Speech emotion recognition with deep convolutional neural networks. Biomed Signal Process Control 59:1\u201311. https:\/\/doi.org\/10.1016\/j.bspc.2020.101894","journal-title":"Biomed Signal Process Control"},{"key":"14656_CR80","unstructured":"Jackson P, Haq S (2011) Surrey Audio-Visual Expressed Emotion (SAVEE) database. http:\/\/kahlan.eps.surrey.ac.uk\/savee. Accessed 17 Sept\u00a02021"},{"key":"14656_CR81","doi-asserted-by":"publisher","unstructured":"Jaiswal JK, Samikannu R (2017) \u201cApplication of Random Forest Algorithm on Feature Subset Selection and Classification and Regression,\u201d in 2017 World congress on computing and communication technologies (WCCCT), pp. 65\u201368, https:\/\/doi.org\/10.1109\/WCCCT.2016.25.","DOI":"10.1109\/WCCCT.2016.25"},{"key":"14656_CR82","doi-asserted-by":"publisher","unstructured":"Jaratrotkamjorn A, Choksuriwong A (2019) \u201cBimodal Emotion Recognition using Deep Belief Network,\u201d ICSEC 2019 - 23rd International Computer Science and Engineering Conference, pp. 103\u2013109, https:\/\/doi.org\/10.1109\/ICSEC47112.2019.8974707.","DOI":"10.1109\/ICSEC47112.2019.8974707"},{"key":"14656_CR83","doi-asserted-by":"publisher","first-page":"90368","DOI":"10.1109\/ACCESS.2019.2927384","volume":"7","author":"P Jiang","year":"2019","unstructured":"Jiang P, Fu H, Tao H, Lei P, Zhao L (2019) Parallelized convolutional recurrent neural network with spectral features for speech emotion recognition. IEEE Access 7:90368\u201390377. https:\/\/doi.org\/10.1109\/ACCESS.2019.2927384","journal-title":"IEEE Access"},{"issue":"1","key":"14656_CR84","doi-asserted-by":"publisher","first-page":"112","DOI":"10.1049\/el.2014.3339","volume":"51","author":"Y Jin","year":"2015","unstructured":"Jin Y, Zha C, Zhao L, Song P (2015) Speech emotion recognition method based on hidden factor analysis. Electron Lett 51(1):112\u2013114. https:\/\/doi.org\/10.1049\/el.2014.3339","journal-title":"Electron Lett"},{"key":"14656_CR85","doi-asserted-by":"publisher","first-page":"216","DOI":"10.1016\/j.dsp.2017.10.016","volume":"72","author":"S Jing","year":"2018","unstructured":"Jing S, Mao X, Chen L (2018) Prominence features: effective emotional features for speech emotion recognition. Digit Signal Process A Rev J 72:216\u2013231. https:\/\/doi.org\/10.1016\/j.dsp.2017.10.016","journal-title":"Digit Signal Process A Rev J"},{"key":"14656_CR86","doi-asserted-by":"publisher","unstructured":"Kamble VV, Gaikwad BP, Rana DM (2014) \u201cSpontaneous emotion recognition for Marathi Spoken Words,\u201d International Conference on Communication and Signal Processing, ICCSP 2014 - Proceedings, pp. 1984\u20131990, https:\/\/doi.org\/10.1109\/ICCSP.2014.6950191.","DOI":"10.1109\/ICCSP.2014.6950191"},{"key":"14656_CR87","doi-asserted-by":"publisher","unstructured":"Kandali AB, Routray A, Basu TK (2008) \u201cEmotion recognition from Assamese speeches using MFCC features and GMM classifier,\u201d IEEE Region 10 Annual International Conference, Proceedings\/TENCON, https:\/\/doi.org\/10.1109\/TENCON.2008.4766487.","DOI":"10.1109\/TENCON.2008.4766487"},{"key":"14656_CR88","unstructured":"Kate Dupuis MKP-F (2010) Toronto emotional speech set (TESS). University of Toronto, Psychology Department. https:\/\/tspace.library.utoronto.ca\/handle\/1807\/24487. Accessed 08.10.2021"},{"key":"14656_CR89","doi-asserted-by":"publisher","unstructured":"Kattubadi IB, Garimella RM (2019) \u201cEmotion Classification: Novel Deep Learning Architectures,\u201d 2019 5th International Conference on Advanced Computing and Communication Systems, ICACCS 2019, pp. 285\u2013290, https:\/\/doi.org\/10.1109\/ICACCS.2019.8728519.","DOI":"10.1109\/ICACCS.2019.8728519"},{"key":"14656_CR90","doi-asserted-by":"publisher","first-page":"117327","DOI":"10.1109\/access.2019.2936124","volume":"7","author":"RA Khalil","year":"2019","unstructured":"Khalil RA, Jones E, Babar MI, Jan T, Zafar MH, Alhussain T (2019) Speech emotion recognition using deep learning techniques: a review. IEEE Access 7:117327\u2013117345. https:\/\/doi.org\/10.1109\/access.2019.2936124","journal-title":"IEEE Access"},{"key":"14656_CR91","doi-asserted-by":"publisher","unstructured":"Khan A, Islam M (2016) Deep belief networks. In: Proceedings of Introduction to Deep Neural Networks At: PIEAS, Islamabad, Pakistan. https:\/\/doi.org\/10.13140\/RG.2.2.17217.15200","DOI":"10.13140\/RG.2.2.17217.15200"},{"key":"14656_CR92","first-page":"1755","volume":"10","author":"DE King","year":"2009","unstructured":"King DE (2009) Dlib-ml: a machine learning toolkit. J Mach Learn Res 10:1755\u20131758","journal-title":"J Mach Learn Res"},{"issue":"2","key":"14656_CR93","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1007\/s10772-011-9125-1","volume":"15","author":"SG Koolagudi","year":"2012","unstructured":"Koolagudi SG, Rao KS (2012) Emotion recognition from speech: a review. Int J Speech Technol 15(2):99\u2013117. https:\/\/doi.org\/10.1007\/s10772-011-9125-1","journal-title":"Int J Speech Technol"},{"issue":"2","key":"14656_CR94","doi-asserted-by":"publisher","first-page":"265","DOI":"10.1007\/s10772-012-9139-3","volume":"15","author":"SG Koolagudi","year":"2012","unstructured":"Koolagudi SG, Rao KS (2012) Emotion recognition from speech using source, system, and prosodic features. Int J Speech Technol 15(2):265\u2013289. https:\/\/doi.org\/10.1007\/s10772-012-9139-3","journal-title":"Int J Speech Technol"},{"key":"14656_CR95","doi-asserted-by":"crossref","unstructured":"Koolagudi SG, Maity S, Kumar VA, Chakrabarti S, Rao KS (2009) \u201cIITKGP-SESC: Speech Database for Emotion Analysis,\u201d in Contemporary Computing, pp. 485\u2013492","DOI":"10.1007\/978-3-642-03547-0_46"},{"key":"14656_CR96","doi-asserted-by":"publisher","unstructured":"Koolagudi SG, Reddy R, Yadav J, Rao KS (2011) \u201cIITKGP-SEHSC\u00a0: Hindi speech corpus for emotion analysis,\u201d 2011 International Conference on Devices and Communications, ICDeCom 2011 - Proceedings, https:\/\/doi.org\/10.1109\/ICDECOM.2011.5738540.","DOI":"10.1109\/ICDECOM.2011.5738540"},{"issue":"1","key":"14656_CR97","doi-asserted-by":"publisher","first-page":"167","DOI":"10.1007\/s10772-018-9495-8","volume":"21","author":"SG Koolagudi","year":"2018","unstructured":"Koolagudi SG, Murthy YVS, Bhaskar SP (2018) Choice of a classifier, based on properties of a dataset: case study-speech emotion recognition. Int J Speech Technol 21(1):167\u2013183. https:\/\/doi.org\/10.1007\/s10772-018-9495-8","journal-title":"Int J Speech Technol"},{"issue":"2","key":"14656_CR98","doi-asserted-by":"publisher","first-page":"181","DOI":"10.1007\/s10772-012-9175-z","volume":"16","author":"SR Krothapalli","year":"2013","unstructured":"Krothapalli SR, Koolagudi SG (2013) Characterization and recognition of emotions from speech using excitation source information. Int J Speech Technol 16(2):181\u2013201. https:\/\/doi.org\/10.1007\/s10772-012-9175-z","journal-title":"Int J Speech Technol"},{"issue":"4","key":"14656_CR99","doi-asserted-by":"publisher","first-page":"409","DOI":"10.1017\/S0269888998214044","volume":"13","author":"M Kubat","year":"1999","unstructured":"Kubat M (1999) Neural networks: a comprehensive foundation. Knowl Eng Rev 13(4):409\u2013412. https:\/\/doi.org\/10.1017\/S0269888998214044","journal-title":"Knowl Eng Rev"},{"issue":"4","key":"14656_CR100","doi-asserted-by":"publisher","first-page":"401","DOI":"10.1007\/s10772-014-9239-3","volume":"17","author":"S Kuchibhotla","year":"2014","unstructured":"Kuchibhotla S, Vankayalapati HD, Vaddi RS, Anne KR (2014) A comparative analysis of classifiers in emotion recognition through acoustic features. Int J Speech Technol 17(4):401\u2013408. https:\/\/doi.org\/10.1007\/s10772-014-9239-3","journal-title":"Int J Speech Technol"},{"issue":"4","key":"14656_CR101","doi-asserted-by":"publisher","first-page":"657","DOI":"10.1007\/s10772-016-9358-0","volume":"19","author":"S Kuchibhotla","year":"2016","unstructured":"Kuchibhotla S, Deepthi H, Koteswara V, Anne R (2016) An optimal two stage feature selection for speech emotion recognition using acoustic features. Int J Speech Technol 19(4):657\u2013667. https:\/\/doi.org\/10.1007\/s10772-016-9358-0","journal-title":"Int J Speech Technol"},{"key":"14656_CR102","doi-asserted-by":"publisher","unstructured":"Kwon O-W, Chan K, Hao J, Lee T-W (2003) Emotion recognition by speech signals. In: 8th European Conference on Speech Communication and Technology. https:\/\/doi.org\/10.21437\/eurospeech.2003-80","DOI":"10.21437\/eurospeech.2003-80"},{"key":"14656_CR103","doi-asserted-by":"publisher","unstructured":"Lalitha S, Mudupu A, Nandyala BV, Munagala R (2015) \u201cSpeech emotion recognition using DWT,\u201d 2015 IEEE international conference on computational intelligence and computing research, ICCIC, 2016, https:\/\/doi.org\/10.1109\/ICCIC.2015.7435630.","DOI":"10.1109\/ICCIC.2015.7435630"},{"issue":"3","key":"14656_CR104","doi-asserted-by":"publisher","first-page":"497","DOI":"10.1007\/s10772-018-09572-8","volume":"22","author":"S Lalitha","year":"2019","unstructured":"Lalitha S, Tripathi S, Gupta D (2019) Enhanced speech emotion detection using deep neural networks. Int J Speech Technol 22(3):497\u2013510. https:\/\/doi.org\/10.1007\/s10772-018-09572-8","journal-title":"Int J Speech Technol"},{"key":"14656_CR105","unstructured":"Langley P, Iba W, Thompson K (1998) \u201cAn Analysis of Bayesian Classifiers,\u201d Proceedings of the Tenth National Conference on Artificial Intelligence, vol. 90"},{"key":"14656_CR106","doi-asserted-by":"publisher","first-page":"961","DOI":"10.1016\/j.asoc.2017.04.055","volume":"68","author":"C-Y Lee","year":"2018","unstructured":"Lee C-Y, Chen B-S (2018) Mutually-exclusive-and-collectively-exhaustive feature selection scheme. Appl Soft Comput 68:961\u2013971. https:\/\/doi.org\/10.1016\/j.asoc.2017.04.055","journal-title":"Appl Soft Comput"},{"key":"14656_CR107","first-page":"737","volume":"1","author":"C Lee","year":"2002","unstructured":"Lee C, Narayanan SS, Pieraccini R (2002) Classifying emotions in human-machine spoken dialogs. Proceed IEEE Int Conf Multimed Expo 1:737\u2013740","journal-title":"Proceed IEEE Int Conf Multimed Expo"},{"key":"14656_CR108","doi-asserted-by":"publisher","unstructured":"Lee KH, Kyun Choi H, Jang BT, Kim DH (2019) \u201cA Study on Speech Emotion Recognition Using a Deep Neural Network,\u201d ICTC 2019 - 10th International Conference on ICT Convergence: ICT Convergence Leading the Autonomous Future, pp. 1162\u20131165, https:\/\/doi.org\/10.1109\/ICTC46691.2019.8939830.","DOI":"10.1109\/ICTC46691.2019.8939830"},{"key":"14656_CR109","unstructured":"Li X (2007) SPEech Feature Toolbox (SPEFT) design and emotional speech feature extraction. https:\/\/epublications.marquette.edu\/theses\/1315. Accessed 25 Aug\u00a02021"},{"key":"14656_CR110","doi-asserted-by":"publisher","unstructured":"Li J, Fu X, Shao Z, Shang Y (2019) \u201cImprovement on Speech Depression Recognition Based on Deep Networks,\u201d Proceedings 2018 Chinese Automation Congress, CAC 2018, pp. 2705\u20132709, https:\/\/doi.org\/10.1109\/CAC.2018.8623055.","DOI":"10.1109\/CAC.2018.8623055"},{"key":"14656_CR111","doi-asserted-by":"publisher","unstructured":"Li Y, Baidoo C, Cai T, Kusi GA (2019) \u201cSpeech Emotion Recognition Using 1D CNN with No Attention,\u201d ICSEC 2019 - 23rd International Computer Science and Engineering Conference, pp. 351\u2013356, https:\/\/doi.org\/10.1109\/ICSEC47112.2019.8974716.","DOI":"10.1109\/ICSEC47112.2019.8974716"},{"key":"14656_CR112","doi-asserted-by":"publisher","unstructured":"Li Z, Li J, Ma S, Ren H (2019) \u201cSpeech emotion recognition based on residual neural network with different classifiers,\u201d Proceedings - 18th IEEE\/ACIS International Conference on Computer and Information Science, ICIS 2019, pp. 186\u2013190, https:\/\/doi.org\/10.1109\/ICIS46139.2019.8940308.","DOI":"10.1109\/ICIS46139.2019.8940308"},{"key":"14656_CR113","doi-asserted-by":"publisher","first-page":"11468","DOI":"10.1016\/j.eswa.2021.114683","volume":"173","author":"D Li","year":"2021","unstructured":"Li D, Liu J, Yang Z, Sun L, Wang Z (2021) Speech Emotion Recognition Using Recurrent Neural Networks with Directional Self-Attention. Exp Syst Appl 173:11468. https:\/\/doi.org\/10.1016\/j.eswa.2021.114683","journal-title":"Exp Syst Appl"},{"key":"14656_CR114","unstructured":"Liberman M (2002) Emotional prosody speech and transcripts. University of Pennsylvania. https:\/\/catalog.ldc.upenn.edu\/LDC2002S28. Accessed 14 Oct\u00a02021"},{"key":"14656_CR115","doi-asserted-by":"publisher","unstructured":"Lim W, Jang D, Lee T (2016) \u201cSpeech emotion recognition using convolutional and Recurrent Neural Networks,\u201d 2016 Asia-Pacific signal and information processing association annual summit and conference, APSIPA, 2017, https:\/\/doi.org\/10.1109\/APSIPA.2016.7820699.","DOI":"10.1109\/APSIPA.2016.7820699"},{"key":"14656_CR116","doi-asserted-by":"publisher","unstructured":"Litman DJ, Forbes-Riley K (2004) \u201cPredicting Student Emotions in Computer-Human Tutoring Dialogues,\u201d in Proceedings of the 42nd Annual Meeting on Association for Computational Linguistics, pp. 351\u2013358, https:\/\/doi.org\/10.3115\/1218955.1219000.","DOI":"10.3115\/1218955.1219000"},{"issue":"4","key":"14656_CR117","doi-asserted-by":"publisher","first-page":"20","DOI":"10.4018\/IJMCMC.2014100102","volume":"6","author":"Y Liu","year":"2014","unstructured":"Liu Y, Zhou Y, Wen S, Tang C (2014) A strategy on selecting performance metrics for classifier evaluation. Int J Mob Comput Multimed Commun 6(4):20\u201335. https:\/\/doi.org\/10.4018\/IJMCMC.2014100102","journal-title":"Int J Mob Comput Multimed Commun"},{"key":"14656_CR118","doi-asserted-by":"publisher","first-page":"101","DOI":"10.1016\/j.neucom.2017.05.097","volume":"277","author":"B Liu","year":"2018","unstructured":"Liu B, Zhou Y, Xia Z, Liu P, Yan Q, Xu H (2018) Spectral regression based marginal Fisher analysis dimensionality reduction algorithm. Neurocomputing 277:101\u2013107. https:\/\/doi.org\/10.1016\/j.neucom.2017.05.097","journal-title":"Neurocomputing"},{"key":"14656_CR119","doi-asserted-by":"publisher","first-page":"145","DOI":"10.1016\/j.neucom.2018.05.005","volume":"309","author":"ZT Liu","year":"2018","unstructured":"Liu ZT, Xie Q, Wu M, Cao WH, Mei Y, Mao JW (2018) Speech emotion recognition based on an improved brain emotion learning model. Neurocomputing 309:145\u2013156. https:\/\/doi.org\/10.1016\/j.neucom.2018.05.005","journal-title":"Neurocomputing"},{"key":"14656_CR120","doi-asserted-by":"publisher","first-page":"271","DOI":"10.1016\/j.neucom.2017.07.050","volume":"273","author":"ZT Liu","year":"2018","unstructured":"Liu ZT, Wu M, Cao WH, Mao JW, Xu JP, Tan GZ (2018) Speech emotion recognition based on feature selection and extreme learning machine decision tree. Neurocomputing 273:271\u2013280. https:\/\/doi.org\/10.1016\/j.neucom.2017.07.050","journal-title":"Neurocomputing"},{"key":"14656_CR121","doi-asserted-by":"publisher","first-page":"e0196391","DOI":"10.1371\/journal.pone.0196391","volume":"13","author":"S Livingstone","year":"2018","unstructured":"Livingstone S, Russo F (2018) The Ryerson audio-visual database of emotional speech and Song (RAVDESS): a dynamic, multimodal set of facial and vocal expressions in north American English. PLoS One 13:e0196391. https:\/\/doi.org\/10.1371\/journal.pone.0196391","journal-title":"PLoS One"},{"key":"14656_CR122","unstructured":"Loizou PC (1998) COLEA: A MATLAB software tool for speech analysis. https:\/\/ecs.utdallas.edu\/loizou\/speech\/colea.htm. Accessed 20 Oct\u00a02021"},{"issue":"4","key":"14656_CR123","doi-asserted-by":"publisher","first-page":"471","DOI":"10.1109\/TAFFC.2017.2736999","volume":"10","author":"R Lotfian","year":"2019","unstructured":"Lotfian R, Busso C (2019) Building naturalistic emotionally balanced speech Corpus by retrieving emotional speech from existing podcast recordings. IEEE Trans Affect Comput 10(4):471\u2013483. https:\/\/doi.org\/10.1109\/TAFFC.2017.2736999","journal-title":"IEEE Trans Affect Comput"},{"issue":"6","key":"14656_CR124","doi-asserted-by":"publisher","first-page":"490","DOI":"10.1109\/TMM.2010.2051872","volume":"12","author":"I Luengo","year":"2010","unstructured":"Luengo I, Navas E, Hern\u00e1ez I (2010) Feature analysis and evaluation for automatic emotion identification in speech. IEEE Trans Multimed 12(6):490\u2013501. https:\/\/doi.org\/10.1109\/TMM.2010.2051872","journal-title":"IEEE Trans Multimed"},{"key":"14656_CR125","doi-asserted-by":"publisher","unstructured":"Majkowski A, Ko\u0142odziej M, Rak RJ, Korczynski R (2016) \u201cClassification of emotions from speech signal,\u201d Signal Processing - Algorithms, Architectures, Arrangements, and Applications Conference Proceedings, SPA, pp. 276\u2013281, https:\/\/doi.org\/10.1109\/SPA.2016.7763627.","DOI":"10.1109\/SPA.2016.7763627"},{"key":"14656_CR126","first-page":"16","volume":"1","author":"R Manjunath","year":"2013","unstructured":"Manjunath R (2013) Dimensionality reduction and classification of color features data using svm and knn. Int J Image Process Vis Commun 1:16\u201321","journal-title":"Int J Image Process Vis Commun"},{"issue":"4","key":"14656_CR127","doi-asserted-by":"publisher","first-page":"779","DOI":"10.1007\/s10772-016-9368-y","volume":"19","author":"K Mannepalli","year":"2016","unstructured":"Mannepalli K, Sastry PN, Suman M (2016) FDBN: design and development of fractional deep belief networks for speaker emotion recognition. Int J Speech Technol 19(4):779\u2013790. https:\/\/doi.org\/10.1007\/s10772-016-9368-y","journal-title":"Int J Speech Technol"},{"issue":"1","key":"14656_CR128","doi-asserted-by":"publisher","first-page":"629","DOI":"10.1109\/TSMCB.2002.804363","volume":"34","author":"KZ Mao","year":"2004","unstructured":"Mao KZ (2004) Orthogonal forward selection and backward elimination algorithms for feature subset selection. IEEE Trans Syst Man, Cybern Part B (Cybernetics) 34(1):629\u2013634. https:\/\/doi.org\/10.1109\/TSMCB.2002.804363","journal-title":"IEEE Trans Syst Man, Cybern Part B (Cybernetics)"},{"issue":"8","key":"14656_CR129","doi-asserted-by":"publisher","first-page":"2324","DOI":"10.1587\/transinf.E93.D.2324","volume":"E93-D","author":"X Mao","year":"2010","unstructured":"Mao X, Chen L (2010) Speech emotion recognition based on parametric filter and fractal dimension. IEICE Trans Inf Syst E93-D(8):2324\u20132326. https:\/\/doi.org\/10.1587\/transinf.E93.D.2324","journal-title":"IEICE Trans Inf Syst"},{"issue":"8","key":"14656_CR130","doi-asserted-by":"publisher","first-page":"2203","DOI":"10.1109\/TMM.2014.2360798","volume":"16","author":"Q Mao","year":"2014","unstructured":"Mao Q, Dong M, Huang Z, Zhan Y (2014) Learning salient features for speech emotion recognition using convolutional neural networks. IEEE Trans Multimed 16(8):2203\u20132213. https:\/\/doi.org\/10.1109\/TMM.2014.2360798","journal-title":"IEEE Trans Multimed"},{"issue":"1","key":"14656_CR131","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1109\/TIT.1963.1057810","volume":"9","author":"T Marill","year":"1963","unstructured":"Marill T, Green D (1963) On the effectiveness of receptors in recognition systems. IEEE Trans Inf Theory 9(1):11\u201317. https:\/\/doi.org\/10.1109\/TIT.1963.1057810","journal-title":"IEEE Trans Inf Theory"},{"key":"14656_CR132","doi-asserted-by":"publisher","unstructured":"Martin O, Kotsia I, Macq B, Pitas I (2006) \u201cThe eNTERFACE\u2019 05 audio-visual emotion database,\u201dhttps:\/\/doi.org\/10.1109\/ICDEW.2006.145.","DOI":"10.1109\/ICDEW.2006.145"},{"issue":"2","key":"14656_CR133","doi-asserted-by":"publisher","first-page":"228","DOI":"10.1109\/34.908974","volume":"23","author":"AM Martinez","year":"2001","unstructured":"Martinez AM, Kak AC (Feb. 2001) PCA versus LDA. IEEE Trans Pattern Anal Mach Intell 23(2):228\u2013233. https:\/\/doi.org\/10.1109\/34.908974","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"14656_CR134","doi-asserted-by":"publisher","unstructured":"McFee B et al. (2015) \u201clibrosa: Audio and Music Signal Analysis in Python,\u201d Proceedings of the 14th Python in Science Conference, no. Scipy, pp. 18\u201324, https:\/\/doi.org\/10.25080\/majora-7b98e3ed-003.","DOI":"10.25080\/majora-7b98e3ed-003"},{"key":"14656_CR135","unstructured":"Meftah A, Alotaibi Y, Selouani SA (2014) Designing, building, and analyzing an Arabic speech emotional Corpus. In: Ninth International Conference on Language Resources and Evaluation at: Reykjavik, Iceland"},{"key":"14656_CR136","doi-asserted-by":"crossref","unstructured":"Meftah A, Alotaibi Y, Selouani S-A (2016) \u201cEmotional Speech Recognition: A Multilingual Perspective,\u201d 2016 International Conference on Bio-Engineering for Smart Technologies(Biosmart)","DOI":"10.1109\/BIOSMART.2016.7835600"},{"issue":"3","key":"14656_CR137","doi-asserted-by":"publisher","first-page":"727","DOI":"10.1016\/j.csl.2013.08.004","volume":"28","author":"A Milton","year":"2014","unstructured":"Milton A, Tamil Selvi S (2014) Class-specific multiple classifiers scheme to recognize emotions from speech signals. Comput Speech Lang 28(3):727\u2013742. https:\/\/doi.org\/10.1016\/j.csl.2013.08.004","journal-title":"Comput Speech Lang"},{"key":"14656_CR138","unstructured":"Mohanta A, Sharma U (2016) \u201cBengali Speech Emotion Recognition,\u201d in 2016 3rd International Conference on Computing for Sustainable Global Development (INDIACom), pp. 2812\u20132814"},{"key":"14656_CR139","doi-asserted-by":"publisher","unstructured":"Montenegro CS, Maravillas EA (2015) \u201cAcoustic-prosodic recognition of emotion in speech,\u201d 8th International Conference on Humanoid, Nanotechnology, Information Technology, Communication and Control, Environment and Management, HNICEM, 2016, https:\/\/doi.org\/10.1109\/HNICEM.2015.7393229.","DOI":"10.1109\/HNICEM.2015.7393229"},{"key":"14656_CR140","doi-asserted-by":"publisher","unstructured":"Morin D (2004) \u201cBeads-on-A-string,\u201d Encyclopedic Dictionary of Genetics, Genomics and Proteomics, pp. 1\u201322, https:\/\/doi.org\/10.1002\/0471684228.egp01270.","DOI":"10.1002\/0471684228.egp01270"},{"issue":"1","key":"14656_CR141","doi-asserted-by":"publisher","first-page":"137","DOI":"10.1007\/s10772-018-9493-x","volume":"21","author":"MB Mustafa","year":"2018","unstructured":"Mustafa MB, Yusoof MAM, Don ZM, Malekzadeh M (2018) Speech emotion recognition research: an analysis of research focus. Int J Speech Technol 21(1):137\u2013156. https:\/\/doi.org\/10.1007\/s10772-018-9493-x","journal-title":"Int J Speech Technol"},{"issue":"1","key":"14656_CR142","doi-asserted-by":"publisher","first-page":"24","DOI":"10.1016\/j.procs.2015.04.223","volume":"49","author":"VV Nanavare","year":"2015","unstructured":"Nanavare VV, Jagtap SK (2015) Recognition of human emotions from speech processing. Proced Comput Sci 49(1):24\u201332. https:\/\/doi.org\/10.1016\/j.procs.2015.04.223","journal-title":"Proced Comput Sci"},{"issue":"2","key":"14656_CR143","doi-asserted-by":"publisher","first-page":"1117","DOI":"10.1007\/s00500-019-03949-w","volume":"24","author":"AF Nematollahi","year":"2020","unstructured":"Nematollahi AF, Rahiminejad A, Vahidi B (2020) A novel meta-heuristic optimization method based on golden ratio in nature. Soft Comput 24(2):1117\u20131151. https:\/\/doi.org\/10.1007\/s00500-019-03949-w","journal-title":"Soft Comput"},{"issue":"4","key":"14656_CR144","doi-asserted-by":"publisher","first-page":"290","DOI":"10.1007\/s005210070006","volume":"9","author":"J Nicholson","year":"2000","unstructured":"Nicholson J, Takahashi K, Nakatsu R (2000) Emotion recognition in speech using neural networks. Neural Comput Applic 9(4):290\u2013296. https:\/\/doi.org\/10.1007\/s005210070006","journal-title":"Neural Comput Applic"},{"key":"14656_CR145","unstructured":"Nooteboom S (1997) The prosody of speech: Melody and rhythm. Handbook Phon Sci, vol 5"},{"key":"14656_CR146","doi-asserted-by":"publisher","unstructured":"O\u2019Shea K, Nash R (2015) An introduction to convolutional neural networks. ArXiv e-prints. https:\/\/doi.org\/10.48550\/arXiv.1511.08458","DOI":"10.48550\/arXiv.1511.08458"},{"key":"14656_CR147","doi-asserted-by":"publisher","DOI":"10.1017\/CBO9780511571299","volume-title":"The cognitive structure of emotions","author":"A Ortony","year":"1988","unstructured":"Ortony A, Clore GL, Collins A (1988) The cognitive structure of emotions. Cambridge University Press, Cambridge"},{"issue":"February","key":"14656_CR148","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.apacoust.2018.02.009","volume":"136","author":"T \u00d6zseven","year":"2018","unstructured":"\u00d6zseven T, D\u00fc\u011fenci M (2018) SPeech ACoustic (SPAC): A novel tool for speech feature extraction and classification. Appl Acoust 136(February):1\u20138. https:\/\/doi.org\/10.1016\/j.apacoust.2018.02.009","journal-title":"Appl Acoust"},{"key":"14656_CR149","doi-asserted-by":"publisher","first-page":"127","DOI":"10.1109\/ICDSBA.2018.00030","volume":"20","author":"HK Palo","year":"2018","unstructured":"Palo HK, Sagar S (2018) Comparison of Neural Network Models for Speech Emotion Recognition. Proceed 2nd Int Conf Data Sci Bus Anal ICDSBA 20:127\u2013131. https:\/\/doi.org\/10.1109\/ICDSBA.2018.00030","journal-title":"Proceed 2nd Int Conf Data Sci Bus Anal ICDSBA"},{"issue":"1","key":"14656_CR150","doi-asserted-by":"publisher","first-page":"135","DOI":"10.1007\/s10772-016-9333-9","volume":"19","author":"HK Palo","year":"2016","unstructured":"Palo HK, Mohanty MN, Chandra M (2016) Efficient feature combination techniques for emotional speech classification. Int J Speech Technol 19(1):135\u2013150. https:\/\/doi.org\/10.1007\/s10772-016-9333-9","journal-title":"Int J Speech Technol"},{"key":"14656_CR151","doi-asserted-by":"publisher","unstructured":"Pandey SK, Shekhawat HS, Prasanna SRM (2019) \u201cDeep Learning Techniques for Speech Emotion Recognition: A Review,\u201d in 2019 29th international conference RADIOELEKTRONIKA (RADIOELEKTRONIKA), pp. 1\u20136, https:\/\/doi.org\/10.1109\/RADIOELEK.2019.8733432.","DOI":"10.1109\/RADIOELEK.2019.8733432"},{"key":"14656_CR152","doi-asserted-by":"publisher","unstructured":"Partila P, Tovarek J, Voznak M, Rozhon J, Sevcik L, Baran R (2018) \u201cMulti-Classifier Speech Emotion Recognition System,\u201d 2018 26th Telecommunications Forum, TELFOR 2018 - Proceedings, pp. 1\u20134, https:\/\/doi.org\/10.1109\/TELFOR.2018.8612050.","DOI":"10.1109\/TELFOR.2018.8612050"},{"key":"14656_CR153","doi-asserted-by":"publisher","unstructured":"Pathak BV, Patil DR, More SD, Mhetre NR (2019) \u201cComparison between five classification techniques for classifying emotions in human speech,\u201d 2019 International Conference on Intelligent Computing and Control Systems, ICCS 2019, pp. 201\u2013207, https:\/\/doi.org\/10.1109\/ICCS45141.2019.9065620.","DOI":"10.1109\/ICCS45141.2019.9065620"},{"key":"14656_CR154","unstructured":"Pedregosa F et al. (2012) \u201cScikit-learn: Machine Learning in Python,\u201d J Mach Learn Res, vol. 12"},{"key":"14656_CR155","unstructured":"Petrushin V (2000) \u201cEmotion in speech: recognition and application to call centers,\u201d Proceedings of Artificial Neural Networks in Engineering"},{"key":"14656_CR156","doi-asserted-by":"crossref","unstructured":"Picard RW (1997) Affective computing. MIT Press. https:\/\/direct.mit.edu\/books\/book\/4296\/Affective-Computing. Accessed 24 Jun\u00a02021","DOI":"10.1037\/e526112012-054"},{"key":"14656_CR157","doi-asserted-by":"publisher","unstructured":"Pratiwi O, Rahardjo B, Supangkat S (2015) \u201cAttribute Selection Based on Information Gain for Automatic Grouping Student System,\u201d in Communications in Computer and Information Science, vol. 516, pp. 205\u2013211, https:\/\/doi.org\/10.1007\/978-3-662-46742-8_19.","DOI":"10.1007\/978-3-662-46742-8_19"},{"key":"14656_CR158","doi-asserted-by":"publisher","unstructured":"Prinz J (2004) Which emotions are basic? Oxford University Press. https:\/\/doi.org\/10.1093\/acprof:oso\/9780198528975.003.0004","DOI":"10.1093\/acprof:oso\/9780198528975.003.0004"},{"key":"14656_CR159","unstructured":"Pudil S, Pavel, N, Jana, Bl\u00e1ha (1991) \u201cStatistical approach to pattern recognition: Theory and practical solution by means of PREDITAS system,\u201d Kybernetika 27, vol. 1, no. 76"},{"key":"14656_CR160","doi-asserted-by":"publisher","unstructured":"Pyrczak F, Oh DM, Pyrczak F, Oh DM (2019) \u201cIntroduction to the t test,\u201d https:\/\/doi.org\/10.4324\/9781315179803-28.","DOI":"10.4324\/9781315179803-28"},{"key":"14656_CR161","unstructured":"Qayyum CSABA, Arefeen A (2019) \u201cConvolutional Neural Network ( CNN ) Based Speech Recognition,\u201d in 2019IEEE International Conference onSignal Processing, Information, Communication & Systems(SPICSCON, pp. 122\u2013125"},{"issue":"2","key":"14656_CR162","doi-asserted-by":"publisher","first-page":"257","DOI":"10.1109\/5.18626","volume":"77","author":"LR Rabiner","year":"1989","unstructured":"Rabiner LR (1989) A tutorial on hidden Markov models and selected applications in speech recognition. Proc IEEE 77(2):257\u2013286. https:\/\/doi.org\/10.1109\/5.18626","journal-title":"Proc IEEE"},{"key":"14656_CR163","doi-asserted-by":"publisher","unstructured":"Rahman MM, Islam MM, Manik MMH, Islam MR, Al-Rakhami MS (2021) \u201cMachine Learning Approaches for Tackling Novel Coronavirus (COVID-19) Pandemic,\u201d SN Comput Sci, vol. 2, no. 5, https:\/\/doi.org\/10.1007\/s42979-021-00774-7.","DOI":"10.1007\/s42979-021-00774-7"},{"key":"14656_CR164","doi-asserted-by":"publisher","first-page":"1097","DOI":"10.1016\/j.protcy.2016.05.242","volume":"24","author":"TM Rajisha","year":"2016","unstructured":"Rajisha TM, Sunija AP, Riyas KS (2016) Performance analysis of Malayalam language speech emotion recognition system using ANN\/SVM. Proced Technol 24:1097\u20131104. https:\/\/doi.org\/10.1016\/j.protcy.2016.05.242","journal-title":"Proced Technol"},{"key":"14656_CR165","doi-asserted-by":"publisher","unstructured":"Rajoo R, Aun CC (2016) \u201cInfluences of languages in speech emotion recognition: A comparative study using Malay, English and Mandarin languages,\u201d ISCAIE 2016\u20132016 IEEE Symposium on Computer Applications and Industrial Electronics, pp. 35\u201339, https:\/\/doi.org\/10.1109\/ISCAIE.2016.7575033.","DOI":"10.1109\/ISCAIE.2016.7575033"},{"key":"14656_CR166","doi-asserted-by":"publisher","unstructured":"Ram CS, Ponnusamy R (2014) \u201cAn effective automatic speech emotion recognition for Tamil language using Support Vector Machine,\u201d in 2014 International Conference on Issues and Challenges in Intelligent Computing Techniques (ICICT), pp. 19\u201323, https:\/\/doi.org\/10.1109\/ICICICT.2014.6781245.","DOI":"10.1109\/ICICICT.2014.6781245"},{"issue":"2","key":"14656_CR167","doi-asserted-by":"publisher","first-page":"143","DOI":"10.1007\/s10772-012-9172-2","volume":"16","author":"KS Rao","year":"2013","unstructured":"Rao KS, Koolagudi SG, Vempada RR (2013) Emotion recognition from speech using global and local prosodic features. Int J Speech Technol 16(2):143\u2013160. https:\/\/doi.org\/10.1007\/s10772-012-9172-2","journal-title":"Int J Speech Technol"},{"issue":"3","key":"14656_CR168","doi-asserted-by":"publisher","first-page":"150","DOI":"10.1016\/j.visinf.2019.10.003","volume":"3","author":"M Ren","year":"2019","unstructured":"Ren M, Nie W, Liu A, Su Y (2019) Multi-modal correlated network for emotion recognition in speech. Vis Inf 3(3):150\u2013155. https:\/\/doi.org\/10.1016\/j.visinf.2019.10.003","journal-title":"Vis Inf"},{"issue":"3","key":"14656_CR169","doi-asserted-by":"publisher","first-page":"473","DOI":"10.1007\/s10772-018-9533-6","volume":"22","author":"A Revathi","year":"2019","unstructured":"Revathi A, Jeyalakshmi C (2019) Emotions recognition: different sets of features and models. Int J Speech Technol 22(3):473\u2013482. https:\/\/doi.org\/10.1007\/s10772-018-9533-6","journal-title":"Int J Speech Technol"},{"issue":"1","key":"14656_CR170","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1186\/s40537-021-00428-8","volume":"8","author":"M Roccetti","year":"2021","unstructured":"Roccetti M, Delnevo G, Casini L, Mirri S (2021) An alternative approach to dimension reduction for pareto distributed data: a case study. J Big Data 8(1):39. https:\/\/doi.org\/10.1186\/s40537-021-00428-8","journal-title":"J Big Data"},{"issue":"3","key":"14656_CR171","doi-asserted-by":"publisher","first-page":"315","DOI":"10.1016\/j.ipm.2008.09.003","volume":"45","author":"J Rong","year":"2009","unstructured":"Rong J, Li G, Chen YPP (2009) Acoustic feature selection for automatic emotion recognition from speech. Inf Process Manag 45(3):315\u2013328. https:\/\/doi.org\/10.1016\/j.ipm.2008.09.003","journal-title":"Inf Process Manag"},{"key":"14656_CR172","doi-asserted-by":"publisher","unstructured":"Roubos H, Setnes M, Abonyi J (2000) Learning fuzzy classification rules from data, vol 150. In: Proceedings Developments in Soft Computing, pp 108\u2013115. https:\/\/doi.org\/10.1007\/978-3-7908-1829-1_13","DOI":"10.1007\/978-3-7908-1829-1_13"},{"key":"14656_CR173","unstructured":"Ryerson RU (2017) Multimedia research lab. RML Emotion Database. http:\/\/shachi.org\/resources\/4965. Accessed 30 Oct\u00a02021"},{"key":"14656_CR174","doi-asserted-by":"publisher","unstructured":"Sadeghyan S (2018) A new robust feature selection method using variance-based sensitivity analysis. arXiv. https:\/\/doi.org\/10.48550\/arXiv.1804.05092","DOI":"10.48550\/arXiv.1804.05092"},{"key":"14656_CR175","doi-asserted-by":"crossref","unstructured":"Sari H, Cochet PY (1996) \u201cTransform-Domain Signal Processing in Digital Communications,\u201d in Signal Processing in Telecommunications, pp. 374\u2013384","DOI":"10.1007\/978-1-4471-1013-2_29"},{"issue":"4","key":"14656_CR176","doi-asserted-by":"publisher","first-page":"778","DOI":"10.1109\/TASLP.2014.2303296","volume":"22","author":"R Sarikaya","year":"2014","unstructured":"Sarikaya R, Hinton GE, Deoras A (2014) Application of deep belief networks for natural language understanding. IEEE Trans Audio Speech Lang Process 22(4):778\u2013784. https:\/\/doi.org\/10.1109\/TASLP.2014.2303296","journal-title":"IEEE Trans Audio Speech Lang Process"},{"key":"14656_CR177","doi-asserted-by":"publisher","unstructured":"Savargiv M, Bastanfard A (2013) \u201cText material design for fuzzy emotional speech corpus based on persian semantic and structure,\u201d in 2013 International Conference on Fuzzy Theory and Its Applications (iFUZZY), pp. 380\u2013384, https:\/\/doi.org\/10.1109\/iFuzzy.2013.6825469.","DOI":"10.1109\/iFuzzy.2013.6825469"},{"key":"14656_CR178","doi-asserted-by":"publisher","unstructured":"Savargiv M, Bastanfard A (2015) \u201cPersian speech emotion recognition,\u201d in 2015 7th Conference on Information and Knowledge Technology (IKT), pp. 1\u20135, https:\/\/doi.org\/10.1109\/IKT.2015.7288756.","DOI":"10.1109\/IKT.2015.7288756"},{"issue":"2","key":"14656_CR179","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1037\/h0054570","volume":"61","author":"H Schlosberg","year":"1954","unstructured":"Schlosberg H (1954) Three dimensions of emotion. Psychol Rev 61(2):81\u201388","journal-title":"Psychol Rev"},{"key":"14656_CR180","doi-asserted-by":"publisher","first-page":"85","DOI":"10.1016\/j.neunet.2014.09.003","volume":"61","author":"J Schmidhuber","year":"2015","unstructured":"Schmidhuber J (2015) Deep learning in neural networks: An overview. Neural Netw 61:85\u2013117. https:\/\/doi.org\/10.1016\/j.neunet.2014.09.003","journal-title":"Neural Netw"},{"key":"14656_CR181","volume-title":"English intonation:its form and function","author":"M Schubiger","year":"1958","unstructured":"Schubiger M (1958) English intonation:its form and function. M. Niemeyer Verlag, T\u00fcbingen"},{"issue":"9\u201310","key":"14656_CR182","doi-asserted-by":"publisher","first-page":"1062","DOI":"10.1016\/j.specom.2011.01.011","volume":"53","author":"B Schuller","year":"2011","unstructured":"Schuller B, Batliner A, Steidl S, Seppi D (2011) Recognising realistic emotions and affect in speech: state of the art and lessons learnt from the first challenge. Speech Comm 53(9\u201310):1062\u20131087. https:\/\/doi.org\/10.1016\/j.specom.2011.01.011","journal-title":"Speech Comm"},{"key":"14656_CR183","doi-asserted-by":"publisher","unstructured":"Schuller B et al (2013) The INTERSPEECH 2013 computational paralinguistics challenge: social signals, conflict, emotion, autism. In: Proceedings 14th Annual Conference of the International Speech Communication Association. https:\/\/doi.org\/10.21437\/Interspeech.2013-56","DOI":"10.21437\/Interspeech.2013-56"},{"key":"14656_CR184","doi-asserted-by":"publisher","first-page":"10295","DOI":"10.1016\/j.dsp.2020.102951","volume":"110","author":"M Shah Fahad","year":"2021","unstructured":"Shah Fahad M, Ranjan A, Yadav J, Deepak A (2021) A survey of speech emotion recognition in natural environment. Digit Signal Process A Rev J 110:10295. https:\/\/doi.org\/10.1016\/j.dsp.2020.102951","journal-title":"Digit Signal Process A Rev J"},{"key":"14656_CR185","doi-asserted-by":"publisher","first-page":"26777","DOI":"10.1109\/ACCESS.2019.2901352","volume":"7","author":"I Shahin","year":"2019","unstructured":"Shahin I, Nassif AB, Hamsa S (2019) Emotion recognition using hybrid Gaussian mixture model and deep neural network. IEEE Access 7:26777\u201326787. https:\/\/doi.org\/10.1109\/ACCESS.2019.2901352","journal-title":"IEEE Access"},{"issue":"1","key":"14656_CR186","doi-asserted-by":"publisher","first-page":"215","DOI":"10.1007\/s00521-012-0814-8","volume":"23","author":"M Sheikhan","year":"2013","unstructured":"Sheikhan M, Bejani M, Gharavian D (2013) Modular neural-SVM scheme for speech emotion recognition using ANOVA feature selection method. Neural Comput & Applic 23(1):215\u2013227. https:\/\/doi.org\/10.1007\/s00521-012-0814-8","journal-title":"Neural Comput & Applic"},{"key":"14656_CR187","doi-asserted-by":"publisher","unstructured":"Spola\u00f4r N, Cherman EA, Monard MC, Lee HD (2013) \u201cReliefF for multi-label feature selection,\u201d Proceedings - 2013 Brazilian Conference on Intelligent Systems, BRACIS 2013, pp. 6\u201311, https:\/\/doi.org\/10.1109\/BRACIS.2013.10.","DOI":"10.1109\/BRACIS.2013.10"},{"key":"14656_CR188","unstructured":"Steidl S (2009) Automatic classification of emotion related user states in spontaneous children\u2019s speech. Logos Verlag. http:\/\/www5.informatik.uni-erlangen.de. Accessed 03.06.2021"},{"key":"14656_CR189","doi-asserted-by":"publisher","unstructured":"Suganya S, Charles E (2019) \u201cSpeech emotion recognition using deep learning on audio recordings,\u201dhttps:\/\/doi.org\/10.1109\/ICTer48817.2019.9023737.","DOI":"10.1109\/ICTer48817.2019.9023737"},{"issue":"4","key":"14656_CR190","doi-asserted-by":"publisher","first-page":"687","DOI":"10.1142\/S0218001409007326","volume":"23","author":"Y Sun","year":"2009","unstructured":"Sun Y, Wong AKC, Kamel MS (2009) Classification of imbalanced data: a review. Int J Pattern Recognit Artif Intell 23(4):687\u2013719. https:\/\/doi.org\/10.1142\/S0218001409007326","journal-title":"Int J Pattern Recognit Artif Intell"},{"key":"14656_CR191","doi-asserted-by":"publisher","unstructured":"Sun L, Zou B, Fu S, Chen J, Wang F (2019) \u201cSpeech emotion recognition based on DNN-decision tree SVM model,\u201d Speech Commun, https:\/\/doi.org\/10.1016\/j.specom.2019.10.004.","DOI":"10.1016\/j.specom.2019.10.004"},{"key":"14656_CR192","doi-asserted-by":"publisher","unstructured":"Swain M, Routray A, Kabisatpathy P, Kundu JN (2017) \u201cStudy of prosodic feature extraction for multidialectal Odia speech emotion recognition,\u201d IEEE Region 10 Annual International Conference, Proceedings\/TENCON, pp 1644\u20131649, https:\/\/doi.org\/10.1109\/TENCON.2016.7848296.","DOI":"10.1109\/TENCON.2016.7848296"},{"issue":"1","key":"14656_CR193","doi-asserted-by":"publisher","first-page":"93","DOI":"10.1007\/s10772-018-9491-z","volume":"21","author":"M Swain","year":"2018","unstructured":"Swain M, Routray A, Kabisatpathy P (2018) Databases, features and classifiers for speech emotion recognition: a review. Int J Speech Technol 21(1):93\u2013120. https:\/\/doi.org\/10.1007\/s10772-018-9491-z","journal-title":"Int J Speech Technol"},{"key":"14656_CR194","doi-asserted-by":"publisher","unstructured":"Tacconi D et al. (2008) \u201cActivity and emotion recognition to support early diagnosis of psychiatric diseases,\u201d in 2008 Second International Conference on Pervasive Computing Technologies for Healthcare, pp. 100\u2013102, https:\/\/doi.org\/10.1109\/PCTHEALTH.2008.4571041.","DOI":"10.1109\/PCTHEALTH.2008.4571041"},{"issue":"17","key":"14656_CR195","doi-asserted-by":"publisher","first-page":"1","DOI":"10.5120\/ijca2018916290","volume":"179","author":"M Taha","year":"2018","unstructured":"Taha M, Adeel A, Hussain A (2018) A survey on techniques for enhancing speech. Int J Comput Appl 179(17):1\u201314. https:\/\/doi.org\/10.5120\/ijca2018916290","journal-title":"Int J Comput Appl"},{"issue":"8","key":"14656_CR196","doi-asserted-by":"publisher","first-page":"1301","DOI":"10.1109\/JSTSP.2017.2764438","volume":"11","author":"G Trigeorgis","year":"2017","unstructured":"Trigeorgis G, Nicolaou MA, Schuller W (2017) End-to-end multimodal emotion recognition using deep neural networks. IEEE J Select Top Signal Process 11(8):1301\u20131309","journal-title":"IEEE J Select Top Signal Process"},{"key":"14656_CR197","doi-asserted-by":"publisher","unstructured":"Uhrin D, Partila P, Voznak M, Chmelikova Z, Hlozak M, Orcik L (2014) \u201cDesign and implementation of Czech database of speech emotions,\u201d 2014 22nd Telecommunications Forum, TELFOR 2014 - Proceedings of Papers, no. November, pp. 529\u2013532, https:\/\/doi.org\/10.1109\/TELFOR.2014.7034463.","DOI":"10.1109\/TELFOR.2014.7034463"},{"key":"14656_CR198","doi-asserted-by":"publisher","first-page":"189","DOI":"10.1016\/j.jbi.2018.07.014","volume":"85","author":"RJ Urbanowicz","year":"2018","unstructured":"Urbanowicz RJ, Meeker M, La Cava W, Olson RS, Moore JH (2018) Relief-based feature selection: Introduction and review. J Biomed Inf 85:189\u2013203. https:\/\/doi.org\/10.1016\/j.jbi.2018.07.014","journal-title":"J Biomed Inf"},{"key":"14656_CR199","doi-asserted-by":"publisher","unstructured":"Valstar M et al. (2014) \u201cAVEC 2014 - 3D dimensional affect and depression recognition challenge,\u201d AVEC 2014 - Proceedings of the 4th International Workshop on Audio\/Visual Emotion Challenge, Workshop of MM 2014, no. January 2021, pp. 3\u201310, https:\/\/doi.org\/10.1145\/2661806.2661807.","DOI":"10.1145\/2661806.2661807"},{"key":"14656_CR200","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1080\/13506280444000102","volume":"10","author":"LJP Van Der Maaten","year":"2009","unstructured":"Van Der Maaten LJP, Postma EO, Van Den Herik HJ (2009) Dimensionality reduction: a comparative review. J Mach Learn Res 10:1\u201341. https:\/\/doi.org\/10.1080\/13506280444000102","journal-title":"J Mach Learn Res"},{"issue":"4","key":"14656_CR201","first-page":"309","volume":"50","author":"K Van Lierde","year":"1996","unstructured":"Van Lierde K, Moerman M, Vermeersch H, Van Cauwenberge P (1996) An introduction to computerised speech lab. Acta Otorhinolaryngol Belg 50(4):309\u2013314","journal-title":"Acta Otorhinolaryngol Belg"},{"issue":"9","key":"14656_CR202","doi-asserted-by":"publisher","first-page":"1162","DOI":"10.1016\/j.specom.2006.04.003","volume":"48","author":"D Ververidis","year":"2006","unstructured":"Ververidis D, Kotropoulos C (2006) Emotional speech recognition: resources, features, and methods. Speech Comm 48(9):1162\u20131181. https:\/\/doi.org\/10.1016\/j.specom.2006.04.003","journal-title":"Speech Comm"},{"key":"14656_CR203","doi-asserted-by":"publisher","unstructured":"Ververidis D, Kotropoulos C (2006) \u201cEmotional speech recognition: resources, features, and methods,\u201d Speech Commun, https:\/\/doi.org\/10.1016\/j.specom.2006.04.003.","DOI":"10.1016\/j.specom.2006.04.003"},{"key":"14656_CR204","doi-asserted-by":"publisher","first-page":"666","DOI":"10.1016\/j.procs.2016.06.032","volume":"89","author":"S Vihari","year":"2016","unstructured":"Vihari S, Murthy AS, Soni P, Naik DC (2016) Comparison of speech enhancement algorithms. Proced Comput Sci 89:666\u2013676. https:\/\/doi.org\/10.1016\/j.procs.2016.06.032","journal-title":"Proced Comput Sci"},{"key":"14656_CR205","first-page":"1","volume":"1","author":"B Vlasenko","year":"2007","unstructured":"Vlasenko B, Wendemuth A (2007) Tuning hidden Markov model for speech emotion recognition. DAGA 1:1","journal-title":"DAGA"},{"key":"14656_CR206","doi-asserted-by":"publisher","unstructured":"Vrebcevic N, Mijic I, Petrinovic D (2019) \u201cEmotion classification based on convolutional neural network using speech data,\u201d 2019 42nd International Convention on Information and Communication Technology, Electronics and Microelectronics, MIPRO 2019 - Proceedings, pp. 1007\u20131012, https:\/\/doi.org\/10.23919\/MIPRO.2019.8756867.","DOI":"10.23919\/MIPRO.2019.8756867"},{"issue":"10","key":"14656_CR207","doi-asserted-by":"publisher","first-page":"2429","DOI":"10.1016\/S0031-3203(03)00044-X","volume":"36","author":"X Wang","year":"2003","unstructured":"Wang X, Paliwal KK (2003) Feature extraction and dimensionality reduction algorithms and their applications in vowel recognition. Pattern Recognit 36(10):2429\u20132439. https:\/\/doi.org\/10.1016\/S0031-3203(03)00044-X","journal-title":"Pattern Recognit"},{"key":"14656_CR208","unstructured":"Wang K, An N, Li BN, Zhang Y, Li L (2015) \u201cSpeech emotion recognition using Fourier parameters,\u201d IEEE Trans Affect Comput"},{"key":"14656_CR209","doi-asserted-by":"publisher","unstructured":"Wei H, Shi X, Yang J, Pu Y (2010) \u201cSpeech Independent Component Analysis,\u201d in 2010 International Conference on Measuring Technology and Mechatronics Automation, vol. 3, pp. 445\u2013448, https:\/\/doi.org\/10.1109\/ICMTMA.2010.604.","DOI":"10.1109\/ICMTMA.2010.604"},{"issue":"9","key":"14656_CR210","doi-asserted-by":"publisher","first-page":"1100","DOI":"10.1109\/T-C.1971.223410","volume":"C\u201320","author":"AW Whitney","year":"1971","unstructured":"Whitney AW (1971) A Direct Method of Nonparametric Measurement Selection. IEEE Trans Comput C\u201320(9):1100\u20131103. https:\/\/doi.org\/10.1109\/T-C.1971.223410","journal-title":"IEEE Trans Comput"},{"key":"14656_CR211","unstructured":"Williams CE, Stevens KN (1981) \u201cVocal correlates of emotional states,\u201d in Speech Eval Psychiatry"},{"key":"14656_CR212","unstructured":"Wu J (2017) Introduction to convolutional neural networks. Introduct Convolutional Neural Netw, pp 1\u201331. https:\/\/cs.nju.edu.cn\/wujx\/paper\/CNN.pdf. Accessed 14 Nov\u00a02021"},{"key":"14656_CR213","doi-asserted-by":"publisher","first-page":"101","DOI":"10.1016\/j.apnum.2020.09.013","volume":"164","author":"G Wu","year":"2021","unstructured":"Wu G, Li F (2021) A randomized exponential canonical correlation analysis method for data analysis and dimensionality reduction. Appl Numer Math 164:101\u2013124. https:\/\/doi.org\/10.1016\/j.apnum.2020.09.013","journal-title":"Appl Numer Math"},{"issue":"April 2019","key":"14656_CR214","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.specom.2020.03.005","volume":"120","author":"Z Yao","year":"2020","unstructured":"Yao Z, Wang Z, Liu W, Liu Y, Pan J (2020) Speech emotion recognition using fusion of three multi-task learning-based classifiers: HSF-DNN, MS-CNN and LLD-RNN. Speech Commun 120(April 2019):11\u201319. https:\/\/doi.org\/10.1016\/j.specom.2020.03.005","journal-title":"Speech Commun"},{"key":"14656_CR215","doi-asserted-by":"crossref","unstructured":"Ye C, Liu J, Chen C, Song M, Bu J (2008) \u201cSpeech Emotion Classification on a Riemannian Manifold,\u201d in Advances in Multimedia Information Processing - PCM 2008, pp. 61\u201369","DOI":"10.1007\/978-3-540-89796-5_7"},{"key":"14656_CR216","unstructured":"Yegnanarayana B (2009) Artificial neural networks. PHI Learning Pvt. Ltd"},{"issue":"1","key":"14656_CR217","doi-asserted-by":"publisher","first-page":"29","DOI":"10.1016\/j.csl.2009.12.004","volume":"25","author":"S Yildirim","year":"2011","unstructured":"Yildirim S, Narayanan S, Potamianos A (2011) Detecting emotional state of a child in a conversational computer game. Comput Speech Lang 25(1):29\u201344. https:\/\/doi.org\/10.1016\/j.csl.2009.12.004","journal-title":"Comput Speech Lang"},{"key":"14656_CR218","doi-asserted-by":"publisher","unstructured":"Yu C, Aoki P, Woodruff A (2004) Detecting user engagement in everyday conversations. ArXiv. https:\/\/doi.org\/10.48550\/arXiv.cs\/0410027","DOI":"10.48550\/arXiv.cs\/0410027"},{"key":"14656_CR219","unstructured":"Zang Q, Wang SLK (2013) A database of elderly emotional speech. In:\u00a02013 International symposium on signal processing, biomedical engineering and informatics"},{"key":"14656_CR220","doi-asserted-by":"crossref","unstructured":"Zeynep Inanoglu RC (2005) Emotive alert: hmm-based emotion detection in voicemail messages. https:\/\/vismod.media.mit.edu\/tech-reports\/TR-585.pdf","DOI":"10.1145\/1040830.1040885"},{"key":"14656_CR221","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TAFFC.2016.2553038","volume":"PP","author":"S Zhalehpour","year":"2016","unstructured":"Zhalehpour S, Onder O, Akhtar Z, Erdem C (2016) BAUM-1: A Spontaneous Audio-Visual Face Database of Affective and Mental States. IEEE Trans Affect Comput PP:1. https:\/\/doi.org\/10.1109\/TAFFC.2016.2553038","journal-title":"IEEE Trans Affect Comput"},{"issue":"1","key":"14656_CR222","doi-asserted-by":"publisher","first-page":"115","DOI":"10.1109\/TASLP.2014.2375558","volume":"23","author":"Z Zhang","year":"2015","unstructured":"Zhang Z, Coutinho E, Deng J, Schuller B (2015) Cooperative learning and its application to emotion recognition from speech. IEEE\/ACM Trans Audio Speech Lang Process 23(1):115\u2013126. https:\/\/doi.org\/10.1109\/TASLP.2014.2375558","journal-title":"IEEE\/ACM Trans Audio Speech Lang Process"},{"issue":"6","key":"14656_CR223","doi-asserted-by":"publisher","first-page":"1576","DOI":"10.1109\/TMM.2017.2766843","volume":"20","author":"S Zhang","year":"2018","unstructured":"Zhang S, Zhang S, Huang T, Gao W (2018) Speech emotion recognition using deep convolutional neural network and discriminant temporal pyramid matching. IEEE Trans Multimed 20(6):1576\u20131590. https:\/\/doi.org\/10.1109\/TMM.2017.2766843","journal-title":"IEEE Trans Multimed"},{"key":"14656_CR224","doi-asserted-by":"publisher","unstructured":"Zhang X, Wu G, Ren F (2018) \u201cSearching Audio-Visual Clips for Dual-mode Chinese Emotional Speech Database,\u201d 2018 1st Asian Conference on Affective Computing and Intelligent Interaction, ACII Asia 2018, pp. 1\u20136, https:\/\/doi.org\/10.1109\/ACIIAsia.2018.8470387.","DOI":"10.1109\/ACIIAsia.2018.8470387"},{"key":"14656_CR225","doi-asserted-by":"publisher","first-page":"23496","DOI":"10.1109\/ACCESS.2020.2969032","volume":"8","author":"S Zhang","year":"2020","unstructured":"Zhang S, Chen A, Guo W, Cui Y, Zhao X, Liu L (2020) Learning deep binaural representations with deep convolutional neural networks for spontaneous speech emotion recognition. IEEE Access 8:23496\u201323505. https:\/\/doi.org\/10.1109\/ACCESS.2020.2969032","journal-title":"IEEE Access"},{"key":"14656_CR226","doi-asserted-by":"publisher","first-page":"5332","DOI":"10.1109\/ACCESS.2020.3047395","volume":"9","author":"H Zhang","year":"2021","unstructured":"Zhang H, Huang H, Han H (2021) Attention-based convolution skip bidirectional long short-term memory network for speech emotion recognition. IEEE Access 9:5332\u20135342. https:\/\/doi.org\/10.1109\/ACCESS.2020.3047395","journal-title":"IEEE Access"},{"key":"14656_CR227","doi-asserted-by":"crossref","unstructured":"Zhang C, Liu Y, Fu H (n.d.) \u201cAE 2 -Nets\u00a0: Autoencoder in Autoencoder Networks,\u201d in 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2577\u20132585","DOI":"10.1109\/CVPR.2019.00268"},{"issue":"6","key":"14656_CR228","doi-asserted-by":"publisher","first-page":"713","DOI":"10.1049\/iet-spr.2017.0320","volume":"12","author":"J Zhao","year":"2018","unstructured":"Zhao J, Mao X, Chen L (2018) Learning deep features to recognise speech emotion using merged deep CNN. IET Signal Process 12(6):713\u2013721. https:\/\/doi.org\/10.1049\/iet-spr.2017.0320","journal-title":"IET Signal Process"},{"key":"14656_CR229","doi-asserted-by":"publisher","first-page":"97515","DOI":"10.1109\/ACCESS.2019.2928625","volume":"7","author":"Z Zhao","year":"2019","unstructured":"Zhao Z, Bao Z, Zhao Y, Zhang Z, Cummins N, Ren Z, Schuller B (2019) Exploring deep Spectrum representations via attention-based recurrent and convolutional neural networks for speech emotion recognition. IEEE Access 7:97515\u201397525. https:\/\/doi.org\/10.1109\/ACCESS.2019.2928625","journal-title":"IEEE Access"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-14656-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-023-14656-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-14656-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,7,22]],"date-time":"2023-07-22T10:28:11Z","timestamp":1690021691000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-023-14656-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,2,22]]},"references-count":229,"journal-issue":{"issue":"19","published-print":{"date-parts":[[2023,8]]}},"alternative-id":["14656"],"URL":"https:\/\/doi.org\/10.1007\/s11042-023-14656-y","relation":{},"ISSN":["1380-7501","1573-7721"],"issn-type":[{"value":"1380-7501","type":"print"},{"value":"1573-7721","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,2,22]]},"assertion":[{"value":"15 December 2021","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 April 2022","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 February 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 February 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of Interests"}}]}}