{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T23:09:05Z","timestamp":1772147345531,"version":"3.50.1"},"reference-count":60,"publisher":"Springer Science and Business Media LLC","issue":"12","license":[{"start":{"date-parts":[[2023,7,29]],"date-time":"2023-07-29T00:00:00Z","timestamp":1690588800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,7,29]],"date-time":"2023-07-29T00:00:00Z","timestamp":1690588800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Circuits Syst Signal Process"],"published-print":{"date-parts":[[2023,12]]},"DOI":"10.1007\/s00034-023-02446-8","type":"journal-article","created":{"date-parts":[[2023,7,29]],"date-time":"2023-07-29T06:02:12Z","timestamp":1690610532000},"page":"7357-7376","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["Improved Speech Emotion Classification Using Deep Neural Network"],"prefix":"10.1007","volume":"42","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4962-4239","authenticated-orcid":false,"given":"Mariwan","family":"Hama Saeed","sequence":"first","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,7,29]]},"reference":[{"key":"2446_CR1","doi-asserted-by":"publisher","first-page":"150","DOI":"10.1016\/j.ins.2019.09.005","volume":"509","author":"L Chen","year":"2020","unstructured":"L. Chen et al., Two-layer fuzzy multiple random forest for speech emotion recognition in human-robot interaction. Inf. Sci. 509, 150\u2013163 (2020)","journal-title":"Inf. Sci."},{"key":"2446_CR2","doi-asserted-by":"publisher","first-page":"128","DOI":"10.22452\/mjcs.vol35no2.3","volume":"35","author":"H Ibrahim","year":"2022","unstructured":"H. Ibrahim, C.K. Loo, Reservoir computing with truncated normal distribution for speech emotion recognition. Malays. J. Comput. Sci. 35, 128\u2013141 (2022)","journal-title":"Malays. J. Comput. Sci."},{"key":"2446_CR3","unstructured":"N. Truong Pham, Hybrid data augmentation and deep attention-based dilated convolutional-recurrent neural networks for speech emotion recognition. ArXiv arXiv:2109.09026 (2021)."},{"key":"2446_CR4","doi-asserted-by":"publisher","first-page":"10773","DOI":"10.1007\/s11227-021-03690-y","volume":"77","author":"L-N Do","year":"2021","unstructured":"L.-N. Do et al., Deep neural network-based fusion model for emotion recognition using visual data. J. Supercomput. 77, 10773\u201310790 (2021)","journal-title":"J. Supercomput."},{"key":"2446_CR5","unstructured":"L. Rowlands, Emotions: how humans regulate them and why some people can\u2019t | News and Events | Bangor University. https:\/\/www.bangor.ac.uk\/news\/archive\/emotions-how-humans-regulate-them-and-why-some-people-can-t-38444"},{"key":"2446_CR6","doi-asserted-by":"publisher","unstructured":"B. McFee et al. Librosa: audio and music signal analysis in python. In Proceedings of the 14th Python in Science Conference 18\u201324 (SciPy, 2015). doi:https:\/\/doi.org\/10.25080\/majora-7b98e3ed-003.","DOI":"10.25080\/majora-7b98e3ed-003"},{"key":"2446_CR7","doi-asserted-by":"publisher","first-page":"16","DOI":"10.1186\/1687-4722-2012-16","volume":"2012","author":"MC Sezgin","year":"2012","unstructured":"M.C. Sezgin, B. Gunsel, G.K. Kurt, Perceptual audio features for emotion detection. EURASIP J. Audio Speech Music Process. 2012, 16 (2012)","journal-title":"EURASIP J. Audio Speech Music Process."},{"key":"2446_CR8","volume-title":"Machine learning with Python Cookbook: Practical Solutions from Preprocessing to Deep Learning","author":"C Albon","year":"2018","unstructured":"C. Albon, Machine learning with Python Cookbook: Practical Solutions from Preprocessing to Deep Learning (O\u2019Reilly Media, Sebastopol, 2018)"},{"key":"2446_CR9","doi-asserted-by":"publisher","first-page":"125830","DOI":"10.1109\/ACCESS.2021.3111659","volume":"9","author":"S Kanwal","year":"2021","unstructured":"S. Kanwal, S. Asghar, Speech emotion recognition using clustering based GA-optimized feature set. IEEE Access 9, 125830\u2013125842 (2021)","journal-title":"IEEE Access"},{"key":"2446_CR10","doi-asserted-by":"publisher","first-page":"1176","DOI":"10.1109\/ICAIS50930.2021.9395844","volume":"2021","author":"K Chauhan","year":"2021","unstructured":"K. Chauhan, K.K. Sharma, T. Varma, Speech emotion recognition using convolution neural networks. Proc. Int. Conf. Artif. Intell. Smart Syst. 2021, 1176\u20131181 (2021). https:\/\/doi.org\/10.1109\/ICAIS50930.2021.9395844","journal-title":"Proc. Int. Conf. Artif. Intell. Smart Syst."},{"key":"2446_CR11","doi-asserted-by":"publisher","first-page":"122855","DOI":"10.1109\/ACCESS.2021.3107858","volume":"9","author":"H Ibrahim","year":"2021","unstructured":"H. Ibrahim, C.K. Loo, F. Alnajjar, Speech emotion recognition by late fusion for bidirectional reservoir computing with random projection. IEEE Access 9, 122855\u2013122871 (2021)","journal-title":"IEEE Access"},{"key":"2446_CR12","doi-asserted-by":"publisher","first-page":"95925","DOI":"10.1109\/ACCESS.2021.3094355","volume":"9","author":"N Liu","year":"2021","unstructured":"N. Liu et al., Transfer subspace learning for unsupervised cross-corpus speech emotion recognition. IEEE Access 9, 95925\u201395937 (2021)","journal-title":"IEEE Access"},{"key":"2446_CR13","doi-asserted-by":"publisher","first-page":"79861","DOI":"10.1109\/ACCESS.2020.2990405","volume":"8","author":"SM Mustaqeem","year":"2020","unstructured":"S.M. Mustaqeem, S. Kwon, Clustering-Based Speech Emotion Recognition by Incorporating Learned Features and Deep BiLSTM. IEEE Access 8, 79861\u201379875 (2020)","journal-title":"IEEE Access"},{"key":"2446_CR14","doi-asserted-by":"publisher","DOI":"10.48550\/arxiv.2109.09026","author":"NT Pham","year":"2023","unstructured":"N.T. Pham, D.N.M. Dang, N.D. Nguyen, T.T. Nguyen, H. Nguyen, B. Manavalan, C.P. Lim, S.D. Nguyen, Hybrid data augmentation and deep attention-based dilated convolutional-recurrent neural networks for speech emotion recognition. Expert Syst. Appl. (2023). https:\/\/doi.org\/10.48550\/arxiv.2109.09026","journal-title":"Expert Syst. Appl."},{"key":"2446_CR15","doi-asserted-by":"publisher","unstructured":"M. Seknedy El, S. Fawzi, Speech emotion recognition system for human interaction applications. In: Proceedings\u20142021 IEEE 10th International Conference on Intelligent Computing and Information Systems, ICICIS 2021 361\u2013368 (2021) doi:https:\/\/doi.org\/10.1109\/ICICIS52592.2021.9694246.","DOI":"10.1109\/ICICIS52592.2021.9694246"},{"key":"2446_CR16","doi-asserted-by":"publisher","first-page":"9897","DOI":"10.3390\/app11219897","volume":"11","author":"H Zhang","year":"2021","unstructured":"H. Zhang, H. Huang, H. Han, A novel heterogeneous parallel convolution bi-LSTM for speech emotion recognition. Appl. Sci. 11, 9897 (2021)","journal-title":"Appl. Sci."},{"key":"2446_CR17","doi-asserted-by":"publisher","first-page":"251","DOI":"10.1016\/j.procs.2020.08.027","volume":"176","author":"H Aouani","year":"2020","unstructured":"H. Aouani, Y.B. Ayed, Speech emotion recognition with deep learning. Procedia Comput. Sci. 176, 251\u2013260 (2020)","journal-title":"Procedia Comput. Sci."},{"key":"2446_CR18","doi-asserted-by":"publisher","first-page":"2891","DOI":"10.3390\/electronics10232891","volume":"10","author":"S Huang","year":"2021","unstructured":"S. Huang et al., Multi-layer hybrid fuzzy classification based on SVM and improved PSO for speech emotion recognition. Electronics 10, 2891 (2021)","journal-title":"Electronics"},{"key":"2446_CR19","doi-asserted-by":"publisher","first-page":"309","DOI":"10.1016\/j.ins.2021.02.016","volume":"563","author":"ZT Liu","year":"2021","unstructured":"Z.T. Liu, A. Rehman, M. Wu, W.H. Cao, M. Hao, Speech emotion recognition based on formant characteristics feature extraction and phoneme type convergence. Inf. Sci. 563, 309\u2013325 (2021)","journal-title":"Inf. Sci."},{"key":"2446_CR20","doi-asserted-by":"publisher","first-page":"1845","DOI":"10.1007\/s40747-020-00250-4","volume":"7","author":"W Zehra","year":"2021","unstructured":"W. Zehra, A.R. Javed, Z. Jalil, H.U. Khan, T.R. Gadekallu, Cross corpus multi-lingual speech emotion recognition using ensemble learning. Complex Intell. Syst. 7, 1845\u20131854 (2021)","journal-title":"Complex Intell. Syst."},{"key":"2446_CR21","doi-asserted-by":"publisher","first-page":"15563","DOI":"10.1007\/s11042-020-10329-2","volume":"80","author":"MD Pawar","year":"2021","unstructured":"M.D. Pawar, R.D. Kokate, Convolution neural network based automatic speech emotion recognition using Mel-frequency Cepstrum coefficients. Multimed. Tools Appl. 80, 15563\u201315587 (2021)","journal-title":"Multimed. Tools Appl."},{"key":"2446_CR22","doi-asserted-by":"publisher","unstructured":"M.H. Pham, F.M. Noori, J.Torresen, Emotion recognition using speech data with convolutional neural network. In: 2021 IEEE 2nd International Conference on Signal, Control and Communication, SCC 2021 182\u2013187 (2021) doi:https:\/\/doi.org\/10.1109\/SCC53769.2021.9768372.","DOI":"10.1109\/SCC53769.2021.9768372"},{"key":"2446_CR23","doi-asserted-by":"publisher","first-page":"166518","DOI":"10.1109\/ACCESS.2021.3135658","volume":"9","author":"KK Sahoo","year":"2021","unstructured":"K.K. Sahoo, I. Dutta, M.F. Ijaz, M. Wozniak, P.K. Singh, TLEFuzzyNet: fuzzy rank-based ensemble of transfer learning models for emotion recognition from human speeches. IEEE Access 9, 166518\u2013166530 (2021)","journal-title":"IEEE Access"},{"key":"2446_CR24","first-page":"201","volume-title":"Nature Inspired Computing for Data Science. In Studies in Computational Intelligence","author":"AK Sahoo","year":"2020","unstructured":"A.K. Sahoo, C. Pradhan, H. Das, Performance evaluation of different machine learning methods and deep-learning based convolutional neural network for health decision making, in Nature Inspired Computing for Data Science. In Studies in Computational Intelligence. (Springer, Cham, 2020), pp.201\u2013212"},{"key":"2446_CR25","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.specom.2022.03.002","volume":"140","author":"BT Atmaja","year":"2022","unstructured":"B.T. Atmaja, A. Sasou, M. Akagi, Survey on bimodal speech emotion recognition from acoustic and linguistic information fusion. Speech Commun. 140, 11\u201328 (2022)","journal-title":"Speech Commun."},{"key":"2446_CR26","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4614-5143-3","volume-title":"Emotion Recognition using Speech Features","author":"KS Rao","year":"2013","unstructured":"K.S. Rao, S.G. Koolagudi, Emotion Recognition using Speech Features (Springer, Cham, 2013). https:\/\/doi.org\/10.1007\/978-1-4614-5143-3"},{"key":"2446_CR27","doi-asserted-by":"publisher","first-page":"93","DOI":"10.1007\/s10772-018-9491-z","volume":"21","author":"M Swain","year":"2018","unstructured":"M. Swain, A. Routray, P. Kabisatpathy, Databases, features and classifiers for speech emotion recognition: a review. Int. J. Speech Technol. 21, 93\u2013120 (2018)","journal-title":"Int. J. Speech Technol."},{"key":"2446_CR28","doi-asserted-by":"publisher","unstructured":"G. Degottex, J. Kane, T. Drugman, T. Raitio, S. Scherer, COVAREP\u2014A collaborative voice analysis repository for speech technologies. In: ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings 960\u2013964 (2014) doi:https:\/\/doi.org\/10.1109\/ICASSP.2014.6853739.","DOI":"10.1109\/ICASSP.2014.6853739"},{"key":"2446_CR29","doi-asserted-by":"crossref","unstructured":"Eyben F., B. Schuller. openSMILE:. ACM SIGMultimedia Records 6, (2015).","DOI":"10.1145\/2729095.2729097"},{"key":"2446_CR30","doi-asserted-by":"publisher","unstructured":"Kejriwal J., Benus, S. & Trnka, M. Stress detection using non-semantic speech representation. 2022 32nd International Conference Radioelektronika, RADIOELEKTRONIKA 2022 - Proceedings (2022) doi:https:\/\/doi.org\/10.1109\/RADIOELEKTRONIKA54537.2022.9764916.","DOI":"10.1109\/RADIOELEKTRONIKA54537.2022.9764916"},{"key":"2446_CR31","doi-asserted-by":"publisher","first-page":"1424","DOI":"10.1109\/TASLP.2021.3068598","volume":"29","author":"S Mai","year":"2021","unstructured":"S. Mai, S. Xing, H. Hu, Analyzing multimodal sentiment via acoustic- and visual-LSTM with channel-aware temporal convolution network. IEEE\/ACM Trans. Audio Speech Lang. Process. 29, 1424\u20131437 (2021)","journal-title":"IEEE\/ACM Trans. Audio Speech Lang. Process."},{"key":"2446_CR32","doi-asserted-by":"publisher","first-page":"73","DOI":"10.1109\/MSP.2021.3052181","volume":"38","author":"M Mueller","year":"2021","unstructured":"M. Mueller, B. McFee, Interactive learning of signal processing through music: making fourier analysis concrete for students. IEEE Signal Process Mag. 38, 73\u201384 (2021)","journal-title":"IEEE Signal Process Mag."},{"key":"2446_CR33","doi-asserted-by":"publisher","unstructured":"M. Muthumari, V. Akash, K. Prudhvicharan, P. Akhil, A novel model for emotion detection with multilayer perceptron neural network. Proceedings\u20142022 6th International Conference on Intelligent Computing and Control Systems, ICICCS 2022 1126\u20131131 (2022) doi:https:\/\/doi.org\/10.1109\/ICICCS53718.2022.9788269.","DOI":"10.1109\/ICICCS53718.2022.9788269"},{"key":"2446_CR34","doi-asserted-by":"crossref","unstructured":"V.S. Nallanthighal, A. H\u00e4rm\u00e4, H. Strik, Detection of COPD exacerbation from speech: comparison of acoustic features and deep learning based speech breathing models. In: ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings 2022-May, 2505\u20132509 (2022).","DOI":"10.1109\/ICASSP43922.2022.9747785"},{"key":"2446_CR35","doi-asserted-by":"publisher","first-page":"409","DOI":"10.1007\/978-981-19-0182-9_41","volume-title":"Visualization of Audio Files Using Librosa","author":"S Suman","year":"2022","unstructured":"S. Suman, K.S. Sahoo, C. Das, N.Z. Jhanjhi, A. Mitra, Visualization of Audio Files Using Librosa (Springer Nature Singapore, Singapore, 2022), pp.409\u2013418. https:\/\/doi.org\/10.1007\/978-981-19-0182-9_41"},{"key":"2446_CR36","doi-asserted-by":"publisher","first-page":"e0247655","DOI":"10.1371\/journal.pone.0247655","volume":"16","author":"M Tomprou","year":"2021","unstructured":"M. Tomprou, Y.J. Kim, P. Chikersal, A.W. Woolley, L.A. Dabbish, Speaking out of turn: How video conferencing reduces vocal synchrony and collective intelligence. PLoS ONE 16, e0247655 (2021)","journal-title":"PLoS ONE"},{"key":"2446_CR37","doi-asserted-by":"publisher","first-page":"6558","DOI":"10.18653\/v1\/P19-1656","volume":"2019","author":"YHH Tsai","year":"2019","unstructured":"Y.H.H. Tsai et al., Multimodal transformer for unaligned multimodal language sequences. Proc. Conf. Assoc. Comput. Linguist. Meet. 2019, 6558 (2019)","journal-title":"Proc. Conf. Assoc. Comput. Linguist. Meet."},{"key":"2446_CR38","doi-asserted-by":"publisher","first-page":"1025","DOI":"10.3390\/e24081025","volume":"24","author":"J Krzywanski","year":"2022","unstructured":"J. Krzywanski et al., Multi-stream convolution-recurrent neural networks based on attention mechanism fusion for speech emotion recognition. Entropy 24, 1025 (2022)","journal-title":"Entropy"},{"key":"2446_CR39","doi-asserted-by":"publisher","first-page":"51231","DOI":"10.1109\/ACCESS.2021.3069818","volume":"9","author":"C Zhang","year":"2021","unstructured":"C. Zhang, L. Xue, Autoencoder with emotion embedding for speech emotion recognition. IEEE Access 9, 51231\u201351241 (2021)","journal-title":"IEEE Access"},{"key":"2446_CR40","doi-asserted-by":"publisher","first-page":"5116","DOI":"10.1002\/int.22505","volume":"36","author":"SM Mustaqeem","year":"2021","unstructured":"S.M. Mustaqeem, S. Kwon, Optimal feature selection based speech emotion recognition using two-stream deep convolutional neural network. Int. J. Intell. Syst. 36, 5116\u20135135 (2021)","journal-title":"Int. J. Intell. Syst."},{"key":"2446_CR41","doi-asserted-by":"publisher","DOI":"10.1016\/j.matpr.2021.12.246","author":"N Senthilkumar","year":"2022","unstructured":"N. Senthilkumar, S. Karpakam, M. Gayathri Devi, R. Balakumaresan, P. Dhilipkumar, Speech emotion recognition based on Bi-directional LSTM architecture and deep belief networks. Mater. Today Proc. (2022). https:\/\/doi.org\/10.1016\/j.matpr.2021.12.246","journal-title":"Mater. Today Proc."},{"key":"2446_CR42","doi-asserted-by":"publisher","first-page":"1328","DOI":"10.3390\/electronics11091328","volume":"11","author":"B Maji","year":"2022","unstructured":"B. Maji, M. Swain, M. Mustaqeem, Advanced fusion-based speech emotion recognition system using a dual-attention mechanism with conv-caps and bi-GRU features. Electronics 11, 1328 (2022)","journal-title":"Electronics"},{"key":"2446_CR43","doi-asserted-by":"publisher","first-page":"119633","DOI":"10.1016\/j.eswa.2023.119633","volume":"218","author":"M Rayhan Ahmed","year":"2023","unstructured":"M. Rayhan Ahmed, S. Islam, A.K.M. Muzahidul Islam, S. Shatabda, An ensemble 1D-CNN-LSTM-GRU model with data augmentation for speech emotion recognition. Expert Syst. Appl. 218, 119633 (2023)","journal-title":"Expert Syst. Appl."},{"key":"2446_CR44","doi-asserted-by":"publisher","first-page":"185","DOI":"10.1109\/CENTCON56610.2022.10051557","volume":"2","author":"YL Prasanna","year":"2023","unstructured":"Y.L. Prasanna, Y. Tarakaram, Y. Mounika, S. Palaniswamy, S. Vekkot, Comparative deep network analysis of speech emotion recognition models using data augmentation. Int. Conf. Disruptive Technol. Multi-Discipl. Res. Appl. 2, 185\u2013190 (2023). https:\/\/doi.org\/10.1109\/CENTCON56610.2022.10051557","journal-title":"Int. Conf. Disruptive Technol. Multi-Discipl. Res. Appl."},{"key":"2446_CR45","volume-title":"Surrey Audio-Visual Expressed Emotion (Savee) Database","author":"P Jackson","year":"2014","unstructured":"P. Jackson, S. Haq, Surrey Audio-Visual Expressed Emotion (Savee) Database (University of Surrey, Guildford, 2014)"},{"key":"2446_CR46","doi-asserted-by":"crossref","unstructured":"F. Burkhardt, A. Paeschke, M. Rolfes, W.F. Sendlmeier, B. Weiss, A database of German emotional speech. In INTERSPEECH (2005).","DOI":"10.21437\/Interspeech.2005-446"},{"key":"2446_CR47","doi-asserted-by":"publisher","unstructured":"Gournay P, O. Lahaie, R. Lefebvre, A Canadian French emotional speech dataset. Proc. 9th ACM Multimed. Syst. Conf. (2018). https:\/\/doi.org\/10.5281\/ZENODO.1478765","DOI":"10.5281\/ZENODO.1478765"},{"key":"2446_CR48","doi-asserted-by":"publisher","DOI":"10.48550\/arxiv.2003.07996","author":"S Goel","year":"2020","unstructured":"S. Goel, H. Beigi, Cross lingual cross corpus speech emotion recognition. arXiv preprint (2020). https:\/\/doi.org\/10.48550\/arxiv.2003.07996","journal-title":"arXiv preprint"},{"key":"2446_CR49","doi-asserted-by":"publisher","unstructured":"S.R. Krothapalli, Koolagudi, S. G. Emotion Recognition Using Vocal Tract Information. in 67\u201378 (2013). doi:https:\/\/doi.org\/10.1007\/978-1-4614-5143-3_4.","DOI":"10.1007\/978-1-4614-5143-3_4"},{"key":"2446_CR50","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-49220-9","volume-title":"Speech Recognition Using Articulatory and Excitation Source Features","author":"KS Rao","year":"2017","unstructured":"K.S. Rao, K.E. Manjunath, Speech Recognition Using Articulatory and Excitation Source Features (Springer International Publishing, Cham, 2017). https:\/\/doi.org\/10.1007\/978-3-319-49220-9"},{"key":"2446_CR51","doi-asserted-by":"publisher","first-page":"182868","DOI":"10.1109\/ACCESS.2020.3028121","volume":"8","author":"S Guha","year":"2020","unstructured":"S. Guha et al., Hybrid feature selection method based on harmony search and naked mole-rat algorithms for spoken language identification from audio signals. IEEE Access 8, 182868\u2013182887 (2020)","journal-title":"IEEE Access"},{"key":"2446_CR52","doi-asserted-by":"publisher","first-page":"1088","DOI":"10.1109\/JSTSP.2011.2112333","volume":"5","author":"M M\u00fcller","year":"2011","unstructured":"M. M\u00fcller, D.P.W. Ellis, A. Klapuri, G. Richard, Signal processing for music analysis. IEEE J. Sel. Top. Sign. Proces. 5, 1088\u20131110 (2011)","journal-title":"IEEE J. Sel. Top. Sign. Proces."},{"key":"2446_CR53","doi-asserted-by":"publisher","DOI":"10.1002\/9781118393550","volume-title":"An Introduction to Audio Content Analysis: Applications in Signal Processing and Music Informatics","author":"A Lerch","year":"2012","unstructured":"A. Lerch, An Introduction to Audio Content Analysis: Applications in Signal Processing and Music Informatics (Wiley, New York, 2012)"},{"key":"2446_CR54","unstructured":"J. Brownlee, Data Preparation for Machine Learning: Data Cleaning, Feature Selection, and Data Transforms in Python. (2020)."},{"key":"2446_CR55","unstructured":"J. Brownlee, Imbalanced Classification with Python\u2014Choose Better Metrics, Balance Skewed Classes, and Apply Cost-Sensitive Learning. Machine Learning Mastery (2020)."},{"key":"2446_CR56","volume-title":"Hands-on Machine Learning with Scikit-Learn, Keras and TensorFlow: Concepts, Tools, and Techniques to Build Intelligent Systems","author":"A G\u00e9ron","year":"2019","unstructured":"A. G\u00e9ron, Hands-on Machine Learning with Scikit-Learn, Keras and TensorFlow: Concepts, Tools, and Techniques to Build Intelligent Systems (O\u2019Reilly Media, Sebastopol, 2019)"},{"key":"2446_CR57","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4842-7915-1","author":"L Long","year":"2022","unstructured":"L. Long, X. Zeng, Beginning deep learning with tensorflow. Begin. Deep Learn. TensorFlow (2022). https:\/\/doi.org\/10.1007\/978-1-4842-7915-1","journal-title":"Begin. Deep Learn. TensorFlow"},{"key":"2446_CR58","doi-asserted-by":"publisher","unstructured":"F. Daneshfar, S.J. Kabudian,  Speech Emotion Recognition Using a New Hybrid Quaternion-Based Echo State Network-Bilinear Filter, Proceedings - 2021 7th International Conference on Signal Processing and Intelligent Systems, ICSPIS (2021). https:\/\/doi.org\/10.1109\/ICSPIS54653.2021.9729337","DOI":"10.1109\/ICSPIS54653.2021.9729337"},{"key":"2446_CR59","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/S41870-022-00996-9","volume":"2022","author":"A Thakur","year":"2022","unstructured":"A. Thakur, S.K. Dhull, Language-independent hyperparameter optimization based speech emotion recognition system. Int J Inform Technol 2022, 1\u20139 (2022). https:\/\/doi.org\/10.1007\/S41870-022-00996-9","journal-title":"Int J Inform Technol"},{"key":"2446_CR60","doi-asserted-by":"publisher","DOI":"10.1016\/J.APACOUST.2021.108046","volume":"179","author":"J Ancilin","year":"2021","unstructured":"J. Ancilin, A. Milton, Improved speech emotion recognition with Mel frequency magnitude coefficient. Applied Acoustics 179, 108046 (2021). https:\/\/doi.org\/10.1016\/J.APACOUST.2021.108046","journal-title":"Applied Acoustics"}],"container-title":["Circuits, Systems, and Signal Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00034-023-02446-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00034-023-02446-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00034-023-02446-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,10,17]],"date-time":"2023-10-17T15:09:35Z","timestamp":1697555375000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00034-023-02446-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,7,29]]},"references-count":60,"journal-issue":{"issue":"12","published-print":{"date-parts":[[2023,12]]}},"alternative-id":["2446"],"URL":"https:\/\/doi.org\/10.1007\/s00034-023-02446-8","relation":{},"ISSN":["0278-081X","1531-5878"],"issn-type":[{"value":"0278-081X","type":"print"},{"value":"1531-5878","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,7,29]]},"assertion":[{"value":"16 November 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"28 June 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 June 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 July 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Not applicable.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}