{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,24]],"date-time":"2026-02-24T16:41:55Z","timestamp":1771951315773,"version":"3.50.1"},"reference-count":29,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2024,5,7]],"date-time":"2024-05-07T00:00:00Z","timestamp":1715040000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,5,7]],"date-time":"2024-05-07T00:00:00Z","timestamp":1715040000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Data Sci Anal"],"published-print":{"date-parts":[[2025,4]]},"DOI":"10.1007\/s41060-024-00553-6","type":"journal-article","created":{"date-parts":[[2024,5,7]],"date-time":"2024-05-07T05:10:51Z","timestamp":1715058651000},"page":"453-467","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Multi-language: ensemble learning-based speech emotion recognition"],"prefix":"10.1007","volume":"19","author":[{"given":"Anumula","family":"Sruthi","sequence":"first","affiliation":[]},{"given":"Anumula Kalyan","family":"Kumar","sequence":"additional","affiliation":[]},{"given":"Kishore","family":"Dasari","sequence":"additional","affiliation":[]},{"given":"Yenugu","family":"Sivaramaiah","sequence":"additional","affiliation":[]},{"given":"Garikapati","family":"Divya","sequence":"additional","affiliation":[]},{"given":"Gunupudi Sai Chaitanya","family":"Kumar","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,5,7]]},"reference":[{"key":"553_CR1","volume":"167","author":"S Kwon","year":"2021","unstructured":"Kwon, S.: MLT-DNet: speech emotion recognition using 1D dilated CNN based on multi-learning trick approach. Expert Syst. Appl. 167, 114177 (2021)","journal-title":"Expert Syst. Appl."},{"key":"553_CR2","doi-asserted-by":"crossref","first-page":"73","DOI":"10.1016\/j.specom.2020.12.009","volume":"127","author":"S Zhang","year":"2021","unstructured":"Zhang, S., Tao, X., Chuang, Y., Zhao, X.: Learning deep multimodal affective features for spontaneous speech emotion recognition. Speech Commun. 127, 73\u201381 (2021)","journal-title":"Speech Commun."},{"issue":"9","key":"553_CR3","doi-asserted-by":"crossref","first-page":"5116","DOI":"10.1002\/int.22505","volume":"36","author":"S Kwon","year":"2021","unstructured":"Kwon, S.: Optimal feature selection based speech emotion recognition using two-stream deep convolutional neural network. Int. J. Intell. Syst. 36(9), 5116\u20135135 (2021)","journal-title":"Int. J. Intell. Syst."},{"key":"553_CR4","doi-asserted-by":"crossref","first-page":"100243","DOI":"10.1016\/j.dajour.2023.100243","volume":"7","author":"G Meena","year":"2023","unstructured":"Meena, G., Mohbey, K.K., Kumar, S., Lokesh, K.: A hybrid deep learning approach for detecting sentiment polarities and knowledge graph representation on monkeypox tweets. Decis. Anal. J. 7, 100243 (2023)","journal-title":"Decis. Anal. J."},{"key":"553_CR5","doi-asserted-by":"crossref","first-page":"106547","DOI":"10.1016\/j.knosys.2020.106547","volume":"211","author":"T Tuncer","year":"2021","unstructured":"Tuncer, T., Dogan, S., Acharya, U.R.: Automated accurate speech emotion recognition system using twine shuffle pattern and iterative neighborhood component analysis techniques. Knowl. Syst. 211, 106547 (2021)","journal-title":"Knowl. Syst."},{"key":"553_CR6","doi-asserted-by":"crossref","first-page":"52","DOI":"10.1016\/j.neunet.2021.03.013","volume":"141","author":"Z Zhao","year":"2021","unstructured":"Zhao, Z., Li, Q., Zhang, Z., Cummins, N., Wang, H., Tao, J., Schuller, B.W.: Combining a parallel 2D CNN with a self-attention dilated residual network for CTC-based discrete speech emotion recognition. Neural Netw. 141, 52\u201360 (2021)","journal-title":"Neural Netw."},{"key":"553_CR7","first-page":"1","volume":"14","author":"KK Mohbey","year":"2023","unstructured":"Mohbey, K.K., Meena, G., Kumar, S., Lokesh, K.: A CNN-LSTM-based hybrid deep learning approach for sentiment analysis on Monkeypox tweets. New Gener. Comput. 14, 1\u201319 (2023)","journal-title":"New Gener. Comput."},{"key":"553_CR8","doi-asserted-by":"crossref","DOI":"10.1016\/j.apacoust.2020.107721","volume":"173","author":"S Yildirim","year":"2021","unstructured":"Yildirim, S., Kaya, Y., K\u0131l\u0131\u00e7, F.: A modified feature selection method based on metaheuristic algorithms for speech emotion recognition. Appl. Acoust. 173, 107721 (2021)","journal-title":"Appl. Acoust."},{"key":"553_CR9","doi-asserted-by":"crossref","first-page":"238","DOI":"10.1016\/j.neucom.2021.02.094","volume":"448","author":"S Li","year":"2021","unstructured":"Li, S., Xing, X., Fan, W., Cai, B., Fordson, P., Xu, X.: Spatiotemporal and frequential cascaded attention networks for speech emotion recognition. Neurocomputing 448, 238\u2013248 (2021)","journal-title":"Neurocomputing"},{"key":"553_CR10","doi-asserted-by":"crossref","first-page":"309","DOI":"10.1016\/j.ins.2021.02.016","volume":"563","author":"ZT Liu","year":"2021","unstructured":"Liu, Z.T., Rehman, A., Wu, M., Cao, W.H., Hao, M.: Speech emotion recognition based on formant characteristics feature extraction and phoneme type convergence. Inf. Sci. 563, 309\u2013325 (2021)","journal-title":"Inf. Sci."},{"key":"553_CR11","doi-asserted-by":"crossref","DOI":"10.1016\/j.compeleceng.2021.107172","volume":"93","author":"HA Abdulmohsin","year":"2021","unstructured":"Abdulmohsin, H.A.: A new proposed statistical feature extraction method in speech emotion recognition. Comput. Electr. Eng. 93, 107172 (2021)","journal-title":"Comput. Electr. Eng."},{"issue":"2","key":"553_CR12","doi-asserted-by":"crossref","first-page":"186","DOI":"10.1111\/acps.13388","volume":"145","author":"L Hansen","year":"2022","unstructured":"Hansen, L., Zhang, Y.P., Wolf, D., Sechidis, K., Ladegaard, N., Fusaroli, R.: A generalizable speech emotion recognition model reveals depression and remission. Acta Psychiatr. Scand. 145(2), 186\u2013199 (2022)","journal-title":"Acta Psychiatr. Scand."},{"key":"553_CR13","doi-asserted-by":"crossref","unstructured":"Fu, C., Dissanayake, T., Hosoda, K., Maekawa, T., & Ishiguro, H.: Similarity of speech emotion in different languages revealed by a neural network with attention. In: 2020 IEEE 14th international conference on semantic computing (ICSC) (pp. 381\u2013386). IEEE (2020)","DOI":"10.1109\/ICSC.2020.00076"},{"key":"553_CR14","doi-asserted-by":"crossref","first-page":"303","DOI":"10.1007\/s10772-020-09792-x","volume":"24","author":"U Kumaran","year":"2021","unstructured":"Kumaran, U., Radha Rammohan, S., Nagarajan, S.M., Prathik, A.: Fusion of mel and gammatone frequency cepstral coefficients for speech emotion recognition using deep C-RNN. Int. J. Speech Technol. 24, 303\u2013314 (2021)","journal-title":"Int. J. Speech Technol."},{"key":"553_CR15","doi-asserted-by":"crossref","first-page":"2180","DOI":"10.1016\/j.matpr.2021.12.246","volume":"57","author":"N Senthilkumar","year":"2022","unstructured":"Senthilkumar, N., Karpakam, S., Devi, M.G., Balakumaresan, R., Dhilipkumar, P.: Speech emotion recognition based on Bi-directional LSTM architecture and deep belief networks. Mater. Today Proc. 57, 2180\u20132184 (2022)","journal-title":"Mater. Today Proc."},{"key":"553_CR16","doi-asserted-by":"crossref","unstructured":"Qadri, S. A. A., Gunawan, T. S., Kartiwi, M., Mansor, H., & Wani, T. M.: Speech emotion recognition using feature fusion of TEO and MFCC on multilingual databases. In: Recent trends in mechatronics towards industry 4.0: selected articles from iM3F 2020, Malaysia (pp. 681\u2013691). Springer Singapore (2022)","DOI":"10.1007\/978-981-33-4597-3_61"},{"issue":"24","key":"553_CR17","doi-asserted-by":"crossref","first-page":"12805","DOI":"10.3390\/app122412805","volume":"12","author":"Y Ma","year":"2022","unstructured":"Ma, Y., Wang, W.: MSFL: explainable multitask-based shared feature learning for multilingual speech emotion recognition. Appl. Sci. 12(24), 12805 (2022)","journal-title":"Appl. Sci."},{"issue":"3","key":"553_CR18","doi-asserted-by":"crossref","first-page":"1386","DOI":"10.3390\/s23031386","volume":"23","author":"W Alsabhan","year":"2023","unstructured":"Alsabhan, W.: Human-computer interaction with a real-time speech emotion recognition with ensembling techniques 1D convolution neural network and attention. Sensors 23(3), 1386 (2023)","journal-title":"Sensors"},{"issue":"1","key":"553_CR19","doi-asserted-by":"crossref","first-page":"155","DOI":"10.1007\/s10772-020-09776-x","volume":"24","author":"M Gomathy","year":"2021","unstructured":"Gomathy, M.: Optimal feature selection for speech emotion recognition using enhanced cat swarm optimization algorithm. Int. J. Speech Technol. 24(1), 155\u2013163 (2021)","journal-title":"Int. J. Speech Technol."},{"issue":"218","key":"553_CR20","doi-asserted-by":"crossref","first-page":"119633","DOI":"10.1016\/j.eswa.2023.119633","volume":"15","author":"MR Ahmed","year":"2023","unstructured":"Ahmed, M.R., Islam, S., Islam, A.M., Shatabda, S.: An ensemble 1D-CNN-LSTM-GRU model with data augmentation for speech emotion recognition. Expert Syst. Appl. 15(218), 119633 (2023)","journal-title":"Expert Syst. Appl."},{"issue":"230","key":"553_CR21","doi-asserted-by":"crossref","first-page":"120608","DOI":"10.1016\/j.eswa.2023.120608","volume":"15","author":"NT Pham","year":"2023","unstructured":"Pham, N.T., Dang, D.N., Nguyen, N.D., Nguyen, T.T., Nguyen, H., Manavalan, B., Lim, C.P., Nguyen, S.D.: Hybrid data augmentation and deep attention-based dilated convolutional-recurrent neural networks for speech emotion recognition. Expert Syst. Appl. 15(230), 120608 (2023)","journal-title":"Expert Syst. Appl."},{"key":"553_CR22","doi-asserted-by":"crossref","first-page":"107440","DOI":"10.1016\/j.patcog.2020.107440","volume":"107","author":"W Chen","year":"2020","unstructured":"Chen, W., Hu, H.: Generative attention adversarial classification network for unsupervised domain adaptation. Pattern Recogn. 107, 107440 (2020)","journal-title":"Pattern Recogn."},{"key":"553_CR23","doi-asserted-by":"crossref","first-page":"107132","DOI":"10.1016\/j.knosys.2021.107132","volume":"226","author":"PR Kanna","year":"2021","unstructured":"Kanna, P.R., Santhi, P.: Unified deep learning approach for efficient intrusion detection system using integrated spatial\u2013temporal features. Knowl. Syst. 226, 107132 (2021)","journal-title":"Knowl. Syst."},{"issue":"2019","key":"553_CR24","first-page":"1","volume":"2019","author":"Z Wang","year":"2019","unstructured":"Wang, Z., Zheng, L., Du, W., Cai, W., Zhou, J., Wang, J., He, G.: A novel method for intelligent fault diagnosis of bearing based on capsule neural network. Complexity 2019(2019), 1 (2019)","journal-title":"Complexity"},{"key":"553_CR25","unstructured":"SAVEE dataset: https:\/\/www.kaggle.com\/datasets\/ejlok1\/surrey-audiovisual-expressed-emotion-savee"},{"key":"553_CR26","unstructured":"EMO-DB dataset: https:\/\/www.kaggle.com\/datasets\/piyushagni5\/berlin-database-of-emotional-speech-emodb"},{"key":"553_CR27","unstructured":"URDU dataset: https:\/\/www.kaggle.com\/datasets\/hazrat\/urdu-speech-dataset?select=files"},{"issue":"18","key":"553_CR28","doi-asserted-by":"crossref","first-page":"9188","DOI":"10.3390\/app12189188","volume":"12","author":"BB Al-onazi","year":"2022","unstructured":"Al-onazi, B.B., Nauman, M.A., Jahangir, R., Malik, M.M., Alkhammash, E.H., Elshewey, A.M.: Transformer-based multilingual speech emotion recognition using data augmentation and feature fusion. Appl. Sci. 12(18), 9188 (2022)","journal-title":"Appl. Sci."},{"key":"553_CR29","first-page":"016555152211372","volume":"12","author":"A Khan","year":"2023","unstructured":"Khan, A.: Improved multi-lingual sentiment analysis and recognition using deep learning. J. Inform. Sci. 12, 01655515221137270 (2023)","journal-title":"J. Inform. Sci."}],"container-title":["International Journal of Data Science and Analytics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s41060-024-00553-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s41060-024-00553-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s41060-024-00553-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,23]],"date-time":"2025-03-23T07:17:50Z","timestamp":1742714270000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s41060-024-00553-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,7]]},"references-count":29,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2025,4]]}},"alternative-id":["553"],"URL":"https:\/\/doi.org\/10.1007\/s41060-024-00553-6","relation":{},"ISSN":["2364-415X","2364-4168"],"issn-type":[{"value":"2364-415X","type":"print"},{"value":"2364-4168","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,5,7]]},"assertion":[{"value":"19 June 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 April 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 May 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}