{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,14]],"date-time":"2026-02-14T21:36:10Z","timestamp":1771104970676,"version":"3.50.1"},"reference-count":72,"publisher":"MDPI AG","issue":"12","license":[{"start":{"date-parts":[[2021,12,6]],"date-time":"2021-12-06T00:00:00Z","timestamp":1638748800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100003141","name":"Consejo Nacional de Ciencia y Tecnolog\u00eda","doi-asserted-by":"publisher","award":["1061809"],"award-info":[{"award-number":["1061809"]}],"id":[{"id":"10.13039\/501100003141","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Data"],"abstract":"<jats:p>In this paper, the Mexican Emotional Speech Database (MESD) that contains single-word emotional utterances for anger, disgust, fear, happiness, neutral and sadness with adult (male and female) and child voices is described. To validate the emotional prosody of the uttered words, a cubic Support Vector Machines classifier was trained on the basis of prosodic, spectral and voice quality features for each case study: (1) male adult, (2) female adult and (3) child. In addition, cultural, semantic, and linguistic shaping of emotional expression was assessed by statistical analysis. This study was registered at BioMed Central and is part of the implementation of a published study protocol. Mean emotional classification accuracies yielded 93.3%, 89.4% and 83.3% for male, female and child utterances respectively. Statistical analysis emphasized the shaping of emotional prosodies by semantic and linguistic features. A cultural variation in emotional expression was highlighted by comparing the MESD with the INTERFACE for Castilian Spanish database. The MESD provides reliable content for linguistic emotional prosody shaped by the Mexican cultural environment. In order to facilitate further investigations, a corpus controlled for linguistic features and emotional semantics, as well as one containing words repeated across voices and emotions are provided. The MESD is made freely available.<\/jats:p>","DOI":"10.3390\/data6120130","type":"journal-article","created":{"date-parts":[[2021,12,7]],"date-time":"2021-12-07T02:48:13Z","timestamp":1638845293000},"page":"130","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":13,"title":["Mexican Emotional Speech Database Based on Semantic, Frequency, Familiarity, Concreteness, and Cultural Shaping of Affective Prosody"],"prefix":"10.3390","volume":"6","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7309-3541","authenticated-orcid":false,"given":"Mathilde Marie","family":"Duville","sequence":"first","affiliation":[{"name":"Escuela de Ingenier\u00eda y Ciencias, Tecnologico de Monterrey, Ave. Eugenio Garza Sada 2501, Monterrey 64849, Mexico"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2256-2958","authenticated-orcid":false,"given":"Luz Mar\u00eda","family":"Alonso-Valerdi","sequence":"additional","affiliation":[{"name":"Escuela de Ingenier\u00eda y Ciencias, Tecnologico de Monterrey, Ave. Eugenio Garza Sada 2501, Monterrey 64849, Mexico"}]},{"given":"David I.","family":"Ibarra-Zarate","sequence":"additional","affiliation":[{"name":"Escuela de Ingenier\u00eda y Ciencias, Tecnologico de Monterrey, Ave. Eugenio Garza Sada 2501, Monterrey 64849, Mexico"}]}],"member":"1968","published-online":{"date-parts":[[2021,12,6]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","unstructured":"Kami\u0144ska, D. (2019). Emotional Speech Recognition Based on the Committee of Classifiers. Entropy, 21.","DOI":"10.3390\/e21100920"},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"572","DOI":"10.1016\/j.patcog.2010.09.020","article-title":"Survey on Speech Emotion Recognition: Features, Classification Schemes, and Databases","volume":"44","author":"Kamel","year":"2011","journal-title":"Pattern Recognit."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"369","DOI":"10.1038\/s41562-019-0533-6","article-title":"The Primacy of Categories in the Recognition of 12 Emotions in Speech Prosody across Two Cultures","volume":"3","author":"Cowen","year":"2019","journal-title":"Nat. Hum. Behav."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"911","DOI":"10.1177\/0956797613517239","article-title":"Cultural Relativity in Perceiving Emotion From Vocalizations","volume":"25","author":"Gendron","year":"2014","journal-title":"Psychol. Sci."},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"686","DOI":"10.1037\/pspi0000066","article-title":"The Expression and Recognition of Emotions in the Voice across Five Nations: A Lens Model Analysis Based on Acoustic Features","volume":"111","author":"Laukka","year":"2016","journal-title":"J. Personal. Soc. Psychol."},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"90","DOI":"10.1177\/1754073912451332","article-title":"Nonverbal Dialects and Accents in Facial Expressions of Emotion","volume":"5","author":"Elfenbein","year":"2013","journal-title":"Emot. Rev."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"3","DOI":"10.1177\/1754073919897295","article-title":"Cross-Cultural Emotion Recognition and In-Group Advantage in Vocal Expression: A Meta-Analysis","volume":"13","author":"Laukka","year":"2021","journal-title":"Emot. Rev."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"34","DOI":"10.1016\/j.specom.2016.07.010","article-title":"Cross-Corpus Speech Emotion Recognition Based on Transfer Non-Negative Matrix Factorization","volume":"83","author":"Song","year":"2016","journal-title":"Speech Commun."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"162093","DOI":"10.1155\/2013\/162093","article-title":"Recognition of Emotions in Mexican Spanish Speech: An Approach Based on Acoustic Modelling of Emotion-Specific Vowels","volume":"2013","year":"2013","journal-title":"Sci. World J."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"5807","DOI":"10.1038\/s41598-020-62761-x","article-title":"Uncovering Electrophysiological and Vascular Signatures of Implicit Emotional Prosody","volume":"10","author":"Steber","year":"2020","journal-title":"Sci. Rep."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"1351","DOI":"10.1093\/scan\/nsx059","article-title":"Effects of Emotional Valence and Arousal on the Voice Perception Network","volume":"12","author":"Bestelmeyer","year":"2017","journal-title":"Soc. Cogn. Affect. Neurosci."},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"5406","DOI":"10.1038\/s41598-020-62225-2","article-title":"Assessing Autonomic Function from Electrodermal Activity and Heart Rate Variability During Cold-Pressor Test and Emotional Challenge","volume":"10","author":"Ghiasi","year":"2020","journal-title":"Sci. Rep."},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"401","DOI":"10.1080\/00207594.2011.626049","article-title":"In the Eye of the Beholder? Universality and Cultural Specificity in the Expression and Perception of Emotion","volume":"46","author":"Scherer","year":"2011","journal-title":"Int. J. Psychol."},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"218","DOI":"10.1016\/j.csl.2013.10.002","article-title":"Comparing the Acoustic Expression of Emotion in the Speaking and the Singing Voice","volume":"29","author":"Scherer","year":"2015","journal-title":"Comput. Speech Lang."},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Arruti, A., Cearreta, I., \u00c1lvarez, A., Lazkano, E., and Sierra, B. (2014). Feature Selection for Speech Emotion Recognition in Spanish and Basque: On the Use of Machine Learning to Improve Human-Computer Interaction. PLoS ONE, 9.","DOI":"10.1371\/journal.pone.0108975"},{"key":"ref_16","doi-asserted-by":"crossref","first-page":"271","DOI":"10.1016\/j.neucom.2017.07.050","article-title":"Speech Emotion Recognition Based on Feature Selection and Extreme Learning Machine Decision Tree","volume":"273","author":"Liu","year":"2018","journal-title":"Neurocomputing"},{"key":"ref_17","doi-asserted-by":"crossref","first-page":"56","DOI":"10.1016\/j.specom.2019.12.001","article-title":"Speech Emotion Recognition: Emotional Models, Databases, Features, Preprocessing Methods, Supporting Modalities, and Classifiers","volume":"116","year":"2020","journal-title":"Speech Commun."},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"104886","DOI":"10.1016\/j.knosys.2019.104886","article-title":"Bagged Support Vector Machines for Emotion Recognition from Speech","volume":"184","author":"Bhavan","year":"2019","journal-title":"Knowl.-Based Syst."},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"190","DOI":"10.1109\/TAFFC.2015.2457417","article-title":"The Geneva Minimalistic Acoustic Parameter Set (GeMAPS) for Voice Research and Affective Computing","volume":"7","author":"Eyben","year":"2016","journal-title":"IEEE Trans. Affect. Comput."},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"145","DOI":"10.1016\/j.neucom.2018.05.005","article-title":"Speech Emotion Recognition Based on an Improved Brain Emotion Learning Model","volume":"309","author":"Liu","year":"2018","journal-title":"Neurocomputing"},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"941","DOI":"10.1007\/s10772-018-9524-7","article-title":"Speaker Identification Based on Normalized Pitch Frequency and Mel Frequency Cepstral Coefficients","volume":"21","author":"Nasr","year":"2018","journal-title":"Int. J. Speech Technol."},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"93","DOI":"10.1007\/s10772-018-9491-z","article-title":"Databases, Features and Classifiers for Speech Emotion Recognition: A Review","volume":"21","author":"Swain","year":"2018","journal-title":"Int. J. Speech Technol."},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"205","DOI":"10.1007\/978-3-319-31056-5_11","article-title":"Analysis of Emotional Speech\u2014A Review","volume":"Volume 105","author":"Esposito","year":"2016","journal-title":"Toward Robotic Socially Believable Behaving Systems"},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"4459","DOI":"10.1007\/s00034-020-01377-y","article-title":"Excitation Features of Speech for Emotion Recognition Using Neutral Speech as Reference","volume":"39","author":"Kadiri","year":"2020","journal-title":"Circuits Syst. Signal Process."},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"12","DOI":"10.1177\/1754073920934544","article-title":"Beyond Correlation: Acoustic Transformation Methods for the Experimental Study of Emotional Voice and Speech","volume":"13","author":"Arias","year":"2021","journal-title":"Emot. Rev."},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Lammert, A.C., and Narayanan, S.S. (2015). On Short-Time Estimation of Vocal Tract Length from Formant Frequencies. PLoS ONE, 10.","DOI":"10.1371\/journal.pone.0132193"},{"key":"ref_27","doi-asserted-by":"crossref","first-page":"101100","DOI":"10.1016\/j.csl.2020.101100","article-title":"Vocal Tract Shaping of Emotional Speech","volume":"64","author":"Kim","year":"2020","journal-title":"Comput. Speech Lang."},{"key":"ref_28","doi-asserted-by":"crossref","first-page":"108046","DOI":"10.1016\/j.apacoust.2021.108046","article-title":"Improved Speech Emotion Recognition with Mel Frequency Magnitude Coefficient","volume":"179","author":"Ancilin","year":"2021","journal-title":"Appl. Acoust."},{"key":"ref_29","doi-asserted-by":"crossref","first-page":"328","DOI":"10.1016\/j.ins.2020.09.047","article-title":"Exploiting the Potentialities of Features for Speech Emotion Recognition","volume":"548","author":"Li","year":"2021","journal-title":"Inf. Sci."},{"key":"ref_30","unstructured":"Hozjan, V., Kacic, Z., Moreno, A., Bonafonte, A., and Nogueiras, A. (2002, January 29\u201331). Interface Databases: Design and Collection of a Multilingual Emotional Speech Database. Proceedings of the 3rd International Conference on Language Resources and Evaluation, LREC 2002 2024\u20132028 5, Las Palmas, Spain."},{"key":"ref_31","doi-asserted-by":"crossref","first-page":"9411","DOI":"10.1007\/s11042-020-10073-7","article-title":"Automatic Speech Recognition: A Survey","volume":"80","author":"Malik","year":"2020","journal-title":"Multimed. Tools Appl."},{"key":"ref_32","doi-asserted-by":"crossref","first-page":"1925","DOI":"10.3233\/JIFS-179963","article-title":"Recognize Basic Emotional Statesin Speech by Machine Learning Techniques Using Mel-Frequency Cepstral Coefficient Features","volume":"39","author":"Yang","year":"2020","journal-title":"J. Intell. Fuzzy Syst."},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"9","DOI":"10.1016\/j.specom.2020.11.003","article-title":"Two-Stage Dimensional Emotion Recognition by Fusing Predictions of Acoustic and Text Networks Using SVM","volume":"126","author":"Atmaja","year":"2021","journal-title":"Speech Commun."},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"69","DOI":"10.1109\/TAFFC.2015.2392101","article-title":"Speech Emotion Recognition Using Fourier Parameters","volume":"6","author":"Wang","year":"2015","journal-title":"IEEE Trans. Affect. Comput."},{"key":"ref_35","doi-asserted-by":"crossref","first-page":"22","DOI":"10.1016\/j.specom.2019.09.002","article-title":"Automatic Speech Emotion Recognition Using an Optimal Combination of Features Based on EMD-TKEO","volume":"114","author":"Kerkeni","year":"2019","journal-title":"Speech Commun."},{"key":"ref_36","doi-asserted-by":"crossref","first-page":"e12439","DOI":"10.1111\/spc3.12439","article-title":"Fostering Social and Emotional Intelligence: What Are the Best Current Strategies in Parenting?","volume":"13","author":"Segrin","year":"2019","journal-title":"Soc. Pers. Psychol. Compass"},{"key":"ref_37","doi-asserted-by":"crossref","unstructured":"Coutinho, E., and Schuller, B. (2017). Shared Acoustic Codes Underlie Emotional Communication in Music and Speech\u2014Evidence from Deep Transfer Learning. PLoS ONE, 12.","DOI":"10.1371\/journal.pone.0179289"},{"key":"ref_38","doi-asserted-by":"crossref","first-page":"243","DOI":"10.1017\/S0265051717000067","article-title":"The Role of Emotional Skills in Music Education","volume":"34","year":"2017","journal-title":"Brit. J. Music Ed."},{"key":"ref_39","doi-asserted-by":"crossref","unstructured":"Amado-Alonso, D., Le\u00f3n-del-Barco, B., Mendo-L\u00e1zaro, S., S\u00e1nchez-Miguel, P., and Iglesias Gallego, D. (2019). Emotional Intelligence and the Practice of Organized Physical-Sport Activity in Children. Sustainability, 11.","DOI":"10.3390\/su11061615"},{"key":"ref_40","doi-asserted-by":"crossref","unstructured":"LoBue, V., P\u00e9rez-Edgar, K., and Buss, K.A. (2019). Culture in Emotional Development. Handbook of Emotional Development, Springer International Publishing.","DOI":"10.1007\/978-3-030-17332-6"},{"key":"ref_41","doi-asserted-by":"crossref","first-page":"62","DOI":"10.1007\/978-3-642-24571-8_7","article-title":"EmoWisconsin: An Emotional Children Speech Database in Mexican Spanish","volume":"Volume 6975","author":"Graesser","year":"2011","journal-title":"Affective Computing and Intelligent Interaction"},{"key":"ref_42","doi-asserted-by":"crossref","first-page":"55","DOI":"10.1016\/j.csl.2019.06.006","article-title":"IESC-Child: An Interactive Emotional Children\u2019s Speech Corpus","volume":"59","year":"2020","journal-title":"Comput. Speech Lang."},{"key":"ref_43","doi-asserted-by":"crossref","first-page":"531","DOI":"10.1016\/j.jvoice.2006.03.002","article-title":"Acoustical Correlates of Affective Prosody","volume":"21","author":"Hammerschmidt","year":"2007","journal-title":"J. Voice"},{"key":"ref_44","doi-asserted-by":"crossref","first-page":"2","DOI":"10.1057\/s41599-020-0499-z","article-title":"Emotion Recognition and Confidence Ratings Predicted by Vocal Stimulus Type and Prosodic Parameters","volume":"7","author":"Lausen","year":"2020","journal-title":"Humanit. Soc. Sci. Commun."},{"key":"ref_45","doi-asserted-by":"crossref","first-page":"55","DOI":"10.3758\/s13415-014-0315-6","article-title":"The Dynamic Influence of Emotional Words on Sentence Processing","volume":"15","author":"Ding","year":"2015","journal-title":"Cogn. Affect. Behav. Neurosci."},{"key":"ref_46","doi-asserted-by":"crossref","first-page":"813","DOI":"10.1080\/23273798.2019.1620957","article-title":"Affective Neurolinguistics: Towards a Framework for Reconciling Language and Emotion","volume":"35","author":"Hinojosa","year":"2020","journal-title":"Lang. Cogn. Neurosci."},{"key":"ref_47","doi-asserted-by":"crossref","first-page":"107","DOI":"10.1016\/j.brainres.2007.03.015","article-title":"When Emotional Prosody and Semantics Dance Cheek to Cheek: ERP Evidence","volume":"1151","author":"Kotz","year":"2007","journal-title":"Brain Res."},{"key":"ref_48","doi-asserted-by":"crossref","first-page":"59","DOI":"10.1016\/j.bandl.2007.11.005","article-title":"An ERP Investigation on the Temporal Dynamics of Emotional Prosody and Emotional Semantics in Pseudo- and Lexical-Sentence Context","volume":"105","author":"Paulmann","year":"2008","journal-title":"Brain Lang."},{"key":"ref_49","doi-asserted-by":"crossref","first-page":"109","DOI":"10.1016\/j.bandc.2014.03.008","article-title":"Temporospatial Analysis of Explicit and Implicit Processing of Negative Content during Word Comprehension","volume":"87","author":"Hinojosa","year":"2014","journal-title":"Brain Cogn."},{"key":"ref_50","doi-asserted-by":"crossref","first-page":"231","DOI":"10.1016\/j.ijpsycho.2016.07.499","article-title":"Effects of Valence and Arousal on Emotional Word Processing Are Modulated by Concreteness: Behavioral and ERP Evidence from a Lexical Decision Task","volume":"110","author":"Yao","year":"2016","journal-title":"Int. J. Psychophysiol."},{"key":"ref_51","doi-asserted-by":"crossref","first-page":"14439","DOI":"10.1038\/s41598-019-50755-3","article-title":"Differential Impact of Emotion on Semantic Processing of Abstract and Concrete Words: ERP and FMRI Evidence","volume":"9","author":"Pauligk","year":"2019","journal-title":"Sci. Rep."},{"key":"ref_52","doi-asserted-by":"crossref","first-page":"95","DOI":"10.1016\/j.biopsycho.2008.03.010","article-title":"Early Emotion Word Processing: Evidence from Event-Related Potentials","volume":"80","author":"Scott","year":"2009","journal-title":"Biol. Psychol."},{"key":"ref_53","doi-asserted-by":"crossref","first-page":"250","DOI":"10.1016\/j.neulet.2011.03.026","article-title":"Word Frequency Modulates the Processing of Emotional Words: Convergent Behavioral and Electrophysiological Data","volume":"494","author":"Pozo","year":"2011","journal-title":"Neurosci. Lett."},{"key":"ref_54","doi-asserted-by":"crossref","unstructured":"Hinojosa, J.A., Rinc\u00f3n-P\u00e9rez, I., Romero-Ferreiro, M.V., Mart\u00ednez-Garc\u00eda, N., Villalba-Garc\u00eda, C., Montoro, P.R., and Pozo, M.A. (2016). The Madrid Affective Database for Spanish (MADS): Ratings of Dominance, Familiarity, Subjective Age of Acquisition and Sensory Experience. PLoS ONE, 11.","DOI":"10.1371\/journal.pone.0155866"},{"key":"ref_55","doi-asserted-by":"crossref","first-page":"395","DOI":"10.3758\/s13428-011-0165-x","article-title":"Affective Norms for 380 Spanish Words Belonging to Three Different Semantic Categories","volume":"44","author":"Guasch","year":"2012","journal-title":"Behav. Res."},{"key":"ref_56","doi-asserted-by":"crossref","first-page":"75","DOI":"10.1037\/emo0000302","article-title":"Universals and Cultural Variations in 22 Emotional Expressions across Five Cultures","volume":"18","author":"Cordaro","year":"2018","journal-title":"Emotion"},{"key":"ref_57","doi-asserted-by":"crossref","first-page":"145","DOI":"10.1016\/j.copsyc.2017.07.014","article-title":"Revisiting Diversity: Cultural Variation Reveals the Constructed Nature of Emotion Perception","volume":"17","author":"Gendron","year":"2017","journal-title":"Curr. Opin. Psychol."},{"key":"ref_58","doi-asserted-by":"crossref","first-page":"445","DOI":"10.1037\/a0036048","article-title":"Evidence for Cultural Dialects in Vocal Emotion Expression: Acoustic Classification within and across Five Nations","volume":"14","author":"Laukka","year":"2014","journal-title":"Emotion"},{"key":"ref_59","doi-asserted-by":"crossref","first-page":"117","DOI":"10.1037\/emo0000100","article-title":"The Voice Conveys Emotion in Ten Globalized Cultures and One Remote Village in Bhutan","volume":"16","author":"Cordaro","year":"2016","journal-title":"Emotion"},{"key":"ref_60","doi-asserted-by":"crossref","first-page":"230","DOI":"10.1080\/02699931.2013.812033","article-title":"Cross-Cultural Emotional Prosody Recognition: Evidence from Chinese and British Listeners","volume":"28","author":"Paulmann","year":"2014","journal-title":"Cogn. Emot."},{"key":"ref_61","doi-asserted-by":"crossref","first-page":"101366","DOI":"10.1016\/j.infbeh.2019.101366","article-title":"Similarities and Differences between Western Cultures: Toddler Temperament and Parent-Child Interactions in the United States (US) and Germany","volume":"57","author":"Kirchhoff","year":"2019","journal-title":"Infant Behav. Dev."},{"key":"ref_62","doi-asserted-by":"crossref","first-page":"204","DOI":"10.1002\/(SICI)1098-2302(199911)35:3<204::AID-DEV5>3.0.CO;2-V","article-title":"Prenatal Experience and Neonatal Responsiveness to Vocal Expressions of Emotion","volume":"35","author":"Mastropieri","year":"1999","journal-title":"Dev. Psychobiol."},{"key":"ref_63","doi-asserted-by":"crossref","first-page":"8659","DOI":"10.1038\/s41598-018-26889-1","article-title":"The Development of Cross-Cultural Recognition of Vocal Emotion during Childhood and Adolescence","volume":"8","author":"Chronaki","year":"2018","journal-title":"Sci. Rep."},{"key":"ref_64","doi-asserted-by":"crossref","first-page":"106","DOI":"10.1016\/j.neubiorev.2016.08.016","article-title":"The Development of Social Cognition in Adolescence: An Integrated Perspective","volume":"70","author":"Kilford","year":"2016","journal-title":"Neurosci. Biobehav. Rev."},{"key":"ref_65","doi-asserted-by":"crossref","first-page":"46474","DOI":"10.1109\/ACCESS.2019.2905597","article-title":"Automatic Voice Pathology Monitoring Using Parallel Deep Models for Smart Healthcare","volume":"7","author":"Alhussein","year":"2019","journal-title":"IEEE Access"},{"key":"ref_66","doi-asserted-by":"crossref","first-page":"115885","DOI":"10.1016\/j.eswa.2021.115885","article-title":"Automatic Children\u2019s Personality Assessment from Emotional Speech","volume":"187","year":"2022","journal-title":"Expert Syst. Appl."},{"key":"ref_67","doi-asserted-by":"crossref","unstructured":"Duville, M.M., Alonso-Valerdi, L.M., and Ibarra-Zarate, D.I. (2020). Improving Emotional Speech Processing in Autism Spectrum Disorders: Toward the Elaboration of a Drug-Free Intervention Based on Social StoriesTM and NAO Social Robot Interactions. ISRCTN.","DOI":"10.1186\/ISRCTN18117434"},{"key":"ref_68","doi-asserted-by":"crossref","first-page":"626146","DOI":"10.3389\/fnhum.2021.626146","article-title":"Electroencephalographic Correlate of Mexican Spanish Emotional Speech Processing in Autism Spectrum Disorder: To a Social Story and Robot-Based Intervention","volume":"15","author":"Duville","year":"2021","journal-title":"Front. Hum. Neurosci."},{"key":"ref_69","doi-asserted-by":"crossref","first-page":"272","DOI":"10.3758\/s13428-015-0572-5","article-title":"Affective Norms of 875 Spanish Words for Five Discrete Emotional Categories and Two Emotional Dimensions","volume":"48","author":"Hinojosa","year":"2016","journal-title":"Behav. Res."},{"key":"ref_70","doi-asserted-by":"crossref","first-page":"189","DOI":"10.1016\/j.neucom.2019.10.118","article-title":"A Comprehensive Survey on Support Vector Machine Classification: Applications, Challenges and Trends","volume":"408","author":"Cervantes","year":"2020","journal-title":"Neurocomputing"},{"key":"ref_71","doi-asserted-by":"crossref","first-page":"520","DOI":"10.1016\/j.catena.2018.03.003","article-title":"Review on Landslide Susceptibility Mapping Using Support Vector Machines","volume":"165","author":"Huang","year":"2018","journal-title":"CATENA"},{"key":"ref_72","unstructured":"Duville, M.M., Alonso-Valerdi, L.M., and Ibarra-Zarate, D.I. (2021). Mexican Emotional Speech Database (MESD). Mendeley Data V2."}],"container-title":["Data"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2306-5729\/6\/12\/130\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T07:42:09Z","timestamp":1760168529000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2306-5729\/6\/12\/130"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,12,6]]},"references-count":72,"journal-issue":{"issue":"12","published-online":{"date-parts":[[2021,12]]}},"alternative-id":["data6120130"],"URL":"https:\/\/doi.org\/10.3390\/data6120130","relation":{},"ISSN":["2306-5729"],"issn-type":[{"value":"2306-5729","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,12,6]]}}}