{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,21]],"date-time":"2026-03-21T21:09:25Z","timestamp":1774127365088,"version":"3.50.1"},"reference-count":67,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2016,10,28]],"date-time":"2016-10-28T00:00:00Z","timestamp":1477612800000},"content-version":"unspecified","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/100000009","name":"Foundation for the National Institutes of Health","doi-asserted-by":"publisher","award":["Grant R01 HD060789"],"award-info":[{"award-number":["Grant R01 HD060789"]}],"id":[{"id":"10.13039\/100000009","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Speech Technol"],"published-print":{"date-parts":[[2017,3]]},"DOI":"10.1007\/s10772-016-9364-2","type":"journal-article","created":{"date-parts":[[2016,10,28]],"date-time":"2016-10-28T03:18:21Z","timestamp":1477624701000},"page":"27-41","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":24,"title":["Enhanced multiclass SVM with thresholding fusion for speech-based emotion classification"],"prefix":"10.1007","volume":"20","author":[{"given":"Na","family":"Yang","sequence":"first","affiliation":[]},{"given":"Jianbo","family":"Yuan","sequence":"additional","affiliation":[]},{"given":"Yun","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Ilker","family":"Demirkol","sequence":"additional","affiliation":[]},{"given":"Zhiyao","family":"Duan","sequence":"additional","affiliation":[]},{"given":"Wendi","family":"Heinzelman","sequence":"additional","affiliation":[]},{"given":"Melissa","family":"Sturge-Apple","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2016,10,28]]},"reference":[{"key":"9364_CR4","unstructured":"Al\u00a0Machot, F., Mosa, A. H., Dabbour, K., Fasih, A., Schwarzlmuller, C., Ali, M., & Kyamakya, K. (2011). A novel real-time emotion detection system from audio streams based on Bayesian quadratic discriminate classifier for ADAS. In Nonlinear Dynamics and Synchronization 16th Int\u2019l Symposium on Theoretical Electrical Engineering, Joint 3rd Int\u2019l Workshop on."},{"key":"9364_CR5","doi-asserted-by":"crossref","unstructured":"Ang, J., Dhillon, R., Krupski, A., Shriberg, E., & Stolcke, A. (2002). Prosody-based automatic detection of annoyance and frustration in human-computer dialog. In Proceeings of International Conference on Spoken Language Processing (pp. 2037\u20132040).","DOI":"10.21437\/ICSLP.2002-559"},{"key":"9364_CR6","volume-title":"Handbook of research methods in social psychology","author":"R Bakeman","year":"1997","unstructured":"Bakeman, R. (1997). Behavioral observation and coding. Handbook of research methods in social psychology. Cambridge: Cambridge University Press."},{"issue":"1","key":"9364_CR7","doi-asserted-by":"crossref","first-page":"31","DOI":"10.1007\/s10919-013-0165-x","volume":"38","author":"T B\u00e4nziger","year":"2014","unstructured":"B\u00e4nziger, T., Patel, S., & Scherer, K. R. (2014). The role of perceived voice and speech characteristics in vocal emotion communication. Journal of nonverbal behavior, 38(1), 31\u201352.","journal-title":"Journal of nonverbal behavior"},{"key":"9364_CR8","doi-asserted-by":"crossref","unstructured":"Bao, H., Xu, M. X., & Zheng, T. F. (2007). Emotion attribute projection for speaker recognition on emotional speech. In Procceedings of Interspeech (pp. 758\u2013761).","DOI":"10.21437\/Interspeech.2007-142"},{"issue":"5","key":"9364_CR9","doi-asserted-by":"crossref","first-page":"394","DOI":"10.1016\/j.specom.2009.12.007","volume":"52","author":"R Barra-Chicote","year":"2010","unstructured":"Barra-Chicote, R., Yamagishi, J., King, S., Montero, J. M., & Macias-Guarasa, J. (2010). Analysis of statistical parametric and unit selection speech synthesis systems applied to emotional speech. Speech Communication, 52(5), 394\u2013404.","journal-title":"Speech Communication"},{"key":"9364_CR10","unstructured":"Batliner, A., Steidl, S., Schuller, B., Seppi, D., Laskowski, K., Vogt, T., Devillers, L., Vidrascu, L., Amir, N., Kessous, L., & Aharonson, V. (2006). Combining efforts for improving automatic classification of emotional user states. In Proceedings of the Fifth Slovenian and First International Language Technologies Conference."},{"issue":"3","key":"9364_CR11","doi-asserted-by":"crossref","first-page":"506","DOI":"10.1111\/j.1467-8640.2012.00457.x","volume":"29","author":"JR Bellegarda","year":"2013","unstructured":"Bellegarda, J. R. (2013). Data-driven analysis of emotion in text using latent affective folding and embedding. Computational Intelligence, 29(3), 506\u2013526.","journal-title":"Computational Intelligence"},{"issue":"7\u20138","key":"9364_CR12","doi-asserted-by":"crossref","first-page":"613","DOI":"10.1016\/j.specom.2010.02.010","volume":"52","author":"D Bitouk","year":"2010","unstructured":"Bitouk, D., Ragini, V., & Ani, N. (2010). Class-level spectral features for emotion recognition. Journal of Speech Communication, 52(7\u20138), 613\u2013625.","journal-title":"Journal of Speech Communication"},{"issue":"1","key":"9364_CR13","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.specom.2011.12.003","volume":"55","author":"MP Black","year":"2013","unstructured":"Black, M. P., Katsamanis, A., Baucom, B. R., Lee, C. C., Lammert, A. C., & Christensen, A. (2013). Toward automating a human behavioral coding system for married couples\u2019 interactions using speech acoustic features. Speech communication, 55(1), 1\u201321.","journal-title":"Speech communication"},{"key":"9364_CR14","unstructured":"Chang, K., Fisher, D., & Canny, J. (2011). AMMON: a speech analysis library for analyzing affect, stress, and mental health on mobile phones. In 2nd International Workshop on Sensing Applications on Mobile Phones."},{"issue":"1","key":"9364_CR15","doi-asserted-by":"crossref","first-page":"321","DOI":"10.1613\/jair.953","volume":"16","author":"NV Chawla","year":"2002","unstructured":"Chawla, N. V., Bowyer, K. W., Hall, L. O., & Kegelmeyer, W. P. (2002). SMOTE: synthetic minority over-sampling technique. Journal of Artificial Intelligence Research, 16(1), 321\u2013357.","journal-title":"Journal of Artificial Intelligence Research"},{"key":"9364_CR16","unstructured":"Cowie, R., Douglas-Cowie, E., Savvidou, S., McMahon, E., Sawey, M., & Schr\u00f6der, M. (2000). \u2018FEELTRACE\u2019: an instrument for recording perceived emotion in real time. In ISCA Tutorial and Research Workshop (ITRW) on Speech and Emotion."},{"key":"9364_CR17","doi-asserted-by":"crossref","unstructured":"Eskimez, S. E., Imade, K., Yang, N., Sturge-Appley, M., Duan, Z., & Heinzelman, W. (2016). Emotion classification: How does an automated system compare to naive human coders? In Acoustics, Speech and Signal Processing, Proceedings of the IEEE International Conference on.","DOI":"10.1109\/ICASSP.2016.7472082"},{"key":"9364_CR18","doi-asserted-by":"crossref","unstructured":"Farr\u00fas, M., Ejarque, P., Temko, A., & Hernando, J. (2007). Histogram equalization in SVM multimodal person verification. In Proceedings of IAPR\/IEEE International Conference on Biometrics.","DOI":"10.1007\/978-3-540-74549-5_86"},{"key":"9364_CR19","doi-asserted-by":"crossref","unstructured":"Goudbeek, M., Goldman, J. P., & Scherer, K. R. (2009). Emotion dimensions and formant position. In INTERSPEECH (pp. 1575\u20131578).","DOI":"10.21437\/Interspeech.2009-469"},{"key":"9364_CR20","unstructured":"Goyal, A., Riloff, E., Daum\u00e9 III, H., & Gilbert, N. (2010). Toward plot units: automatic affect state analysis. In Proceedings of HLT\/NAACL Workshop on Computational Approaches to Analysis and Generation of Emotion in Text (CAET)."},{"key":"9364_CR21","doi-asserted-by":"crossref","unstructured":"Gupta P., & Rajput N. (2007). Two-stream emotion recognition for call center monitoring. In INTERSPEECH (pp. 2241\u20132244).","DOI":"10.21437\/Interspeech.2007-609"},{"key":"9364_CR22","doi-asserted-by":"crossref","first-page":"42","DOI":"10.1007\/11821830_4","volume-title":"Intelligent virtual agents","author":"M Hoque","year":"2006","unstructured":"Hoque, M., Yeasin, M., & Louwerse, M. (2006). Robust recognition of emotion from speech. Intelligent virtual agents (pp. 42\u201353). Berlin: Springer. Lecture notes in computer science."},{"key":"9364_CR23","unstructured":"Hsu, C.W., Chang ,C.C., & Lin, C.J. (2003). A practical guide to support vector classification."},{"key":"9364_CR24","volume-title":"Spoken language processing","author":"X Huang","year":"2001","unstructured":"Huang, X., Acero, A., & Hon, H. W. (2001). Spoken language processing. New Jersey: Prentice Hall PTR."},{"key":"9364_CR25","doi-asserted-by":"crossref","unstructured":"Huisman, G., Van\u00a0Hout, M., van Dijk, E., van\u00a0der Geest, T., & Heylen, D. (2013). Lemtool\u2014measuring emotions in visual interfaces. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems.","DOI":"10.1145\/2470654.2470706"},{"key":"9364_CR1","unstructured":"Implementation of extracting MFCCs included in the VOICEBOX toolkit. http:\/\/www.ee.ic.ac.uk\/hp\/staff\/dmb\/voicebox\/voicebox.html ."},{"issue":"2","key":"9364_CR26","first-page":"49","volume":"2","author":"NHD Jong","year":"2007","unstructured":"Jong, N. H. D., & Wempe, T. (2007). Automatic measurement of speech rate in spoken Dutch. ACLC Working Papers, 2(2), 49\u201358.","journal-title":"ACLC Working Papers"},{"key":"9364_CR27","doi-asserted-by":"crossref","unstructured":"Kawanami, H., Iwami, Y., Toda, T., Saruwatari, H., & Shikano, K. (2003). GMM-based voice conversion applied to emotional speech synthesis. In Proceedings of Eurospeech.","DOI":"10.21437\/Eurospeech.2003-661"},{"key":"9364_CR28","doi-asserted-by":"crossref","DOI":"10.4324\/9781410610843","volume-title":"Couple observational coding systems","author":"P Kerig","year":"2004","unstructured":"Kerig, P., & Baucom, D. (2004). Couple observational coding systems. Abington: Routledge."},{"key":"9364_CR29","doi-asserted-by":"crossref","unstructured":"Kwon, O.W., Chan, K., Hao, J., & Lee, T. W. (2003). Emotion recognition by speech signals. In EUROSPEECH. (pp. 125\u2013128).","DOI":"10.21437\/Eurospeech.2003-80"},{"key":"9364_CR30","doi-asserted-by":"crossref","unstructured":"Lee, C., & Lee, G. G. (2007). Emotion recognition for affective user interfaces using natural language dialogs. In Procceedings of IEEE International Symposium on Robot and Human interactive Communication. (pp. 798\u2013801).","DOI":"10.1109\/ROMAN.2007.4415194"},{"issue":"2","key":"9364_CR31","doi-asserted-by":"crossref","first-page":"293","DOI":"10.1109\/TSA.2004.838534","volume":"13","author":"CM Lee","year":"2005","unstructured":"Lee, C. M., & Narayanan, S. S. (2005). Toward detecting emotions in spoken dialogs. IEEE Transactions on Speech and Audio Processing, 13(2), 293\u2013303.","journal-title":"IEEE Transactions on Speech and Audio Processing"},{"key":"9364_CR32","doi-asserted-by":"crossref","unstructured":"Lee, C. M., Narayanan, S. S., & Pieraccini, R. (2002). Combining acoustic and language information for emotion recognition. In Proceeding of 7th International Conference on Spoken Language Processing.","DOI":"10.21437\/ICSLP.2002-296"},{"key":"9364_CR33","doi-asserted-by":"crossref","unstructured":"Lee, L., & Rose, R. C. (1996). Speaker normalization using efficient frequency warping procedures. In IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), vol\u00a01 (pp. 353\u2013356).","DOI":"10.1109\/ICASSP.1996.541105"},{"key":"9364_CR34","volume-title":"Emotional prosody speech and transcripts","author":"M Liberman","year":"2002","unstructured":"Liberman, M., Davis, K., Grossman, M., Martey, N., & Bell, J. (2002). Emotional prosody speech and transcripts. Philadelphia: Linguistic Data Consortium (LDC)."},{"key":"9364_CR35","unstructured":"Ling, C., Dong, M., Li, H., Yu, Z. L., & Chan, P. (2010). Machine learning methods in the application of speech emotion recognition. In Application of Machine Learning (pp. 1\u201319)."},{"key":"9364_CR2","unstructured":"MATLAB implementation of mutual information. 2007. http:\/\/www.mathworks.com\/matlabcentral\/fileexchange\/14888-mutual-information-computation ."},{"key":"9364_CR36","unstructured":"\u00d6zkul, S., Bozkurt, E., Asta, S., Yemez, Y., & Erzin, E. (2012). Multimodal analysis of upper-body gestures, facial expressions and speech. In Procceedings of the 4th International Workshop on Corpora for Research on Emotion Sentiment and Social Signals."},{"issue":"8","key":"9364_CR37","doi-asserted-by":"crossref","first-page":"1226","DOI":"10.1109\/TPAMI.2005.159","volume":"27","author":"H Peng","year":"2005","unstructured":"Peng, H., Long, F., & Ding, C. (2005). Feature selection based on mutual information: criteria of max-dependency, max-relevance, and min-redundancy. IEEE Transactions on Pattern Analysis and Mchine Intelligence, 27(8), 1226\u20131238.","journal-title":"IEEE Transactions on Pattern Analysis and Mchine Intelligence"},{"key":"9364_CR38","first-page":"185","volume-title":"Advances in Kernel Methods","author":"JC Platt","year":"1999","unstructured":"Platt, J. C. (1999). Fast training of support vector machines using sequential minimal optimization. Advances in Kernel Methods (pp. 185\u2013208). Cambridge: MIT Press."},{"key":"9364_CR39","doi-asserted-by":"crossref","unstructured":"Qin, L., Ling, Z. H., Wu, Y. J., Zhang, B. F., & Wang, R. H. (2006). Hmm-based emotional speech synthesis using average emotion model. In Procceedings of Chinese Spoken Language Processing (pp. 233\u2013240).","DOI":"10.1007\/11939993_27"},{"key":"9364_CR40","doi-asserted-by":"crossref","unstructured":"Rachuri, K. K., Musolesi, M., Mascolo, C., Rentfrow, P. J., Longworth, C., & Aucinas, A. (2010). EmotionSense: a mobile phones based adaptive platform for experimental social psychology research. In Proceedings of the 12th ACM International Conference on Ubiquitous Computing (pp. 281\u2013290).","DOI":"10.1145\/1864349.1864393"},{"issue":"4","key":"9364_CR41","doi-asserted-by":"crossref","first-page":"537","DOI":"10.1109\/72.298224","volume":"5","author":"B Roberto","year":"1994","unstructured":"Roberto, B. (1994). Using mutual information for selecting features in supervised neural net learning. IEEE Transactions on Neural Networks, 5(4), 537\u2013550.","journal-title":"IEEE Transactions on Neural Networks"},{"issue":"3","key":"9364_CR42","doi-asserted-by":"crossref","first-page":"315","DOI":"10.1016\/j.ipm.2008.09.003","volume":"45","author":"J Rong","year":"2009","unstructured":"Rong, J., Li, G., & Chen, Y. P. P. (2009). Acoustic feature selection for automatic emotion recognition from speech. Information Processing and Management, 45(3), 315\u2013328.","journal-title":"Information Processing and Management"},{"issue":"11","key":"9364_CR43","first-page":"2251","volume":"63","author":"DA Sauter","year":"2010","unstructured":"Sauter, D. A., Eisner, F., Calder, A. J., & Scott, S. K. (2010). Perceptual cues in nonverbal vocal expressions of emotion, 63(11), 2251\u20132272.","journal-title":"Perceptual cues in nonverbal vocal expressions of emotion"},{"issue":"1\u20132","key":"9364_CR44","doi-asserted-by":"crossref","first-page":"227","DOI":"10.1016\/S0167-6393(02)00084-5","volume":"40","author":"KR Scherer","year":"2003","unstructured":"Scherer, K. R. (2003). Vocal communication of emotion: a review of research paradigms. Speech Communication, 40(1\u20132), 227\u2013256.","journal-title":"Speech Communication"},{"issue":"4","key":"9364_CR45","doi-asserted-by":"crossref","first-page":"695","DOI":"10.1177\/0539018405058216","volume":"44","author":"KR Scherer","year":"2005","unstructured":"Scherer, K. R. (2005). What are emotions? And how can they be measured? Social Science Information, 44(4), 695\u2013729.","journal-title":"Social Science Information"},{"key":"9364_CR46","unstructured":"Schuller, B., Rigoll, G., & Lang, M. (2003). Hidden markov model-based speech emotion recognition. In: IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), vol\u00a02 (p. 1)."},{"key":"9364_CR47","doi-asserted-by":"crossref","unstructured":"Schuller, B., Rigoll, G., & Lang, M. (2004). Speech emotion recognition combining acoustic features and linguistic information in a hybrid support vector machine-belief network architecture. In IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), vol\u00a01 (p. 577).","DOI":"10.1109\/ICASSP.2004.1326051"},{"key":"9364_CR48","doi-asserted-by":"crossref","unstructured":"Schuller, B., Vlasenko, B., Minguez, R., Rigoll, G., & Wendemuth, A. (2007). Comparing one and two-stage acoustic modeling in the recognition of emotion in speech. In: IEEE Workshop on Automatic Speech Recognition Understanding (pp. 596\u2013600).","DOI":"10.1109\/ASRU.2007.4430180"},{"key":"9364_CR49","doi-asserted-by":"crossref","unstructured":"Sethu, V., Ambikairajah, E., & Epps, J. (2008). Empirical mode decomposition based weighted frequency feature for speech-based emotion classification. In IEEE International Conference on Acoustics, Speech and Signal Processing (pp. 5017\u20135020).","DOI":"10.1109\/ICASSP.2008.4518785"},{"key":"9364_CR50","doi-asserted-by":"crossref","unstructured":"Shafran, I. (2005). A comparison of classifiers for detecting emotion from speech. In IEEE International Conference on Acoustics, Speech and Signal Processing.","DOI":"10.1109\/ICASSP.2005.1415120"},{"key":"9364_CR51","unstructured":"Shrawankar, U., & Thakare, V.M. (2013). Adverse conditions and ASR techniques for robust speech user interface. arXiv preprint arXiv:13035515 ."},{"key":"9364_CR52","doi-asserted-by":"crossref","unstructured":"Steidl, S., Polzehl, T., Bunnell, H. T., Dou, Y., Muthukumar, P. K., Perry, D., Prahallad, K., Vaughn, C., Black, A. W., & Metze, F. (2012). Emotion identification for evaluation of synthesized emotional speech. In Procceedings of Speech Prosody.","DOI":"10.21437\/SpeechProsody.2012-165"},{"key":"9364_CR53","doi-asserted-by":"crossref","unstructured":"Tacconi, D., Mayora, O., Lukowicz, P., Arnrich, B., Setz, C., Troster, G., & Haring, C. (2008). Activity and emotion recognition to support early diagnosis of psychiatric diseases. In Pervasive Computing Technologies for Healthcare (PervasiveHealth), Second International Conference on (pp. 100\u2013102).","DOI":"10.1109\/PCTHEALTH.2008.4571041"},{"key":"9364_CR54","doi-asserted-by":"crossref","unstructured":"Tang, H., Chu, S. M., Hasegawa-Johnson, M., & Huang, T. S. (2009). Emotion recognition from speech via boosted Gaussian mixture models. In IEEE International Conference on Multimedia and Expo (ICME) (pp. 294\u2013297).","DOI":"10.1109\/ICME.2009.5202493"},{"key":"9364_CR55","doi-asserted-by":"crossref","DOI":"10.1007\/978-1-4757-2440-0","volume-title":"The nature of statistical learning theory","author":"VN Vapnik","year":"1995","unstructured":"Vapnik, V. N. (1995). The nature of statistical learning theory. New York: Springer."},{"key":"9364_CR56","volume-title":"Statistical learning theory","author":"VN Vapnik","year":"1998","unstructured":"Vapnik, V. N. (1998). Statistical learning theory. New Jersey: Wiley."},{"key":"9364_CR57","unstructured":"Varga A. P., Steeneken H. J. M., Tomlinson M., & Jones D. (1992). NOISEX-92 study on the effect of additive noise on automatic speech recognition. http:\/\/spib.ece.rice.edu\/spib\/data\/signals\/noise\/ ."},{"key":"9364_CR58","doi-asserted-by":"crossref","unstructured":"Vlasenko, B., Schuller, B., Wendemuth, A., & Rigoll, G. (2007). Combining frame and turn-level information for robust recognition of emotions within speech. In INTERSPEECH. (pp. 2249\u20132252).","DOI":"10.21437\/Interspeech.2007-611"},{"key":"9364_CR3","unstructured":"Wireless communication and networking group, University of Rochester. 2016. http:\/\/www.ece.rochester.edu\/projects\/wcng\/project_bridge.html ."},{"issue":"8","key":"9364_CR59","doi-asserted-by":"crossref","first-page":"1880","DOI":"10.1109\/TMM.2013.2269314","volume":"15","author":"CH Wu","year":"2013","unstructured":"Wu, C. H., Kung, C., Lin, J. C., & Wei, W. L. (2013). Two-level hierarchical alignment for semi-coupled hmm-based audiovisual emotion recognition with temporal course. IEEE Transactions on Multimedia, 15(8), 1880\u20131895.","journal-title":"IEEE Transactions on Multimedia"},{"key":"9364_CR60","unstructured":"Wu, G., & Chang, E. Y. (2003). Class-boundary alignment for imbalanced dataset learning. In Workshop on Learning from Imbalanced Datasets II, ICML (pp. 49\u201356)."},{"key":"9364_CR61","doi-asserted-by":"crossref","unstructured":"Wu, S., Falk, T. H., & Chan, W. Y. (2009). Automatic recognition of speech emotion using long-term spectro-temporal features. In Procceedings of the 16th International Conference on Digital Signal Processing.","DOI":"10.1109\/ICDSP.2009.5201047"},{"key":"9364_CR62","doi-asserted-by":"crossref","unstructured":"Xia, R., & Liu, Y. (2012). Using i-vector space model for emotion recognition. In Procceedings of Interspeech.","DOI":"10.21437\/Interspeech.2012-128"},{"issue":"12","key":"9364_CR63","doi-asserted-by":"crossref","first-page":"1833","DOI":"10.1109\/TASLP.2014.2352453","volume":"22","author":"N Yang","year":"2014","unstructured":"Yang, N., Ba, H., Cai, W., Demirkol, I., & Heinzelman, W. (2014). BaNa: a noise resilient fundamental frequency detection algorithm for speech and music. IEEE\/ACM Transactions on Audio, Speech, and Language Processing, 22(12), 1833\u20131848.","journal-title":"IEEE\/ACM Transactions on Audio, Speech, and Language Processing"},{"key":"9364_CR64","doi-asserted-by":"crossref","unstructured":"Yang, N., Muraleedharan, R., Kohl, J., Demirkol, I., Heinzelman, W., & Sturge-Apple, M. (2012). Speech-based emotion classification using multiclass SVM with hybrid kernel and thresholding fusion. In Spoken Language Technology Workshop (SLT), 2012 IEEE (pp. 455\u2013460).","DOI":"10.1109\/SLT.2012.6424267"},{"issue":"2","key":"9364_CR65","doi-asserted-by":"crossref","first-page":"142","DOI":"10.1109\/T-AFFC.2012.38","volume":"4","author":"Y Yang","year":"2013","unstructured":"Yang, Y., Fairbairn, C., & Cohn, J. F. (2013). Detecting depression severity from vocal prosody. IEEE Transactions on Affective Computing, 4(2), 142\u2013150.","journal-title":"IEEE Transactions on Affective Computing"},{"issue":"2","key":"9364_CR66","doi-asserted-by":"crossref","first-page":"585","DOI":"10.1109\/TASL.2011.2162405","volume":"20","author":"S Yun","year":"2012","unstructured":"Yun, S., & Yoo, C. D. (2012). Loss-scaled large-margin gaussian mixture models for speech emotion classification. IEEE Transactions on Audio, Speech, and Language Processing, 20(2), 585\u2013598.","journal-title":"IEEE Transactions on Audio, Speech, and Language Processing"},{"key":"9364_CR67","doi-asserted-by":"publisher","unstructured":"Zhang S., Zhao X., & Lei B. (2013). Speech emotion recognition using an enhanced kernel isomap for human-robot interaction. International Journal of Advanced Robotic Systems. doi: 10.5772\/55403 .","DOI":"10.5772\/55403"}],"container-title":["International Journal of Speech Technology"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s10772-016-9364-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/article\/10.1007\/s10772-016-9364-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s10772-016-9364-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,6,20]],"date-time":"2024-06-20T06:58:20Z","timestamp":1718866700000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/s10772-016-9364-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2016,10,28]]},"references-count":67,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2017,3]]}},"alternative-id":["9364"],"URL":"https:\/\/doi.org\/10.1007\/s10772-016-9364-2","relation":{},"ISSN":["1381-2416","1572-8110"],"issn-type":[{"value":"1381-2416","type":"print"},{"value":"1572-8110","type":"electronic"}],"subject":[],"published":{"date-parts":[[2016,10,28]]}}}