{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,8]],"date-time":"2025-09-08T06:53:38Z","timestamp":1757314418098,"version":"3.37.3"},"reference-count":65,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2019,8,20]],"date-time":"2019-08-20T00:00:00Z","timestamp":1566259200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2019,8,20]],"date-time":"2019-08-20T00:00:00Z","timestamp":1566259200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/100000002","name":"National Institutes of Health","doi-asserted-by":"publisher","award":["R01 R01DC006859"],"award-info":[{"award-number":["R01 R01DC006859"]}],"id":[{"id":"10.13039\/100000002","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J AUDIO SPEECH MUSIC PROC."],"published-print":{"date-parts":[[2019,12]]},"DOI":"10.1186\/s13636-019-0157-9","type":"journal-article","created":{"date-parts":[[2019,8,20]],"date-time":"2019-08-20T14:09:12Z","timestamp":1566310152000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":8,"title":["Articulation constrained learning with application to speech emotion recognition"],"prefix":"10.1186","volume":"2019","author":[{"given":"Mohit","family":"Shah","sequence":"first","affiliation":[]},{"given":"Ming","family":"Tu","sequence":"additional","affiliation":[]},{"given":"Visar","family":"Berisha","sequence":"additional","affiliation":[]},{"given":"Chaitali","family":"Chakrabarti","sequence":"additional","affiliation":[]},{"given":"Andreas","family":"Spanias","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2019,8,20]]},"reference":[{"key":"157_CR1","doi-asserted-by":"crossref","unstructured":"P. Partila, J. Tovarek, J. Frnda, M. Voznak, M. Penhaker, T. Peterek, in Intelligent Data Analysis and Its Applications, Volume II. Emotional impact on neurological characteristics and human speech (Springer, 2014), pp. 527\u2013533.","DOI":"10.1007\/978-3-319-07773-4_52"},{"issue":"1","key":"157_CR2","doi-asserted-by":"publisher","first-page":"5","DOI":"10.1016\/S0167-6393(02)00071-7","volume":"40","author":"R. Cowie","year":"2003","unstructured":"R. Cowie, R. R. Cornelius, Describing the emotional states that are expressed in speech. Speech Comm.40(1), 5\u201332 (2003).","journal-title":"Speech Comm."},{"issue":"2","key":"157_CR3","doi-asserted-by":"publisher","first-page":"143","DOI":"10.1037\/0033-2909.99.2.143","volume":"99","author":"K. R. Scherer","year":"1986","unstructured":"K. R. Scherer, Vocal affect expression: a review and a model for future research. Psychol. Bull.99(2), 143 (1986).","journal-title":"Psychol. Bull."},{"key":"157_CR4","doi-asserted-by":"crossref","unstructured":"L. Vidrascu, L. Devillers, Detection of real-life emotions in call centers. Proc. INTERSPEECH, 1841\u20131844 (2005).","DOI":"10.21437\/Interspeech.2005-582"},{"issue":"5","key":"157_CR5","doi-asserted-by":"publisher","first-page":"776","DOI":"10.1037\/0012-1649.40.5.776","volume":"40","author":"G. I. Roisman","year":"2004","unstructured":"G. I. Roisman, J. L. Tsai, K. -H. S. Chiang, The emotional integration of childhood experience: physiological, facial expressive, and self-reported emotional response during the adult attachment interview. Dev. Psychol.40(5), 776 (2004).","journal-title":"Dev. Psychol."},{"issue":"5","key":"157_CR6","doi-asserted-by":"publisher","first-page":"1203","DOI":"10.1109\/JPROC.2012.2236291","volume":"101","author":"S. Narayanan","year":"2013","unstructured":"S. Narayanan, P. G. Georgiou, Behavioral signal processing: deriving human behavioral informatics from speech and language. Proc. IEEE. 101(5), 1203\u20131233 (2013).","journal-title":"Proc. IEEE"},{"key":"157_CR7","doi-asserted-by":"crossref","unstructured":"A. Metallinou, S. Lee, S. Narayanan, Decision level combination of multiple modalities for recognition and analysis of emotional expression, (2010).","DOI":"10.1109\/ICASSP.2010.5494890"},{"issue":"1","key":"157_CR8","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1109\/TPAMI.2008.52","volume":"31","author":"Z. Zeng","year":"2009","unstructured":"Z. Zeng, M. Pantic, G. I. Roisman, T. S. Huang, A survey of affect recognition methods: audio, visual, and spontaneous expressions. IEEE Trans. Patt. Anal. Mach. Intell.31(1), 39\u201358 (2009).","journal-title":"IEEE Trans. Patt. Anal. Mach. Intell."},{"issue":"5","key":"157_CR9","doi-asserted-by":"publisher","first-page":"1057","DOI":"10.1109\/TASL.2010.2076804","volume":"19","author":"E. Mower","year":"2011","unstructured":"E. Mower, M. J. Mataric, S. Narayanan, A framework for automatic human emotion classification using emotion profiles. IEEE Trans. Audio Speech Lang. Process.19(5), 1057\u20131070 (2011).","journal-title":"IEEE Trans. Audio Speech Lang. Process."},{"key":"157_CR10","doi-asserted-by":"crossref","unstructured":"Z. Aldeneh, S. Khorram, D. Dimitriadis, E. M. Provost, in Proceedings of the 19 th ACM International Conference on Multimodal Interaction. Pooling acoustic and lexical features for the prediction of valence (ACM, 2017), pp. 68\u201372.","DOI":"10.1145\/3136755.3136760"},{"issue":"3-4","key":"157_CR11","doi-asserted-by":"publisher","first-page":"399","DOI":"10.1177\/002383099804100408","volume":"41","author":"D. Erickson","year":"1998","unstructured":"D. Erickson, O. Fujimura, B. Pardo, Articulatory correlates of prosodic control: Emotion and emphasis. Lang. Speech. 41(3-4), 399\u2013417 (1998).","journal-title":"Lang. Speech"},{"issue":"1","key":"157_CR12","doi-asserted-by":"publisher","first-page":"187","DOI":"10.1016\/j.specom.2004.09.003","volume":"44","author":"M. Nordstrand","year":"2004","unstructured":"M. Nordstrand, G. Svanfeldt, B. Granstr\u00f6m, D. House, Measurements of articulatory variation in expressive speech for a set of swedish vowels. Speech Comm.44(1), 187\u2013196 (2004).","journal-title":"Speech Comm."},{"key":"157_CR13","doi-asserted-by":"crossref","unstructured":"S. Lee, S. Yildirim, A. Kazemzadeh, S. Narayanan, An articulatory study of emotional speech production. Proc. INTERSPEECH, 497\u2013500 (2005).","DOI":"10.21437\/Interspeech.2005-325"},{"issue":"4","key":"157_CR14","doi-asserted-by":"publisher","first-page":"603","DOI":"10.1016\/S0167-6393(03)00099-2","volume":"41","author":"T. L. Nwe","year":"2003","unstructured":"T. L. Nwe, S. W. Foo, L. C. De Silva, Speech emotion recognition using hidden Markov models. Speech Commun.41(4), 603\u2013623 (2003).","journal-title":"Speech Commun."},{"issue":"2","key":"157_CR15","doi-asserted-by":"publisher","first-page":"293","DOI":"10.1109\/TSA.2004.838534","volume":"13","author":"C. M. Lee","year":"2005","unstructured":"C. M. Lee, S. S. Narayanan, Toward detecting emotions in spoken dialogs. IEEE Trans. Speech Audio Process. 13(2), 293\u2013303 (2005).","journal-title":"IEEE Trans. Speech Audio Process"},{"issue":"9","key":"157_CR16","doi-asserted-by":"publisher","first-page":"1062","DOI":"10.1016\/j.specom.2011.01.011","volume":"53","author":"B. Schuller","year":"2011","unstructured":"B. Schuller, A. Batliner, S. Steidl, D. Seppi, Recognising realistic emotions and affect in speech: state of the art and lessons learnt from the first challenge. Speech Commun.53(9), 1062\u20131087 (2011).","journal-title":"Speech Commun."},{"issue":"3","key":"157_CR17","doi-asserted-by":"publisher","first-page":"572","DOI":"10.1016\/j.patcog.2010.09.020","volume":"44","author":"M. El Ayadi","year":"2011","unstructured":"M. El Ayadi, M. S. Kamel, F. Karray, Survey on speech emotion recognition: features, classification schemes, and databases. Pattern Recog.44(3), 572\u2013587 (2011).","journal-title":"Pattern Recog."},{"issue":"1","key":"157_CR18","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s13636-014-0049-y","volume":"2015","author":"M. Shah","year":"2015","unstructured":"M. Shah, C. Chakrabarti, A. Spanias, Within and cross-corpus speech emotion recognition using latent topic model-based features. EURASIP J. Audio Speech Music Process. 2015(1), 1\u201317 (2015).","journal-title":"EURASIP J. Audio Speech Music Process"},{"key":"157_CR19","unstructured":"G. Zhou, J. H. Hansen, J. F. Kaiser, Proc. IEEE Int. Conf. Acoust. Speech Signal Process.1:, 549\u2013552 (1998)."},{"issue":"4B","key":"157_CR20","doi-asserted-by":"publisher","first-page":"1238","DOI":"10.1121\/1.1913238","volume":"52","author":"C. E. Williams","year":"2005","unstructured":"C. E. Williams, K. N. Stevens, Emotions and speech: some acoustical correlates. J. Acoust. Soc. Am.52(4B), 1238\u20131250 (2005).","journal-title":"J. Acoust. Soc. Am."},{"key":"157_CR21","doi-asserted-by":"crossref","unstructured":"B. Schuller, A. Batliner, D. Seppi, S. Steidl, T. Vogt, J. Wagner, L. Devillers, L. Vidrascu, N. Amir, L. Kessous, et al., The relevance of feature type for the automatic classification of emotional user states: low level descriptors and functionals. Proc. INTERSPEECH, 1\u20134 (2007).","DOI":"10.21437\/Interspeech.2007-612"},{"issue":"4","key":"157_CR22","doi-asserted-by":"publisher","first-page":"582","DOI":"10.1109\/TASL.2008.2009578","volume":"17","author":"C. Busso","year":"2009","unstructured":"C. Busso, S. Lee, S. Narayanan, Analysis of emotionally salient aspects of fundamental frequency for emotion detection. IEEE Tran. Audio Speech Lang. Process. 17(4), 582\u2013596 (2009).","journal-title":"IEEE Tran. Audio Speech Lang. Process"},{"key":"157_CR23","doi-asserted-by":"crossref","unstructured":"B. Schuller, S. Steidl, A. Batliner, The INTERSPEECH 2009 emotion challenge. Proc. INTERSPEECH, 312\u2013315 (2009).","DOI":"10.21437\/Interspeech.2009-103"},{"issue":"9","key":"157_CR24","doi-asserted-by":"publisher","first-page":"1162","DOI":"10.1016\/j.specom.2011.06.004","volume":"53","author":"C. -C. Lee","year":"2011","unstructured":"C. -C. Lee, E. Mower, C. Busso, S. Lee, S. Narayanan, Emotion recognition using a hierarchical binary decision tree approach. Speech Commun.53(9), 1162\u20131171 (2011).","journal-title":"Speech Commun."},{"key":"157_CR25","doi-asserted-by":"crossref","unstructured":"F. Eyben, M. Wollmer, B. Schuller, OpenEAR\u2014introducing the Munich open-source emotion and affect recognition toolkit. Int. Conf. Affect. Comput. Intell. Interact. Workshops, 1\u20136 (2009).","DOI":"10.1109\/ACII.2009.5349350"},{"issue":"4","key":"157_CR26","doi-asserted-by":"publisher","first-page":"290","DOI":"10.1007\/s005210070006","volume":"9","author":"J. Nicholson","year":"2000","unstructured":"J. Nicholson, K. Takahashi, R. Nakatsu, Emotion recognition in speech using neural networks. Neural Comput. Applic.9(4), 290\u2013296 (2000).","journal-title":"Neural Comput. Applic."},{"key":"157_CR27","doi-asserted-by":"crossref","unstructured":"O. -W. Kwon, K. Chan, J. Hao, T. -W. Lee, Emotion recognition by speech signals. Proc. INTERSPEECH (2003).","DOI":"10.21437\/Eurospeech.2003-80"},{"issue":"9","key":"157_CR28","doi-asserted-by":"publisher","first-page":"1162","DOI":"10.1016\/j.specom.2011.06.004","volume":"53","author":"C. -C. Lee","year":"2011","unstructured":"C. -C. Lee, E. Mower, C. Busso, S. Lee, S. Narayanan, Emotion recognition using a hierarchical binary decision tree approach. Speech Commun.53(9), 1162\u20131171 (2011).","journal-title":"Speech Commun."},{"issue":"11","key":"157_CR29","doi-asserted-by":"publisher","first-page":"2108","DOI":"10.1109\/TASLP.2016.2593944","volume":"24","author":"S. Parthasarathy","year":"2016","unstructured":"S. Parthasarathy, R. Cowie, C. Busso, Using agreement on direction of change to build rank-based emotion classifiers. IEEE\/ACM Trans. Audio Speech Lang. Process.24(11), 2108\u20132121 (2016).","journal-title":"IEEE\/ACM Trans. Audio Speech Lang. Process."},{"key":"157_CR30","doi-asserted-by":"crossref","unstructured":"Z. Huang, J. Epps, in Acoustics, Speech and Signal Processing (ICASSP), 2017 IEEE International Conference On. A PLLR and multi-stage staircase regression framework for speech-based emotion prediction (IEEE, 2017), pp. 5145\u20135149.","DOI":"10.1109\/ICASSP.2017.7953137"},{"key":"157_CR31","doi-asserted-by":"crossref","unstructured":"S. Parthasarathy, C. Busso, Jointly predicting arousal, valence and dominance with multi-task learning. INTERSPEECH Stockholm Sweden (2017).","DOI":"10.21437\/Interspeech.2017-1494"},{"key":"157_CR32","doi-asserted-by":"crossref","unstructured":"D. Le, Z. Aldeneh, E. M. Provost, Discretized Continuous Speech Emotion Recognition with Multi-Task Deep Recurrent Neural Network. In INTERSPEECH, 1108\u20131112 (2017).","DOI":"10.21437\/Interspeech.2017-94"},{"key":"157_CR33","doi-asserted-by":"crossref","unstructured":"S. Khorram, Z. Aldeneh, D. Dimitriadis, M. McInnis, E. M. Provost, Capturing long-term temporal dependencies with convolutional networks for continuous emotion recognition. Proc. Interspeech 2017, 1253\u20131257 (2017).","DOI":"10.21437\/Interspeech.2017-548"},{"key":"157_CR34","doi-asserted-by":"crossref","unstructured":"Z. Aldeneh, E. M. Provost, in Acoustics, Speech and Signal Processing (ICASSP), 2017 IEEE International Conference On. Using regional saliency for speech emotion recognition (IEEE, 2017), pp. 2741\u20132745.","DOI":"10.1109\/ICASSP.2017.7952655"},{"key":"157_CR35","doi-asserted-by":"crossref","unstructured":"S. Mirsamadi, E. Barsoum, C. Zhang, in Acoustics, Speech and Signal Processing (ICASSP), 2017 IEEE International Conference On. Automatic speech emotion recognition using recurrent neural networks with local attention (IEEE, 2017), pp. 2227\u20132231.","DOI":"10.1109\/ICASSP.2017.7952552"},{"key":"157_CR36","doi-asserted-by":"crossref","unstructured":"M. Shah, C. Chakrabarti, A. Spanias, A multi-modal approach to emotion recognition using undirected topic models, (2014).","DOI":"10.1109\/ISCAS.2014.6865245"},{"issue":"4","key":"157_CR37","doi-asserted-by":"publisher","first-page":"2162","DOI":"10.1121\/1.3455847","volume":"128","author":"P. K. Ghosh","year":"2010","unstructured":"P. K. Ghosh, S. Narayanan, A generalized smoothness criterion for acoustic-to-articulatory inversion. J. Acoust. Soc. Am.128(4), 2162\u20132172 (2010).","journal-title":"J. Acoust. Soc. Am."},{"key":"157_CR38","unstructured":"J. Kim, P. Ghosh, S. Lee, S. S. Narayanan, in Signal & Information Processing Association Annual Summit and Conference (APSIPA ASC), 2012 Asia-Pacific. A study of emotional information present in articulatory movements estimated using acoustic-to-articulatory inversion (IEEE, 2012), pp. 1\u20134."},{"issue":"4","key":"157_CR39","doi-asserted-by":"publisher","first-page":"251","DOI":"10.1121\/1.3634122","volume":"130","author":"P. K. Ghosh","year":"2011","unstructured":"P. K. Ghosh, S. Narayanan, Automatic speech recognition using articulatory features from subject-independent acoustic-to-articulatory inversion. J. Acoust. Soc. Am.130(4), 251\u2013257 (2011).","journal-title":"J. Acoust. Soc. Am."},{"key":"157_CR40","doi-asserted-by":"publisher","first-page":"173","DOI":"10.1016\/j.csl.2015.05.005","volume":"36","author":"L. Badino","year":"2016","unstructured":"L. Badino, C. Canevari, L. Fadiga, G. Metta, Integrating articulatory data in deep neural network-based acoustic modeling. Comput. Speech Lang.36:, 173\u2013195 (2016).","journal-title":"Comput. Speech Lang."},{"key":"157_CR41","doi-asserted-by":"publisher","first-page":"196","DOI":"10.1016\/j.csl.2015.05.003","volume":"36","author":"M. Li","year":"2016","unstructured":"M. Li, J. Kim, A. Lammert, P. K. Ghosh, V. Ramanarayanan, S. Narayanan, Speaker verification based on the fusion of speech acoustics and inverted articulatory signals. Comput. Speech Lang.36:, 196\u2013211 (2016).","journal-title":"Comput. Speech Lang."},{"key":"157_CR42","unstructured":"D. Ververidis, C. Kotropoulos, in Proc. Panhellenic Conference on Informatics (PCI). A review of emotional speech databases, (2003), pp. 560\u2013574."},{"key":"157_CR43","doi-asserted-by":"crossref","unstructured":"J. Deng, Z. Zhang, E. Marchi, B. Schuller, in Affective Computing and Intelligent Interaction (ACII), 2013 Humaine Association Conference On. Sparse autoencoder-based feature transfer learning for speech emotion recognition (IEEE, 2013), pp. 511\u2013516.","DOI":"10.1109\/ACII.2013.90"},{"issue":"9","key":"157_CR44","doi-asserted-by":"publisher","first-page":"2530","DOI":"10.1587\/transinf.2014EDL8038","volume":"97","author":"P. Song","year":"2014","unstructured":"P. Song, Y. Jin, L. Zhao, M. Xin, Speech emotion recognition using transfer learning. IEICE Trans. Inf. Syst.97(9), 2530\u20132532 (2014).","journal-title":"IEICE Trans. Inf. Syst."},{"key":"157_CR45","doi-asserted-by":"crossref","unstructured":"J. Gideon, S. Khorram, Z. Aldeneh, D. Dimitriadis, E. M. Provost, Progressive neural networks for transfer learning in emotion recognition. arXiv preprint arXiv:1706.03256 (2017).","DOI":"10.21437\/Interspeech.2017-1637"},{"key":"157_CR46","doi-asserted-by":"crossref","unstructured":"M. Tu, V. Berisha, J. Liss, Interpretable objective assessment of dysarthric speech based on deep neural networks. Proc. Interspeech 2017, 1849\u20131853 (2017).","DOI":"10.21437\/Interspeech.2017-1222"},{"issue":"4","key":"157_CR47","doi-asserted-by":"publisher","first-page":"335","DOI":"10.1007\/s10579-008-9076-6","volume":"42","author":"C. Busso","year":"2008","unstructured":"C. Busso, M. Bulut, C. -C. Lee, A. Kazemzadeh, E. Mower, S. Kim, J. N. Chang, S. Lee, S. S. Narayanan, IEMOCAP: Interactive emotional dyadic motion capture database. Lang. Resour. Eval.42(4), 335\u2013359 (2008).","journal-title":"Lang. Resour. Eval."},{"key":"157_CR48","doi-asserted-by":"crossref","unstructured":"C. M. Lee, S. Yildirim, M. Bulut, A. Kazemzadeh, C. Busso, Z. Deng, S. Lee, S. Narayanan, Emotion recognition based on phoneme classes. Proc. INTERSPEECH, 205\u2013211 (2004).","DOI":"10.21437\/Interspeech.2004-322"},{"issue":"4","key":"157_CR49","doi-asserted-by":"publisher","first-page":"644","DOI":"10.2307\/3094912","volume":"47","author":"S. G. Barsade","year":"2002","unstructured":"S. G. Barsade, The ripple effect: emotional contagion and its influence on group behavior. Adm. Sci. Q.47(4), 644\u2013675 (2002).","journal-title":"Adm. Sci. Q."},{"key":"157_CR50","unstructured":"A. Katsamanis, M. Black, P. G. Georgiou, L. Goldstein, S. Narayanan, Sailalign: Robust long speech-text alignment, (2011)."},{"key":"157_CR51","unstructured":"P. Boersma, Praat, a system for doing phonetics by computer. Glot international, 5 (2002)."},{"issue":"2","key":"157_CR52","doi-asserted-by":"publisher","first-page":"119","DOI":"10.1109\/T-AFFC.2010.8","volume":"1","author":"B. Schuller","year":"2010","unstructured":"B. Schuller, B. Vlasenko, F. Eyben, M. Wollmer, A. Stuhlsatz, A. Wendemuth, G. Rigoll, Cross-corpus acoustic emotion recognition: variances and strategies. IEEE Trans. Affect. Comput.1(2), 119\u2013131 (2010).","journal-title":"IEEE Trans. Affect. Comput."},{"key":"157_CR53","doi-asserted-by":"crossref","unstructured":"B. Schuller, B. Vlasenko, F. Eyben, G. Rigoll, A. Wendemuth, in Automatic Speech Recognition & Understanding, 2009. ASRU 2009. IEEE Workshop On. Acoustic emotion recognition: a benchmark comparison of performances (IEEE, 2009), pp. 552\u2013557.","DOI":"10.1109\/ASRU.2009.5372886"},{"key":"157_CR54","unstructured":"A. Y. Ng, Feature selection, l1 vs. l2 regularization, and rotational invariance, (2004)."},{"key":"157_CR55","unstructured":"I. Guyon, Feature Extraction: Foundations and Applications, vol. 207, (2006)."},{"key":"157_CR56","unstructured":"L. -J. Li, H. Su, L. Fei-Fei, E. P. Xing, Object bank: a high-level image representation for scene classification & semantic feature sparsification. Advances in Neural Information Processing Systems, 1378\u20131386 (2010)."},{"issue":"1","key":"157_CR57","first-page":"824","volume":"45","author":"J. Gao","year":"2007","unstructured":"J. Gao, G. Andrew, M. Johnson, K. Toutanova, A comparative study of parameter estimation methods for statistical natural language processing. Annu. Meet.-Assoc. Comput. Linguist.45(1), 824 (2007).","journal-title":"Annu. Meet.-Assoc. Comput. Linguist."},{"issue":"1","key":"157_CR58","first-page":"401","volume":"21","author":"S. -I. Lee","year":"2006","unstructured":"S. -I. Lee, H. Lee, P. Abbeel, A. Y. Ng, Efficient L1 regularized logistic regression. Proc.Natl. Conf. Artif. Intell.21(1), 401 (2006).","journal-title":"Proc.Natl. Conf. Artif. Intell."},{"key":"157_CR59","doi-asserted-by":"crossref","unstructured":"H. Lee, A. Battle, R. Raina, A. Y. Ng, Efficient sparse coding algorithms. Adv. Neural Inf. Process. Syst., 801\u2013808 (2006).","DOI":"10.7551\/mitpress\/7503.003.0105"},{"key":"157_CR60","first-page":"3183","volume":"11","author":"G. -X. Yuan","year":"2010","unstructured":"G. -X. Yuan, K. -W. Chang, C. -J. Hsieh, C. -J. Lin, A comparison of optimization methods and software for large-scale L1-regularized linear classification. J. Mach. Learn. Res.11:, 3183\u20133234 (2010).","journal-title":"J. Mach. Learn. Res."},{"key":"157_CR61","unstructured":"M. Grant, S. Boyd, Y. Ye, (2015) CVX: Matlab software for disciplined convex programming (2008). http:\/\/stanford.edu\/~boyd\/cvx ."},{"key":"157_CR62","unstructured":"F. Eyben, A. Batliner, B. Schuller, D. Seppi, S. Steidl, Cross-corpus classification of realistic emotions some pilot experiments, (2010)."},{"key":"157_CR63","doi-asserted-by":"crossref","unstructured":"D. Neiberg, P. Laukka, H. A. Elfenbein, Intra-, inter-, and cross-cultural classification of vocal affect. Proc. INTERSPEECH, 1581\u20131584 (2011).","DOI":"10.21437\/Interspeech.2011-475"},{"key":"157_CR64","unstructured":"B. Zhang, E. M. Provost, G. Essl, Cross-corpus acoustic emotion recognition with multi-task learning: seeking common ground while preserving differences. IEEE Trans. Affect. Comput. (2017)."},{"issue":"3","key":"157_CR65","doi-asserted-by":"publisher","first-page":"1411","DOI":"10.1121\/1.4908284","volume":"137","author":"J. Kim","year":"2015","unstructured":"J. Kim, A. Toutios, S. Lee, S. S. Narayanan, A kinematic study of critical and non-critical articulators in emotional speech production. J. Acoust. Soc. Am.137(3), 1411\u20131429 (2015).","journal-title":"J. Acoust. Soc. Am."}],"container-title":["EURASIP Journal on Audio, Speech, and Music Processing"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1186\/s13636-019-0157-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/article\/10.1186\/s13636-019-0157-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1186\/s13636-019-0157-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,9,19]],"date-time":"2023-09-19T08:12:19Z","timestamp":1695111139000},"score":1,"resource":{"primary":{"URL":"https:\/\/asmp-eurasipjournals.springeropen.com\/articles\/10.1186\/s13636-019-0157-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,8,20]]},"references-count":65,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2019,12]]}},"alternative-id":["157"],"URL":"https:\/\/doi.org\/10.1186\/s13636-019-0157-9","relation":{},"ISSN":["1687-4722"],"issn-type":[{"type":"electronic","value":"1687-4722"}],"subject":[],"published":{"date-parts":[[2019,8,20]]},"assertion":[{"value":"3 February 2019","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 July 2019","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 August 2019","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"All authors were employed by Arizona State University at the time this research was conducted. Berisha was also the Chief Science Officer at Aural Analytics.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"14"}}