{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,13]],"date-time":"2025-05-13T18:05:40Z","timestamp":1747159540449,"version":"3.40.3"},"publisher-location":"Cham","reference-count":19,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030007935"},{"type":"electronic","value":"9783030007942"}],"license":[{"start":{"date-parts":[[2018,1,1]],"date-time":"2018-01-01T00:00:00Z","timestamp":1514764800000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2018]]},"DOI":"10.1007\/978-3-030-00794-2_56","type":"book-chapter","created":{"date-parts":[[2018,9,7]],"date-time":"2018-09-07T19:50:24Z","timestamp":1536349824000},"page":"518-526","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Classification of Formal and Informal Dialogues Based on Emotion Recognition Features"],"prefix":"10.1007","author":[{"given":"Gy\u00f6rgy","family":"Kov\u00e1cs","sequence":"first","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2018,9,8]]},"reference":[{"key":"56_CR1","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"178","DOI":"10.1007\/978-3-540-24842-2_17","volume-title":"Affective Dialogue Systems","author":"E Andr\u00e9","year":"2004","unstructured":"Andr\u00e9, E., Rehm, M., Minker, W., B\u00fchler, D.: Endowing spoken language dialogue systems with emotional intelligence. In: Andr\u00e9, E., Dybkj\u00e6r, L., Minker, W., Heisterkamp, P. (eds.) ADS 2004. LNCS (LNAI), vol. 3068, pp. 178\u2013187. Springer, Heidelberg (2004). https:\/\/doi.org\/10.1007\/978-3-540-24842-2_17"},{"key":"56_CR2","first-page":"549","volume":"13","author":"D Benbouzid","year":"2012","unstructured":"Benbouzid, D., Busa-Fekete, R., Casagrande, N., Collin, F.D., K\u00e9gl, B.: MULTIBOOST: a multi-purpose boosting package. J. Mach. Learn. Res. 13, 549\u2013553 (2012)","journal-title":"J. Mach. Learn. Res."},{"key":"56_CR3","unstructured":"Bradley, J., Schapire, R.: FilterBoost: regression and classification on large datasets. In: Advances in Neural Information Processing Systems, vol. 20, pp. 185\u2013192. The MIT Press (2008)"},{"key":"56_CR4","doi-asserted-by":"publisher","first-page":"27:1","DOI":"10.1145\/1961189.1961199","volume":"2","author":"CC Chang","year":"2011","unstructured":"Chang, C.C., Lin, C.J.: LIBSVM: a library for support vector machines. ACM Trans. Intell. Syst. Technol. 2, 27:1\u201327:27 (2011)","journal-title":"ACM Trans. Intell. Syst. Technol."},{"key":"56_CR5","unstructured":"Eyben, F., W\u00f6llmer, M., Schuller, B.: The Munich open speech and music interpretation by large space extraction toolkit (2010)"},{"key":"56_CR6","doi-asserted-by":"crossref","unstructured":"Eyben, F., W\u00f6llmer, M., Schuller, B.: openSMILE: the Munich versatile and fast open-source audio feature extractor. In: Proceedings of ACM (MM), pp. 1459\u20131462 (2010)","DOI":"10.1145\/1873951.1874246"},{"key":"56_CR7","doi-asserted-by":"publisher","first-page":"119","DOI":"10.1006\/jcss.1997.1504","volume":"55","author":"Y Freund","year":"1997","unstructured":"Freund, Y., Schapire, R.E.: A decision-theoretic generalization of on-line learning and an application to boosting. J. Comput. Syst. Sci 55, 119\u2013139 (1997)","journal-title":"J. Comput. Syst. Sci"},{"key":"56_CR8","doi-asserted-by":"publisher","first-page":"129","DOI":"10.1007\/3-540-45164-1_14","volume-title":"Machine Learning: ECML 2000","author":"Gerard Escudero","year":"2000","unstructured":"G. Escudero, L.M., Rigau, G.: Boosting applied to word sense disambiguation. In: Proceedings of ECML, pp. 129\u2013141 (2000)"},{"key":"56_CR9","unstructured":"Hunyadi, L.: Multimodal human-computer interaction technologies. Theoretical modeling and application in speech processing. Argumentum, pp. 240\u2013260 (2011)"},{"key":"56_CR10","unstructured":"Hunyadi, L., V\u00e1radi, T., Szekr\u00e9nyes, I.: Language technology tools and resources for the analysis of multimodal communication. In: Proceedings of LT4DH, pp. 117\u2013124. University of T\u00fcbingen, T\u00fcbingen (2016)"},{"key":"56_CR11","doi-asserted-by":"publisher","DOI":"10.1017\/CBO9780511618963","volume-title":"Neurolinguistics","author":"JCL Ingram","year":"2007","unstructured":"Ingram, J.C.L.: Neurolinguistics. Cambridge University Press, Cambridge (2007)"},{"key":"56_CR12","doi-asserted-by":"publisher","first-page":"265","DOI":"10.4135\/9781446200957.n20","volume-title":"The SAGE Handbook of Sociolinguistics","author":"T Kristiansen","year":"2011","unstructured":"Kristiansen, T.: Attitudes, ideology and awareness. In: Wodak, R., Johnstone, B., Kerswill, P. (eds.) The SAGE Handbook of Sociolinguistics, pp. 265\u2013278. SAGE Publishing, Thousand Oaks (2011)"},{"key":"56_CR13","volume-title":"The Social Stratification of English in New York City","author":"W Labov","year":"1996","unstructured":"Labov, W.: The Social Stratification of English in New York City. Cambridge University Press, Cambridge (1996)"},{"key":"56_CR14","first-page":"330","volume":"7","author":"K P\u00e1pay","year":"2011","unstructured":"P\u00e1pay, K., Szeghalmy, S., Szekr\u00e9nyes, I.: HuComTech multimodal corpus annotation. Argumentum 7, 330\u2013347 (2011)","journal-title":"Argumentum"},{"key":"56_CR15","doi-asserted-by":"crossref","unstructured":"Schuller, B., Steidl, S., Batliner, A.: The INTERSPEECH 2009 emotion challenge. In: Proceedings of INTERSPEECH, pp. 312\u2013315 (2009)","DOI":"10.21437\/Interspeech.2009-103"},{"key":"56_CR16","doi-asserted-by":"crossref","unstructured":"Schuller, B., et al.: The INTERSPEECH 2010 paralinguistic challenge. In: Proceedings of INTERSPEECH, pp. 2822\u20132825 (2010)","DOI":"10.21437\/Interspeech.2010-739"},{"key":"56_CR17","doi-asserted-by":"publisher","first-page":"17","DOI":"10.1007\/s12193-013-0129-9","volume":"8","author":"I Siegert","year":"2014","unstructured":"Siegert, I., B\u00f6ck, R., Wendmeuth, A.: Inter-rater reliability for emotion annotation in human-computer interaction: comparison and methodological improvements. Multimodal User Interfaces 8, 17\u201328 (2014)","journal-title":"Multimodal User Interfaces"},{"key":"56_CR18","doi-asserted-by":"crossref","unstructured":"Szekr\u00e9nyes, I.: ProsoTool, a method for automatic annotation of fundamental frequency. In: Proceedings of CogInfoCom, pp. 291\u2013296 (2015)","DOI":"10.1109\/CogInfoCom.2015.7390606"},{"key":"56_CR19","doi-asserted-by":"publisher","first-page":"233","DOI":"10.1007\/978-3-319-66429-3_22","volume-title":"Speech and Computer","author":"Istv\u00e1n Szekr\u00e9nyes","year":"2017","unstructured":"Szekr\u00e9nyes, I., Kov\u00e1cs, G.: Classification of formal and informal dialogues based on turn-taking and intonation using deep neural networks. In: Proceedings of SPECOM, pp. 233\u2013243 (2017)"}],"container-title":["Lecture Notes in Computer Science","Text, Speech, and Dialogue"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-00794-2_56","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,7]],"date-time":"2024-03-07T16:39:21Z","timestamp":1709829561000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-00794-2_56"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018]]},"ISBN":["9783030007935","9783030007942"],"references-count":19,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-00794-2_56","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2018]]},"assertion":[{"value":"8 September 2018","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}}]}}