{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,16]],"date-time":"2026-04-16T02:57:11Z","timestamp":1776308231967,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":41,"publisher":"ACM","license":[{"start":{"date-parts":[[2016,10,31]],"date-time":"2016-10-31T00:00:00Z","timestamp":1477872000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2016,10,31]]},"DOI":"10.1145\/2993148.2993151","type":"proceedings-article","created":{"date-parts":[[2016,11,1]],"date-time":"2016-11-01T13:46:03Z","timestamp":1478007963000},"page":"92-99","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":24,"title":["Emotion spotting: discovering regions of evidence in audio-visual emotion expressions"],"prefix":"10.1145","author":[{"given":"Yelin","family":"Kim","sequence":"first","affiliation":[{"name":"SUNY Albany, USA"}]},{"given":"Emily Mower","family":"Provost","sequence":"additional","affiliation":[{"name":"University of Michigan, USA"}]}],"member":"320","published-online":{"date-parts":[[2016,10,31]]},"reference":[{"key":"e_1_3_2_1_1_1","first-page":"563","volume-title":"IEEE Winter Conference on Applications of Computer Vision","author":"Amer M. R.","unstructured":"M. R. Amer , B. Siddiquie , S. Khan , A. Divakaran , and H. Sawhney . Multimodal fusion using dynamic hybrid models . In IEEE Winter Conference on Applications of Computer Vision , pages 556\u2013 563 . IEEE, 2014. M. R. Amer, B. Siddiquie, S. Khan, A. Divakaran, and H. Sawhney. Multimodal fusion using dynamic hybrid models. In IEEE Winter Conference on Applications of Computer Vision, pages 556\u2013563. IEEE, 2014."},{"key":"e_1_3_2_1_2_1","unstructured":"P. Boersma and D. Weenink. Praat: doing phonetics by computer (version 6.0.17){computer program}. retrieved 21 april 2016 from http:\/\/www.praat.org\/.  P. Boersma and D. Weenink. Praat: doing phonetics by computer (version 6.0.17){computer program}. retrieved 21 april 2016 from http:\/\/www.praat.org\/."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10579-008-9076-6"},{"key":"e_1_3_2_1_4_1","first-page":"2228","volume-title":"Interspeech","author":"Busso C.","year":"2007","unstructured":"C. Busso , S. Lee , and S. S. Narayanan . Using neutral speech models for emotional speech analysis . In Interspeech , pages 2225\u2013 2228 , 2007 . C. Busso, S. Lee, and S. S. Narayanan. Using neutral speech models for emotional speech analysis. In Interspeech, pages 2225\u20132228, 2007."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/TASL.2007.905145"},{"key":"e_1_3_2_1_6_1","article-title":"An acted corpus of dyadic interactions to study emotion perception","author":"Busso C.","year":"2015","unstructured":"C. Busso , S. Parthasarathy , A. Burmania , M. AbdelWahab , N. Sadoughi , and E. Mower Provost . MSP-IMPROV : An acted corpus of dyadic interactions to study emotion perception . IEEE Transactions on Affective Computing , 2015 . C. Busso, S. Parthasarathy, A. Burmania, M. AbdelWahab, N. Sadoughi, and E. Mower Provost. MSP-IMPROV: An acted corpus of dyadic interactions to study emotion perception. IEEE Transactions on Affective Computing, 2015.","journal-title":"IEEE Transactions on Affective Computing"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/1961189.1961199"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/79.911197"},{"key":"e_1_3_2_1_9_1","volume-title":"Twelfth Annual Conference of the International Speech Communication Association","author":"Cvejic E.","year":"2011","unstructured":"E. Cvejic , J. Kim , and C. Davis . Temporal relationship between auditory and visual prosodic cues . In Twelfth Annual Conference of the International Speech Communication Association , 2011 . E. Cvejic, J. Kim, and C. Davis. Temporal relationship between auditory and visual prosodic cues. In Twelfth Annual Conference of the International Speech Communication Association, 2011."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1145\/2818346.2830596"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1111\/1467-9280.00054"},{"key":"e_1_3_2_1_12_1","volume-title":"Freitas-Magalh\u02dc aes. Microexpression and macroexpression. Encyclopedia of human behavior, 2:173\u2013183","author":"A.","year":"2012","unstructured":"A. Freitas-Magalh\u02dc aes. Microexpression and macroexpression. Encyclopedia of human behavior, 2:173\u2013183 , 2012 . A. Freitas-Magalh\u02dc aes. Microexpression and macroexpression. Encyclopedia of human behavior, 2:173\u2013183, 2012."},{"key":"e_1_3_2_1_13_1","volume-title":"Springer","author":"Haggard E. A.","year":"1966","unstructured":"E. A. Haggard and K. S. Isaacs . Micromomentary facial expressions as indicators of ego mechanisms in psychotherapy. In Methods of research in psychotherapy, pages 154\u2013165 . Springer , 1966 . E. A. Haggard and K. S. Isaacs. Micromomentary facial expressions as indicators of ego mechanisms in psychotherapy. In Methods of research in psychotherapy, pages 154\u2013165. Springer, 1966."},{"key":"e_1_3_2_1_14_1","volume-title":"Handbook of human-computer interaction","author":"Helander M. G.","year":"2014","unstructured":"M. G. Helander . Handbook of human-computer interaction . Elsevier , 2014 . M. G. Helander. Handbook of human-computer interaction. Elsevier, 2014."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/CW.2005.82"},{"key":"e_1_3_2_1_16_1","volume-title":"Evolving Technologies, and Emerging Applications","author":"Jacko J. A.","year":"2012","unstructured":"J. A. Jacko . Human Computer Interaction Handbook : Fundamentals , Evolving Technologies, and Emerging Applications . CRC press , 2012 . J. A. Jacko. Human Computer Interaction Handbook: Fundamentals, Evolving Technologies, and Emerging Applications. CRC press, 2012."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/2808196.2811637"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/2647868.2654934"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/2808204"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/FG.2011.5771414"},{"key":"e_1_3_2_1_21_1","first-page":"1229","volume-title":"European Signal Processing Conference","author":"Lugger M.","unstructured":"M. Lugger , M.-E. Janoir , and B. Yang . Combining classifiers with diverse feature sets for robust speaker independent emotion recognition . In European Signal Processing Conference , pages 1225\u2013 1229 . IEEE, 2009. M. Lugger, M.-E. Janoir, and B. Yang. Combining classifiers with diverse feature sets for robust speaker independent emotion recognition. In European Signal Processing Conference, pages 1225\u20131229. IEEE, 2009."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-009-0344-2"},{"key":"e_1_3_2_1_23_1","volume-title":"IEEE International Conference on Automatic Face and Gesture Recognition (FG).","author":"Mariooryad S.","unstructured":"S. Mariooryad and C. Busso . Feature and model level compensation of lexical content for facial emotion recognition . In IEEE International Conference on Automatic Face and Gesture Recognition (FG). S. Mariooryad and C. Busso. Feature and model level compensation of lexical content for facial emotion recognition. In IEEE International Conference on Automatic Face and Gesture Recognition (FG)."},{"key":"e_1_3_2_1_24_1","unstructured":"A. Metallinou C. Busso S. Lee and S. Narayanan. Visual emotion recognition using compact facial representations and viseme information. In ICASSP.  A. Metallinou C. Busso S. Lee and S. Narayanan. Visual emotion recognition using compact facial representations and viseme information. In ICASSP."},{"key":"e_1_3_2_1_25_1","first-page":"2465","volume-title":"International Conference on Acoustics Speech and Signal Processing","author":"Metallinou A.","unstructured":"A. Metallinou , S. Lee , and S. Narayanan . Decision level combination of multiple modalities for recognition and analysis of emotional expression . In International Conference on Acoustics Speech and Signal Processing , pages 2462\u2013 2465 . IEEE, 2010. A. Metallinou, S. Lee, and S. Narayanan. Decision level combination of multiple modalities for recognition and analysis of emotional expression. In International Conference on Acoustics Speech and Signal Processing, pages 2462\u20132465. IEEE, 2010."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/T-AFFC.2011.40"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1109\/TASL.2010.2076804"},{"key":"e_1_3_2_1_28_1","first-page":"8","volume-title":"International Conference on Affective Computing and Intelligent Interaction and Workshops","author":"Mower E.","unstructured":"E. Mower , A. Metallinou , C.-C. Lee , A. Kazemzadeh , C. Busso , S. Lee , and S. Narayanan . Interpreting ambiguous emotional expressions . In International Conference on Affective Computing and Intelligent Interaction and Workshops , pages 1\u2013 8 . IEEE, 2009. E. Mower, A. Metallinou, C.-C. Lee, A. Kazemzadeh, C. Busso, S. Lee, and S. Narayanan. Interpreting ambiguous emotional expressions. In International Conference on Affective Computing and Intelligent Interaction and Workshops, pages 1\u20138. IEEE, 2009."},{"key":"e_1_3_2_1_29_1","first-page":"2375","volume-title":"ICASSP","author":"Mower E.","unstructured":"E. Mower and S. Narayanan . A hierarchical static-dynamic framework for emotion classification . In ICASSP , pages 2372\u2013 2375 . IEEE, 2011. E. Mower and S. Narayanan. A hierarchical static-dynamic framework for emotion classification. In ICASSP, pages 2372\u20132375. IEEE, 2011."},{"key":"e_1_3_2_1_30_1","first-page":"3686","volume-title":"Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on","author":"Provost E. Mower","unstructured":"E. Mower Provost . Identifying salient sub-utterance emotion dynamics using flexible units and estimates of affective flow. In Acoustics , Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on , pages 3682\u2013 3686 . IEEE, 2013. E. Mower Provost. Identifying salient sub-utterance emotion dynamics using flexible units and estimates of affective flow. In Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on, pages 3682\u20133686. IEEE, 2013."},{"key":"e_1_3_2_1_31_1","unstructured":"A. V. Nefian L. Liang X. Pi L. Xiaoxiang C. Mao and K. Murphy. A coupled hmm for audio-visual speech recognition.  A. V. Nefian L. Liang X. Pi L. Xiaoxiang C. Mao and K. Murphy. A coupled hmm for audio-visual speech recognition."},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1109\/JPROC.2003.817122"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0027256"},{"key":"e_1_3_2_1_34_1","first-page":"6","volume-title":"International Conference on Crime Detection and Prevention (ICDP)","author":"Polikovsky S.","unstructured":"S. Polikovsky , Y. Kameda , and Y. Ohta . Facial micro-expressions recognition using high speed camera and 3d-gradient descriptor . In International Conference on Crime Detection and Prevention (ICDP) , pages 1\u2013 6 . IET, 2009. S. Polikovsky, Y. Kameda, and Y. Ohta. Facial micro-expressions recognition using high speed camera and 3d-gradient descriptor. In International Conference on Crime Detection and Prevention (ICDP), pages 1\u20136. IET, 2009."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.5555\/2440186"},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW.2013.70"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2011.11.008"},{"key":"e_1_3_2_1_38_1","first-page":"2256","volume-title":"INTERSPEECH","author":"Schuller B.","unstructured":"B. Schuller , A. Batliner , D. Seppi , S. Steidl , T. Vogt , J. Wagner , L. Devillers , L. Vidrascu , N. Amir , L. Kessous , The relevance of feature type for the automatic classification of emotional user states: low level descriptors and functionals . In INTERSPEECH , pages 2253\u2013 2256 . B. Schuller, A. Batliner, D. Seppi, S. Steidl, T. Vogt, J. Wagner, L. Devillers, L. Vidrascu, N. Amir, L. Kessous, et al. The relevance of feature type for the automatic classification of emotional user states: low level descriptors and functionals. In INTERSPEECH, pages 2253\u20132256."},{"key":"e_1_3_2_1_39_1","unstructured":"Citeseer 2007.  Citeseer 2007."},{"key":"e_1_3_2_1_40_1","first-page":"320","volume-title":"ICASSP","author":"Steidl S.","unstructured":"S. Steidl , M. Levit , A. Batliner , E. N\u00f6th , and H. Niemann . \u201c of all things the measure is man\u201d: Automatic classification of emotions and inter-labeler consistency . In ICASSP , pages 317\u2013 320 . Citeseer, 2005. S. Steidl, M. Levit, A. Batliner, E. N\u00f6th, and H. Niemann. \u201cof all things the measure is man\u201d: Automatic classification of emotions and inter-labeler consistency. In ICASSP, pages 317\u2013320. Citeseer, 2005."},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1109\/T-AFFC.2011.12"}],"event":{"name":"ICMI '16: INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION","location":"Tokyo Japan","acronym":"ICMI '16","sponsor":["SIGCHI ACM Special Interest Group on Computer-Human Interaction"]},"container-title":["Proceedings of the 18th ACM International Conference on Multimodal Interaction"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/2993148.2993151","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/2993148.2993151","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T03:50:02Z","timestamp":1750218602000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/2993148.2993151"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2016,10,31]]},"references-count":41,"alternative-id":["10.1145\/2993148.2993151","10.1145\/2993148"],"URL":"https:\/\/doi.org\/10.1145\/2993148.2993151","relation":{},"subject":[],"published":{"date-parts":[[2016,10,31]]},"assertion":[{"value":"2016-10-31","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}