{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,9]],"date-time":"2026-04-09T00:35:33Z","timestamp":1775694933622,"version":"3.50.1"},"reference-count":66,"publisher":"Springer Science and Business Media LLC","issue":"6-7","license":[{"start":{"date-parts":[[2019,2,13]],"date-time":"2019-02-13T00:00:00Z","timestamp":1550016000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"name":"Finland Distinguished Professor Programme of Tekes","award":["1849\/31\/2015"],"award-info":[{"award-number":["1849\/31\/2015"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Comput Vis"],"published-print":{"date-parts":[[2019,6]]},"DOI":"10.1007\/s11263-019-01158-4","type":"journal-article","created":{"date-parts":[[2019,2,13]],"date-time":"2019-02-13T05:22:35Z","timestamp":1550035355000},"page":"907-929","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":257,"title":["Deep Affect Prediction in-the-Wild: Aff-Wild Database and Challenge, Deep Architectures, and Beyond"],"prefix":"10.1007","volume":"127","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8188-3751","authenticated-orcid":false,"given":"Dimitrios","family":"Kollias","sequence":"first","affiliation":[]},{"given":"Panagiotis","family":"Tzirakis","sequence":"additional","affiliation":[]},{"given":"Mihalis A.","family":"Nicolaou","sequence":"additional","affiliation":[]},{"given":"Athanasios","family":"Papaioannou","sequence":"additional","affiliation":[]},{"given":"Guoying","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Bj\u00f6rn","family":"Schuller","sequence":"additional","affiliation":[]},{"given":"Irene","family":"Kotsia","sequence":"additional","affiliation":[]},{"given":"Stefanos","family":"Zafeiriou","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2019,2,13]]},"reference":[{"key":"1158_CR1","first-page":"265","volume":"16","author":"M Abadi","year":"2016","unstructured":"Abadi, M., Barham, P., Chen, J., Chen, Z., Davis, A., Dean, J., et al. (2016). Tensorflow: A system for large-scale machine learning. OSDI, 16, 265\u2013283.","journal-title":"OSDI"},{"key":"1158_CR2","doi-asserted-by":"crossref","unstructured":"Alabort-i-Medina, J., Antonakos, E., Booth, J., Snape, P., & Zafeiriou, S. (2014). Menpo: A comprehensive platform for parametric image alignment and visual deformable models. In Proceedings of the ACM international conference on multimedia, MM \u201914 (pp. 679\u2013682). ACM, New York, NY.","DOI":"10.1145\/2647868.2654890"},{"key":"1158_CR3","unstructured":"Albanie, S., & Vedaldi, A. (2016). Learning grimaces by watching TV. In: Proceedings of the British Machine Vision Conference (BMVC)."},{"key":"1158_CR4","first-page":"1","volume":"1","author":"MS Aung","year":"2016","unstructured":"Aung, M. S., Kaltwang, S., Romera-paredes, B., Martinez, B., Singh, A., Cella, M., et al. (2016). The automatic detection of chronic pain-related expression: Requirements, challenges and a multimodal dataset. IEEE Transactions on Affective Computing, 1, 1.","journal-title":"IEEE Transactions on Affective Computing"},{"key":"1158_CR5","unstructured":"Bartlett, M. S., Littlewort, G., Frank, M., Lainscsek, C., Fasel, I., & Movellan, J. (2006). Fully automatic facial action recognition in spontaneous behavior. In: 7th international conference on automatic face and gesture recognition, 2006, FGR 2006 (pp. 223\u2013230). IEEE."},{"key":"1158_CR6","unstructured":"Chang, W. Y., Hsu, S. H., & Chien, J. H. (2017). Fatauva-net : An integrated deep learning framework for facial attribute recognition, action unit (AU) detection, and valence\u2013arousal estimation. In: Proceedings of the IEEE conference on computer vision and pattern recognition workshop."},{"key":"1158_CR7","unstructured":"Chatfield, K., Simonyan, K., Vedaldi, A., & Zisserman, A. (2014). Return of the devil in the details: Delving deep into convolutional nets. arXiv preprint \n                    arXiv:1405.3531\n                    \n                  ."},{"issue":"2\u20134","key":"1158_CR8","doi-asserted-by":"publisher","first-page":"198","DOI":"10.1007\/s11263-017-0999-5","volume":"126","author":"GG Chrysos","year":"2018","unstructured":"Chrysos, G. G., Antonakos, E., Snape, P., Asthana, A., & Zafeiriou, S. (2018). A comprehensive performance evaluation of deformable face tracking in-the-wild. International Journal of Computer Vision, 126(2\u20134), 198\u2013232.","journal-title":"International Journal of Computer Vision"},{"key":"1158_CR9","unstructured":"Chung, J., Gulcehre, C., Cho, K., & Bengio, Y. (2014). Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint \n                    arXiv:1412.3555\n                    \n                  ."},{"key":"1158_CR10","doi-asserted-by":"publisher","first-page":"1548","DOI":"10.1109\/TPAMI.2016.2515606","volume":"38","author":"C Corneanu","year":"2016","unstructured":"Corneanu, C., Oliu, M., Cohn, J., & Escalera, S. (2016). Survey on rgb, 3d, thermal, and multimodal approaches for facial expression recognition: History, trends, and affect-related applications. IEEE Transactions on Pattern Analysis and Machine Intelligence, 38, 1548\u20131568.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"issue":"1","key":"1158_CR11","doi-asserted-by":"publisher","first-page":"5","DOI":"10.1016\/S0167-6393(02)00071-7","volume":"40","author":"R Cowie","year":"2003","unstructured":"Cowie, R., & Cornelius, R. R. (2003). Describing the emotional states that are expressed in speech. Speech Communication, 40(1), 5\u201332.","journal-title":"Speech Communication"},{"key":"1158_CR12","unstructured":"Cowie, R., Douglas-Cowie, E., Savvidou, S., McMahon, E., Sawey, M., & Schr\u00f6der, M. (2000). \u2018feeltrace\u2019: An instrument for recording perceived emotion in real time. In ISCA tutorial and research workshop (ITRW) on speech and emotion."},{"issue":"1","key":"1158_CR13","doi-asserted-by":"publisher","first-page":"1","DOI":"10.4018\/jse.2012010101","volume":"3","author":"R Cowie","year":"2012","unstructured":"Cowie, R., McKeown, G., & Douglas-Cowie, E. (2012). Tracing emotion: An overview. International Journal of Synthetic Emotions (IJSE), 3(1), 1\u201317.","journal-title":"International Journal of Synthetic Emotions (IJSE)"},{"key":"1158_CR14","volume-title":"Handbook of cognition and emotion","author":"T Dalgleish","year":"2000","unstructured":"Dalgleish, T., & Power, M. (2000). Handbook of cognition and emotion. Hoboken: Wiley."},{"key":"1158_CR15","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L. J., Li, K., & Fei-Fei, L. (2009). Imagenet: A large-scale hierarchical image database. In IEEE conference on computer vision and pattern recognition, 2009. CVPR 2009 (pp. 248\u2013255). IEEE.","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"1158_CR16","doi-asserted-by":"crossref","unstructured":"Dhall, A., Goecke, R., Ghosh, S., Joshi, J., Hoey, J., & Gedeon, T. (2017). From individual to group-level emotion recognition: Emotiw 5.0. In Proceedings of the 19th ACM international conference on multimodal interaction (pp. 524\u2013528). ACM.","DOI":"10.1145\/3136755.3143004"},{"key":"1158_CR17","doi-asserted-by":"crossref","unstructured":"Dhall, A., Goecke, R., Joshi, J., Hoey, J., & Gedeon, T. (2016). Emotiw 2016: Video and group-level emotion recognition challenges. In Proceedings of the 18th ACM international conference on multimodal interaction (pp. 427\u2013432). ACM.","DOI":"10.1145\/2993148.2997638"},{"key":"1158_CR18","doi-asserted-by":"crossref","unstructured":"Dhall, A., Goecke, R., Joshi, J., Sikka, K., & Gedeon, T. (2014). Emotion recognition in the wild challenge 2014: Baseline, data and protocol. In Proceedings of the 16th international conference on multimodal interaction (pp. 461\u2013466). ACM.","DOI":"10.1145\/2663204.2666275"},{"key":"1158_CR19","doi-asserted-by":"crossref","unstructured":"Dhall, A., Goecke, R., Joshi, J., Wagner, M., & Gedeon, T. (2013). Emotion recognition in the wild challenge 2013. In Proceedings of the 15th ACM on international conference on multimodal interaction (pp. 509\u2013516). ACM.","DOI":"10.1145\/2522848.2531739"},{"key":"1158_CR20","unstructured":"Dhall, A., Ramana Murthy, O., Goecke, R., Joshi, J., & Gedeon, T. (2015). Video and image based emotion recognition challenges in the wild: Emotiw 2015. In: Proceedings of the 2015 ACM on international conference on multimodal interaction (pp. 423\u2013426). ACM."},{"key":"1158_CR21","unstructured":"Douglas-Cowie, E., Cowie, R., Cox, C., Amier, N., & Heylen, D. K. (2008). The sensitive artificial listner: An induction technique for generating emotionally coloured conversation. In LREC workshop on corpora for research on emotion and affect. ELRA."},{"issue":"5","key":"1158_CR22","doi-asserted-by":"publisher","first-page":"807","DOI":"10.1016\/j.imavis.2009.08.002","volume":"28","author":"R Gross","year":"2010","unstructured":"Gross, R., Matthews, I., Cohn, J., Kanade, T., & Baker, S. (2010). Multi-pie. Image and Vision Computing, 28(5), 807\u2013813.","journal-title":"Image and Vision Computing"},{"key":"1158_CR23","unstructured":"Hardoon, D.R., Szedmak, S., & Shawe-Taylor, J. (2003). Canonical correlation analysis; An overview with application to learning methods. Technical report, Royal Holloway, University of London."},{"key":"1158_CR24","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 770\u2013778).","DOI":"10.1109\/CVPR.2016.90"},{"issue":"8","key":"1158_CR25","doi-asserted-by":"publisher","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","volume":"9","author":"S Hochreiter","year":"1997","unstructured":"Hochreiter, S., & Schmidhuber, J. (1997). Long short-term memory. Neural Computation, 9(8), 1735\u20131780.","journal-title":"Neural Computation"},{"key":"1158_CR26","doi-asserted-by":"crossref","unstructured":"Hu, P., Cai, D., Wang, S., Yao, A., & Chen, Y. (2017). Learning supervised scoring ensemble for emotion recognition in the wild. In Proceedings of the 19th ACM international conference on multimodal interaction (pp. 553\u2013560). ACM.","DOI":"10.1145\/3136755.3143009"},{"key":"1158_CR27","doi-asserted-by":"crossref","unstructured":"Jung, H., Lee, S., Yim, J., Park, S., & Kim, J. (2015). Joint fine-tuning in deep neural networks for facial expression recognition. In Proceedings of the IEEE international conference on computer vision (pp. 2983\u20132991).","DOI":"10.1109\/ICCV.2015.341"},{"key":"1158_CR28","unstructured":"Knyazev, B., Shvetsov, R., Efremova, N., & Kuharenko, A. (2017). Convolutional neural networks pretrained on large face recognition datasets for emotion classification from video. arXiv preprint \n                    arXiv:1711.04598\n                    \n                  ."},{"issue":"1","key":"1158_CR29","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1109\/T-AFFC.2011.15","volume":"3","author":"S Koelstra","year":"2012","unstructured":"Koelstra, S., Muhl, C., Soleymani, M., Lee, J. S., Yazdani, A., Ebrahimi, T., et al. (2012). Deap: A database for emotion analysis; using physiological signals. IEEE Transactions on Affective Computing, 3(1), 18\u201331.","journal-title":"IEEE Transactions on Affective Computing"},{"key":"1158_CR30","doi-asserted-by":"crossref","unstructured":"Kollias, D., Nicolaou, M. A., Kotsia, I., Zhao, G., & Zafeiriou, S. (2017). Recognition of affect in the wild using deep neural networks. In 2017 IEEE conference on Computer Vision and Pattern Recognition Workshops (CVPRW) (pp. 1972\u20131979). IEEE.","DOI":"10.1109\/CVPRW.2017.247"},{"key":"1158_CR31","doi-asserted-by":"crossref","unstructured":"Kossaifi, J., Tzimiropoulos, G., Todorovic, S., & Pantic, M. (2017). AFEW-VA database for valence and arousal estimation in-the-wild. Image and Vision Computing.","DOI":"10.1016\/j.imavis.2017.02.001"},{"key":"1158_CR32","doi-asserted-by":"publisher","first-page":"255","DOI":"10.2307\/2532051","volume":"45","author":"I Lawrence","year":"1989","unstructured":"Lawrence, I., & Lin, K. (1989). A concordance correlation coefficient to evaluate reproducibility. Biometrics, 45, 255\u2013268.","journal-title":"Biometrics"},{"key":"1158_CR33","unstructured":"Lee, A. (2002). Welcome to virtualdub.org!\u2014virtualdub.org."},{"key":"1158_CR34","doi-asserted-by":"crossref","unstructured":"Li, J., Chen, Y., Xiao, S., Zhao, J., Roy, S., Feng, J., et al. (2017). Estimation of affective level in the wild with multiple memory networks. In Proceedings of the IEEE conference on computer vision and pattern recognition workshop.","DOI":"10.1109\/CVPRW.2017.244"},{"key":"1158_CR35","doi-asserted-by":"crossref","unstructured":"Lucey, P., Cohn, J. F., Kanade, T., Saragih, J., Ambadar, Z., & Matthews, I. (2010). The extended Cohn\u2013Kanade dataset (CK+): A complete dataset for action unit and emotion-specified expression. In 2010 IEEE computer society conference on computer vision and pattern recognition workshops (CVPRW) (pp. 94\u2013101). IEEE.","DOI":"10.1109\/CVPRW.2010.5543262"},{"key":"1158_CR36","doi-asserted-by":"crossref","unstructured":"Lucey, P., Cohn, J. F., Prkachin, K. M., Solomon, P. E., & Matthews, I. (2011). Painful data: The UNBC-McMaster shoulder pain expression archive database. In 2011 IEEE international conference on automatic face and gesture recognition and workshops (FG 2011) (pp. 57\u201364). IEEE.","DOI":"10.1109\/FG.2011.5771462"},{"key":"1158_CR37","unstructured":"Mahoor, M., & Hasani, B. (2017). Facial affect estimation in the wild using deep residual and convolutional networks. In Proceedings of the IEEE conference on computer vision and pattern recognition workshop."},{"key":"1158_CR38","doi-asserted-by":"crossref","unstructured":"Mathias, M., Benenson, R., Pedersoli, M., & Van Gool, L. (2014). Face detection without bells and whistles. In European conference on computer vision (pp. 720\u2013735). Springer.","DOI":"10.1007\/978-3-319-10593-2_47"},{"issue":"1","key":"1158_CR39","doi-asserted-by":"publisher","first-page":"5","DOI":"10.1109\/T-AFFC.2011.20","volume":"3","author":"G McKeown","year":"2012","unstructured":"McKeown, G., Valstar, M., Cowie, R., Pantic, M., & Schr\u00f6der, M. (2012). The semaine database: Annotated multimodal records of emotionally colored conversations between a person and a limited agent. IEEE Transactions on Affective Computing, 3(1), 5\u201317.","journal-title":"IEEE Transactions on Affective Computing"},{"key":"1158_CR40","unstructured":"More, A. (2016). Survey of resampling techniques for improving classification performance in unbalanced datasets. arXiv preprint \n                    arXiv:1608.06048\n                    \n                  ."},{"key":"1158_CR41","doi-asserted-by":"crossref","unstructured":"Pantic, M., Valstar, M., Rademaker, R., & Maat, L. (2005). Web-based database for facial expression analysis. In IEEE international conference on multimedia and expo, 2005. ICME 2005 (p. 5). IEEE.","DOI":"10.1109\/ICME.2005.1521424"},{"key":"1158_CR42","doi-asserted-by":"crossref","unstructured":"Parkhi, O. M., Vedaldi, A., & Zisserman, A. (2015). Deep face recognition. In BMVC (Vol. 1, p. 6).","DOI":"10.5244\/C.29.41"},{"key":"1158_CR43","volume-title":"Emotion: A psychoevolutionary synthesis","author":"R Plutchik","year":"1980","unstructured":"Plutchik, R. (1980). Emotion: A psychoevolutionary synthesis. New York, NY: Harpercollins College Division."},{"key":"1158_CR44","doi-asserted-by":"crossref","unstructured":"Ringeval, F., Schuller, B., Valstar, M., Cowie, R., & Pantic, M. (2015). AVEC 2015: The 5th international audio\/visual emotion challenge and workshop. In Proceedings of the 23rd ACM international conference on multimedia (pp. 1335\u20131336). ACM.","DOI":"10.1145\/2733373.2806408"},{"key":"1158_CR45","unstructured":"Ringeval, F., Schuller, B., Valstar, M., Gratch, J., Cowie, R., Scherer, S., et al. (2017). AVEC 2017\u2014Real-life depression, and affect recognition workshop and challenge."},{"key":"1158_CR46","unstructured":"Ringeval, F., Sonderegger, A., Sauer, J., & Lalanne, D. (2013). Introducing the RECOLA multimodal corpus of remote collaborative and affective interactions. In: 2013 10th IEEE international conference and workshops on automatic face and gesture recognition (FG) (pp. 1\u20138). IEEE."},{"issue":"10","key":"1158_CR47","doi-asserted-by":"publisher","first-page":"1152","DOI":"10.1037\/0022-3514.36.10.1152","volume":"36","author":"JA Russell","year":"1978","unstructured":"Russell, J. A. (1978). Evidence of convergent validity on the dimensions of affect. Journal of Personality and Social Psychology, 36(10), 1152.","journal-title":"Journal of Personality and Social Psychology"},{"issue":"6","key":"1158_CR48","doi-asserted-by":"publisher","first-page":"1113","DOI":"10.1109\/TPAMI.2014.2366127","volume":"37","author":"E Sariyanidi","year":"2015","unstructured":"Sariyanidi, E., Gunes, H., & Cavallaro, A. (2015). Automatic analysis of facial affect: A survey of registration, representation, and recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 37(6), 1113\u20131133.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"1158_CR49","volume-title":"Affective computing and intelligent interaction","author":"B Schuller","year":"2011","unstructured":"Schuller, B., Valstar, M., Eyben, F., McKeown, G., Cowie, R., & Pantic, M. (2011). AVEC 2011\u2014The first international audio\/visual emotion challenge. In S. D\u2019Mello, A. Graesser, B. Schuller, & J. C. Martin (Eds.), Affective computing and intelligent interaction. Berlin: Springer."},{"key":"1158_CR50","doi-asserted-by":"crossref","unstructured":"Schuller, B., Valster, M., Eyben, F., Cowie, R., & Pantic, M. (2012). AVEC 2012: The continuous audio\/visual emotion challenge. In Proceedings of the 14th ACM international conference on multimodal interaction (pp. 449\u2013456). ACM.","DOI":"10.1145\/2388676.2388776"},{"key":"1158_CR51","unstructured":"Simonyan, K., & Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv preprint \n                    arXiv:1409.1556\n                    \n                  ."},{"issue":"1","key":"1158_CR52","doi-asserted-by":"publisher","first-page":"32","DOI":"10.1109\/T-AFFC.2011.26","volume":"3","author":"I Sneddon","year":"2012","unstructured":"Sneddon, I., McRorie, M., McKeown, G., & Hanratty, J. (2012). The belfast induced natural emotion database. IEEE Transactions on Affective Computing, 3(1), 32\u201341.","journal-title":"IEEE Transactions on Affective Computing"},{"issue":"1","key":"1158_CR53","doi-asserted-by":"publisher","first-page":"42","DOI":"10.1109\/T-AFFC.2011.25","volume":"3","author":"M Soleymani","year":"2012","unstructured":"Soleymani, M., Lichtenauer, J., Pun, T., & Pantic, M. (2012). A multimodal database for affect recognition and implicit tagging. IEEE Transactions on Affective Computing, 3(1), 42\u201355.","journal-title":"IEEE Transactions on Affective Computing"},{"key":"1158_CR54","unstructured":"Szegedy, C., Ioffe, S., Vanhoucke, V., & Alemi, A. A. (2017). Inception-v4, inception-ResNet and the impact of residual connections on learning. In: AAAI (Vol. 4, p. 12)."},{"issue":"2","key":"1158_CR55","doi-asserted-by":"publisher","first-page":"97","DOI":"10.1109\/34.908962","volume":"23","author":"Yl Tian","year":"2001","unstructured":"Tian, Yl, Kanade, T., & Cohn, J. F. (2001). Recognizing action units for facial expression analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence, 23(2), 97\u2013115.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"1158_CR56","doi-asserted-by":"crossref","unstructured":"Valstar, M., Gratch, J., Schuller, B., Ringeval, F., Lalanne, D., Torres Torres, M., et al. (2016). Avec 2016: Depression, mood, and emotion recognition workshop and challenge. In Proceedings of the 6th international workshop on audio\/visual emotion challenge (pp. 3\u201310). ACM.","DOI":"10.1145\/2988257.2988258"},{"key":"1158_CR57","unstructured":"Valstar, M., & Pantic, M. (2010). Induced disgust, happiness and surprise: An addition to the mmi facial expression database. In Proceedings of 3rd international workshop on EMOTION (satellite of LREC): Corpora for research on emotion and affect (p. 65)."},{"key":"1158_CR58","doi-asserted-by":"crossref","unstructured":"Valstar, M., Schuller, B., Smith, K., Almaev, T., Eyben, F., Krajewski, J., et al. (2014). Avec 2014: 3d dimensional affect and depression recognition challenge. In Proceedings of the 4th international workshop on audio\/visual emotion challenge (pp. 3\u201310). ACM.","DOI":"10.1145\/2661806.2661807"},{"key":"1158_CR59","doi-asserted-by":"crossref","unstructured":"Valstar, M., Schuller, B., Smith, K., Eyben, F., Jiang, B., Bilakhia, S., et al. (2013). Avec 2013: The continuous audio\/visual emotion and depression recognition challenge. In Proceedings of the 3rd ACM international workshop on audio\/visual emotion challenge (pp. 3\u201310). ACM.","DOI":"10.1145\/2512530.2512533"},{"key":"1158_CR60","unstructured":"Vielzeuf, V., Pateux, S., & Jurie, F. (2017). Temporal multimodal fusion for video emotion classification in the wild. arXiv preprint \n                    arXiv:1709.07200\n                    \n                  ."},{"key":"1158_CR61","volume-title":"The measurement of emotions","author":"C Whissel","year":"1989","unstructured":"Whissel, C. (1989). The dictionary of affect in language, emotion: Theory, research and experience. In R. Plutchik & H. Kellerman (Eds.), The measurement of emotions. New York: Academic."},{"key":"1158_CR62","unstructured":"Yin, L., Chen, X., Sun, Y., Worm, T., & Reale, M. (2008). A high-resolution 3d dynamic facial expression database. In: 8th IEEE international conference on automatic face & gesture recognition, 2008. FG\u201908 (pp. 1\u20136). IEEE."},{"key":"1158_CR63","unstructured":"Yin, L., Wei, X., Sun, Y., Wang, J., & Rosato, M. J. (2006). A 3d facial expression database for facial behavior research. In: 7th international conference on automatic face and gesture recognition, 2006. FGR 2006 (pp. 211\u2013216). IEEE."},{"key":"1158_CR64","unstructured":"YouTube, L. L. C. (2011). Youtube. Retrieved, 27, 2011."},{"key":"1158_CR65","doi-asserted-by":"crossref","unstructured":"Zafeiriou, S., Kollias, D., Nicolaou, M. A., Papaioannou, A., Zhao, G., & Kotsia, I. (2017). Aff-wild: Valence and arousal \u2018in-the-wild\u2019 challenge. In 2017 IEEE conference on Computer Vision and Pattern Recognition Workshops (CVPRW) (pp. 1980\u20131987). IEEE.","DOI":"10.1109\/CVPRW.2017.248"},{"issue":"1","key":"1158_CR66","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1109\/TPAMI.2008.52","volume":"31","author":"Z Zeng","year":"2009","unstructured":"Zeng, Z., Pantic, M., Roisman, G. I., & Huang, T. S. (2009). A survey of affect recognition methods: Audio, visual, and spontaneous expressions. IEEE Transactions on Pattern Analysis and Machine Intelligence, 31(1), 39\u201358.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"}],"container-title":["International Journal of Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/article\/10.1007\/s11263-019-01158-4\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-019-01158-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-019-01158-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2020,2,12]],"date-time":"2020-02-12T19:10:59Z","timestamp":1581534659000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/s11263-019-01158-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,2,13]]},"references-count":66,"journal-issue":{"issue":"6-7","published-print":{"date-parts":[[2019,6]]}},"alternative-id":["1158"],"URL":"https:\/\/doi.org\/10.1007\/s11263-019-01158-4","relation":{},"ISSN":["0920-5691","1573-1405"],"issn-type":[{"value":"0920-5691","type":"print"},{"value":"1573-1405","type":"electronic"}],"subject":[],"published":{"date-parts":[[2019,2,13]]},"assertion":[{"value":"22 February 2018","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 January 2019","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 February 2019","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}