{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,22]],"date-time":"2026-04-22T06:11:25Z","timestamp":1776838285616,"version":"3.51.2"},"reference-count":32,"publisher":"IEEE","license":[{"start":{"date-parts":[[2019,9,1]],"date-time":"2019-09-01T00:00:00Z","timestamp":1567296000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2019,9,1]],"date-time":"2019-09-01T00:00:00Z","timestamp":1567296000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2019,9,1]],"date-time":"2019-09-01T00:00:00Z","timestamp":1567296000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019,9]]},"DOI":"10.1109\/icip.2019.8803603","type":"proceedings-article","created":{"date-parts":[[2019,8,26]],"date-time":"2019-08-26T19:32:48Z","timestamp":1566847968000},"page":"3866-3870","source":"Crossref","is-referenced-by-count":164,"title":["Frame Attention Networks for Facial Expression Recognition in Videos"],"prefix":"10.1109","author":[{"given":"Debin","family":"Meng","sequence":"first","affiliation":[]},{"given":"Xiaojiang","family":"Peng","sequence":"additional","affiliation":[]},{"given":"Kai","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Yu","family":"Qiao","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref32","article-title":"Multi-feature based emotion recognition for video clips","author":"liu","year":"2018","journal-title":"ACM ICMI"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1145\/3242969.3264978"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1145\/3136755.3143009"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1145\/3136755.3143011"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2018.03.068"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.341"},{"key":"ref13","article-title":"Combining modality specific deep neural networks for emotion recognition in video","author":"kahou","year":"2013","journal-title":"ACM ICMI"},{"key":"ref14","article-title":"Attention is all you need","author":"vaswani","year":"2017","journal-title":"NIPS"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.554"},{"key":"ref16","article-title":"Learning to compare: Relation network for few-shot learning","author":"yongxin yang","year":"2018","journal-title":"CVPR"},{"key":"ref17","article-title":"The extended cohnkanade dataset (ck+): A complete dataset for action unit and emotion-specified expression","author":"lucey","year":"2010","journal-title":"cvprw"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.226"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2018.00286"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1145\/2993148.2997632"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1145\/2993148.2997627"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2011.5995345"},{"key":"ref3","article-title":"Deep learning using linear support vector machines","author":"tang","year":"2013","journal-title":"CS"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1145\/2993148.2997639"},{"key":"ref5","article-title":"Convolutional neural networks pretrained on large face recognition datasets for emotion classification from video","author":"knyazev","year":"2017","journal-title":"ACM ICMI"},{"key":"ref8","article-title":"Video-based emotion recognition using cnn-rnn and c3d hybrid networks","author":"liu","year":"2016","journal-title":"ACM ICMI"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.510"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2008.08.005"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1145\/3136755.3143012"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2005.09.011"},{"key":"ref20","article-title":"Emotiw 2018: Audio-video, student engagement and group-level affect prediction","author":"dhall","year":"2018","journal-title":"arXiv preprint 1808 07773"},{"key":"ref22","article-title":"Ms-celeb-1m: A dataset and benchmark for large-scale face recognition","author":"guo","year":"2016","journal-title":"ECCV"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/MMUL.2012.26"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2017.2689999"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1145\/2993148.2993165"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.602"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/FG.2018.00051"}],"event":{"name":"2019 IEEE International Conference on Image Processing (ICIP)","location":"Taipei, Taiwan","start":{"date-parts":[[2019,9,22]]},"end":{"date-parts":[[2019,9,25]]}},"container-title":["2019 IEEE International Conference on Image Processing (ICIP)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/8791230\/8799366\/08803603.pdf?arnumber=8803603","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,7,18]],"date-time":"2022-07-18T15:37:06Z","timestamp":1658158626000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8803603\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,9]]},"references-count":32,"URL":"https:\/\/doi.org\/10.1109\/icip.2019.8803603","relation":{},"subject":[],"published":{"date-parts":[[2019,9]]}}}