{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,20]],"date-time":"2026-02-20T20:58:09Z","timestamp":1771621089754,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":89,"publisher":"ACM","license":[{"start":{"date-parts":[[2020,6,15]],"date-time":"2020-06-15T00:00:00Z","timestamp":1592179200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2020,6,15]]},"DOI":"10.1145\/3386901.3388917","type":"proceedings-article","created":{"date-parts":[[2020,6,7]],"date-time":"2020-06-07T01:27:30Z","timestamp":1591493250000},"page":"448-461","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":27,"title":["EMO"],"prefix":"10.1145","author":[{"given":"Hao","family":"Wu","sequence":"first","affiliation":[{"name":"Nanjing University"}]},{"given":"Jinghao","family":"Feng","sequence":"additional","affiliation":[{"name":"Nanjing University"}]},{"given":"Xuejin","family":"Tian","sequence":"additional","affiliation":[{"name":"Nanjing University"}]},{"given":"Edward","family":"Sun","sequence":"additional","affiliation":[{"name":"Nanjing University"}]},{"given":"Yunxin","family":"Liu","sequence":"additional","affiliation":[{"name":"Microsoft Research"}]},{"given":"Bo","family":"Dong","sequence":"additional","affiliation":[{"name":"Kitware"}]},{"given":"Fengyuan","family":"Xu","sequence":"additional","affiliation":[{"name":"Nanjing University"}]},{"given":"Sheng","family":"Zhong","sequence":"additional","affiliation":[{"name":"Nanjing University"}]}],"member":"320","published-online":{"date-parts":[[2020,6,15]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"https:\/\/github.com\/saki4510t\/UVCCamera. [Online","author":"Camera About","year":"2018","unstructured":"2018. About UVC Camera . https:\/\/github.com\/saki4510t\/UVCCamera. [Online ; accessed 11- December - 2018 ]. 2018. About UVCCamera. https:\/\/github.com\/saki4510t\/UVCCamera. [Online; accessed 11-December-2018]."},{"key":"e_1_3_2_1_2_1","volume-title":"FOVE 0 Eye-tracking VR Devkit. https:\/\/www.getfove.com\/. [Online","year":"2019","unstructured":"2019. FOVE 0 Eye-tracking VR Devkit. https:\/\/www.getfove.com\/. [Online ; accessed 10- April - 2019 ]. 2019. FOVE 0 Eye-tracking VR Devkit. https:\/\/www.getfove.com\/. [Online; accessed 10-April-2019]."},{"key":"e_1_3_2_1_3_1","volume-title":"https:\/\/www.96boards.org\/product\/hikey\/. [Online","year":"2019","unstructured":"2019. HiKey. https:\/\/www.96boards.org\/product\/hikey\/. [Online ; accessed 11- December - 2019 ]. 2019. HiKey. https:\/\/www.96boards.org\/product\/hikey\/. [Online; accessed 11-December-2019]."},{"key":"e_1_3_2_1_4_1","volume-title":"HTC VIVE Pro Eye Head Mounted Display. https:\/\/enterprise.vive.com\/us\/product\/vive-pro-eye\/. [Online","year":"2019","unstructured":"2019. HTC VIVE Pro Eye Head Mounted Display. https:\/\/enterprise.vive.com\/us\/product\/vive-pro-eye\/. [Online ; accessed 05- December - 2019 ]. 2019. HTC VIVE Pro Eye Head Mounted Display. https:\/\/enterprise.vive.com\/us\/product\/vive-pro-eye\/. [Online; accessed 05-December-2019]."},{"key":"e_1_3_2_1_5_1","volume-title":"https:\/\/www.intrinsyc.com\/computing-platforms\/open-q-820-usom\/. [Online","author":"Q","year":"2019","unstructured":"2019. Open- Q 820. https:\/\/www.intrinsyc.com\/computing-platforms\/open-q-820-usom\/. [Online ; accessed 11- December - 2019 ]. 2019. Open-Q 820. https:\/\/www.intrinsyc.com\/computing-platforms\/open-q-820-usom\/. [Online; accessed 11-December-2019]."},{"key":"e_1_3_2_1_6_1","volume-title":"https:\/\/pupil-labs.com\/pupil\/. [Online","year":"2019","unstructured":"2019. Pupil. https:\/\/pupil-labs.com\/pupil\/. [Online ; accessed 10- December - 2019 ]. 2019. Pupil. https:\/\/pupil-labs.com\/pupil\/. [Online; accessed 10-December-2019]."},{"key":"e_1_3_2_1_7_1","volume-title":"https:\/\/www.tobii.com\/tech\/products\/vr\/. [Online","year":"2019","unstructured":"2019. Tobii. https:\/\/www.tobii.com\/tech\/products\/vr\/. [Online ; accessed 10- December - 2019 ]. 2019. Tobii. https:\/\/www.tobii.com\/tech\/products\/vr\/. [Online; accessed 10-December-2019]."},{"key":"e_1_3_2_1_8_1","volume-title":"https:\/\/www.microsoft.com\/en-us\/hololens. [Online","author":"HoloLens","year":"2020","unstructured":"2020. HoloLens 2. https:\/\/www.microsoft.com\/en-us\/hololens. [Online ; accessed 01- Apr- 2020 ]. 2020. HoloLens 2. https:\/\/www.microsoft.com\/en-us\/hololens. [Online; accessed 01-Apr-2020]."},{"key":"e_1_3_2_1_9_1","volume-title":"How Eye Tracking is Driving the Next Generation of AR and VR. https:\/\/vrscout.com\/news\/eye-tracking-driving-next-generation-ar-vr\/. [Online","year":"2020","unstructured":"2020. How Eye Tracking is Driving the Next Generation of AR and VR. https:\/\/vrscout.com\/news\/eye-tracking-driving-next-generation-ar-vr\/. [Online ; accessed 01- Apr- 2020 ]. 2020. How Eye Tracking is Driving the Next Generation of AR and VR. https:\/\/vrscout.com\/news\/eye-tracking-driving-next-generation-ar-vr\/. [Online; accessed 01-Apr-2020]."},{"key":"e_1_3_2_1_10_1","volume-title":"USENIX Symposium on Operating Systems Design and Implementation (OSDI). USENIX Association, 265--283","author":"Abadi Mart\u00edn","year":"2016","unstructured":"Mart\u00edn Abadi , Paul Barham , Jianmin Chen , Zhifeng Chen , Andy Davis , Jeffrey Dean , Matthieu Devin , Sanjay Ghemawat , Geoffrey Irving , Michael Isard , 2016 . Tensorflow: a system for large-scale machine learning .. In USENIX Symposium on Operating Systems Design and Implementation (OSDI). USENIX Association, 265--283 . Mart\u00edn Abadi, Paul Barham, Jianmin Chen, Zhifeng Chen, Andy Davis, Jeffrey Dean, Matthieu Devin, Sanjay Ghemawat, Geoffrey Irving, Michael Isard, et al. 2016. Tensorflow: a system for large-scale machine learning.. In USENIX Symposium on Operating Systems Design and Implementation (OSDI). USENIX Association, 265--283."},{"key":"e_1_3_2_1_11_1","volume-title":"11th International Workshop on Image Analysis for Multimedia Interactive Services WIAMIS 10","author":"Aifanti Niki","year":"2010","unstructured":"Niki Aifanti , Christos Papachristou , and Anastasios Delopoulos . 2010 . The MUG facial expression database . In 11th International Workshop on Image Analysis for Multimedia Interactive Services WIAMIS 10 . IEEE, 1--4. Niki Aifanti, Christos Papachristou, and Anastasios Delopoulos. 2010. The MUG facial expression database. In 11th International Workshop on Image Analysis for Multimedia Interactive Services WIAMIS 10. IEEE, 1--4."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jphysparis.2008.03.012"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/3204493.3204592"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-48881-3_56"},{"key":"e_1_3_2_1_15_1","volume-title":"Advances in neural information processing systems","author":"Bromley Jane","unstructured":"Jane Bromley , Isabelle Guyon , Yann LeCun , Eduard S\u00e4ckinger , and Roopak Shah . 1993. Signature Verification Using a \"Siamese\" Time Delay Neural Network . In Advances in neural information processing systems . Morgan Kaufmann Publishers Inc ., 737--744. Jane Bromley, Isabelle Guyon, Yann LeCun, Eduard S\u00e4ckinger, and Roopak Shah. 1993. Signature Verification Using a \"Siamese\" Time Delay Neural Network. In Advances in neural information processing systems. Morgan Kaufmann Publishers Inc., 737--744."},{"key":"e_1_3_2_1_16_1","volume-title":"Multimedia Content Representation, Classification and Security","author":"Chanel Guillaume","unstructured":"Guillaume Chanel , Julien Kronegg , Didier Grandjean , and Thierry Pun . 2006. Emotion assessment: Arousal evaluation using EEG's and peripheral physiological signals . In Multimedia Content Representation, Classification and Security . Springer Berlin Heidelberg , 530--537. Guillaume Chanel, Julien Kronegg, Didier Grandjean, and Thierry Pun. 2006. Emotion assessment: Arousal evaluation using EEG's and peripheral physiological signals. In Multimedia Content Representation, Classification and Security. Springer Berlin Heidelberg, 530--537."},{"key":"e_1_3_2_1_17_1","volume-title":"The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE Computer Society","author":"Chang Jason","unstructured":"Jason Chang , Donglai Wei , and John W . Fisher, III. 2013. A Video Representation Using Temporal Superpixels . In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE Computer Society , 2051--2058. Jason Chang, Donglai Wei, and John W. Fisher, III. 2013. A Video Representation Using Temporal Superpixels. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE Computer Society, 2051--2058."},{"key":"e_1_3_2_1_18_1","volume-title":"CNN-based pupil center detection for wearable gaze estimation system. Applied Computational Intelligence and Soft Computing 2017","author":"Chinsatit Warapon","year":"2017","unstructured":"Warapon Chinsatit and Takeshi Saitoh . 2017. CNN-based pupil center detection for wearable gaze estimation system. Applied Computational Intelligence and Soft Computing 2017 ( 2017 ). Warapon Chinsatit and Takeshi Saitoh. 2017. CNN-based pupil center detection for wearable gaze estimation system. Applied Computational Intelligence and Soft Computing 2017 (2017)."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2005.202"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.218"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/2971648.2971752"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/79.911197"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/2818346.2830596"},{"key":"e_1_3_2_1_25_1","volume-title":"Recurrent Neural Networks for Emotion Recognition in Video. In ACM on International Conference on Multimodal Interaction. ACM, 467--474","author":"Kahou Samira Ebrahimi","year":"2015","unstructured":"Samira Ebrahimi Kahou , Vincent Michalski , Kishore Konda , Roland Memisevic , and Christopher Pal . 2015 . Recurrent Neural Networks for Emotion Recognition in Video. In ACM on International Conference on Multimodal Interaction. ACM, 467--474 . Samira Ebrahimi Kahou, Vincent Michalski, Kishore Konda, Roland Memisevic, and Christopher Pal. 2015. Recurrent Neural Networks for Emotion Recognition in Video. In ACM on International Conference on Multimodal Interaction. ACM, 467--474."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1080\/02699939208411068"},{"key":"e_1_3_2_1_27_1","volume-title":"Friesen","author":"Ekman Paul","year":"1976","unstructured":"Paul Ekman and Wallace V . Friesen . 1976 . Measuring facial movement. Environmental psychology and nonverbal behavior 1, 1 (1976), 56--75. Paul Ekman and Wallace V. Friesen. 1976. Measuring facial movement. Environmental psychology and nonverbal behavior 1, 1 (1976), 56--75."},{"key":"e_1_3_2_1_28_1","volume-title":"EMFACS-7: Emotional facial action coding system. Unpublished manuscript","author":"Friesen Wallace V","unstructured":"Wallace V Friesen and Paul Ekman . 1983. EMFACS-7: Emotional facial action coding system. Unpublished manuscript , University of California at San Francisco (1983) , 1. Wallace V Friesen and Paul Ekman. 1983. EMFACS-7: Emotional facial action coding system. Unpublished manuscript, University of California at San Francisco (1983), 1."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/2857491.2857505"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1145\/2459236.2459273"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-42051-1_16"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.196"},{"key":"e_1_3_2_1_33_1","volume-title":"Deep Residual Learning for Image Recognition. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE Computer Society, 770--778","author":"He Kaiming","year":"2016","unstructured":"Kaiming He , Xiangyu Zhang , Shaoqing Ren , and Jian Sun . 2016 . Deep Residual Learning for Image Recognition. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE Computer Society, 770--778 . Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep Residual Learning for Image Recognition. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE Computer Society, 770--778."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1109\/WACV.2019.00178"},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/2800835.2800910"},{"key":"e_1_3_2_1_36_1","volume-title":"Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861","author":"Howard Andrew G","year":"2017","unstructured":"Andrew G Howard , Menglong Zhu , Bo Chen , Dmitry Kalenichenko , Weijun Wang , Tobias Weyand , Marco Andreetto , and Hartwig Adam . 2017 . Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861 (2017). http:\/\/arxiv.org\/abs\/1704.04861 Andrew G Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, and Hartwig Adam. 2017. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861 (2017). http:\/\/arxiv.org\/abs\/1704.04861"},{"key":"e_1_3_2_1_37_1","volume-title":"SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and &lt","author":"Iandola Forrest N","year":"2016","unstructured":"Forrest N Iandola , Song Han , Matthew W Moskewicz , Khalid Ashraf , William J Dally , and Kurt Keutzer . 2016. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and &lt ; 0.5 MB model size. arXiv preprint arXiv:1602.07360 ( 2016 ). https:\/\/arxiv.org\/abs\/1602.07360 Forrest N Iandola, Song Han, Matthew W Moskewicz, Khalid Ashraf, William J Dally, and Kurt Keutzer. 2016. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and &lt; 0.5 MB model size. arXiv preprint arXiv:1602.07360 (2016). https:\/\/arxiv.org\/abs\/1602.07360"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2005.03.004"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/2638728.2641695"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1145\/2818346.2830590"},{"key":"e_1_3_2_1_41_1","volume-title":"Hierarchical Committee of Deep CNNs with Exponentially-Weighted Decision Fusion for Static Facial Expression Recognition. In ACM on International Conference on Multimodal Interaction. ACM, 427--434","author":"Kim Bo-Kyeong","year":"2015","unstructured":"Bo-Kyeong Kim , Hwaran Lee , Jihyeon Roh , and Soo-Young Lee . 2015 . Hierarchical Committee of Deep CNNs with Exponentially-Weighted Decision Fusion for Static Facial Expression Recognition. In ACM on International Conference on Multimodal Interaction. ACM, 427--434 . Bo-Kyeong Kim, Hwaran Lee, Jihyeon Roh, and Soo-Young Lee. 2015. Hierarchical Committee of Deep CNNs with Exponentially-Weighted Decision Fusion for Static Facial Expression Recognition. In ACM on International Conference on Multimodal Interaction. ACM, 427--434."},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/2578153.2578209"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1145\/2807442.2807479"},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1145\/2968219.2968337"},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2016.59"},{"key":"e_1_3_2_1_46_1","first-page":"525","article-title":"The emotional brain","volume":"25","author":"LeDoux Joseph","year":"1997","unstructured":"Joseph LeDoux and Jules R Bemporad . 1997 . The emotional brain . Journal of the American Academy of Psychoanalysis 25 , 3 (1997), 525 -- 528 . Joseph LeDoux and Jules R Bemporad. 1997. The emotional brain. Journal of the American Academy of Psychoanalysis 25, 3 (1997), 525--528.","journal-title":"Journal of the American Academy of Psychoanalysis"},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.adhoc.2018.08.021"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1145\/2818346.2830587"},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1145\/2818346.2830587"},{"key":"e_1_3_2_1_50_1","volume-title":"Proceedings of the 11th annual international conference on Mobile systems, applications, and services. ACM, 389--402","author":"LiKamWa Robert","year":"2013","unstructured":"Robert LiKamWa , Yunxin Liu , Nicholas D. Lane , and Lin Zhong . 2013 . MoodScope: building a mood sensor from smartphone usage patterns . In Proceedings of the 11th annual international conference on Mobile systems, applications, and services. ACM, 389--402 . Robert LiKamWa, Yunxin Liu, Nicholas D. Lane, and Lin Zhong. 2013. MoodScope: building a mood sensor from smartphone usage patterns. In Proceedings of the 11th annual international conference on Mobile systems, applications, and services. ACM, 389--402."},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"crossref","unstructured":"G. Littlewort J. Whitehill T. Wu I. Fasel M. Frank J. Movellan and M. Bartlett. 2011. The computer expression recognition toolbox (CERT). In Face and Gesture 2011. IEEE 298--305.  G. Littlewort J. Whitehill T. Wu I. Fasel M. Frank J. Movellan and M. Bartlett. 2011. The computer expression recognition toolbox (CERT). In Face and Gesture 2011. IEEE 298--305.","DOI":"10.1109\/FG.2011.5771414"},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2008.17"},{"key":"e_1_3_2_1_53_1","volume-title":"Edge Assisted Real-time Object Detection for Mobile Augmented Reality. In In Proceedings of The 25th Annual International Conference on Mobile Computing and Networking. ACM.","author":"Liu Luyang","year":"2019","unstructured":"Luyang Liu , Hongyu Li , and Marco Gruteser . 2019 . Edge Assisted Real-time Object Detection for Mobile Augmented Reality. In In Proceedings of The 25th Annual International Conference on Mobile Computing and Networking. ACM. Luyang Liu, Hongyu Li, and Marco Gruteser. 2019. Edge Assisted Real-time Object Detection for Mobile Augmented Reality. In In Proceedings of The 25th Annual International Conference on Mobile Computing and Networking. ACM."},{"key":"e_1_3_2_1_54_1","volume-title":"International Conference and Workshops on Automatic Face and Gesture Recognition. IEEE Computer Society, 1--6.","author":"Liu Mengyi","year":"2013","unstructured":"Mengyi Liu , Shaoxin Li , Shiguang Shan , and Xilin Chen . 2013 . Au-aware deep networks for facial expression recognition . In International Conference and Workshops on Automatic Face and Gesture Recognition. IEEE Computer Society, 1--6. Mengyi Liu, Shaoxin Li, Shiguang Shan, and Xilin Chen. 2013. Au-aware deep networks for facial expression recognition. In International Conference and Workshops on Automatic Face and Gesture Recognition. IEEE Computer Society, 1--6."},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.233"},{"key":"e_1_3_2_1_56_1","volume-title":"Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability","author":"MacQueen J.","year":"1967","unstructured":"J. MacQueen . 1967 . Some methods for classification and analysis of multivariate observations . In Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability . University of California Press, 281--297. J. MacQueen. 1967. Some methods for classification and analysis of multivariate observations. In Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability. University of California Press, 281--297."},{"key":"e_1_3_2_1_57_1","volume-title":"Proceedings of the 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing and the Proceedings of the 2015 ACM International Symposium on Wearable Computers. ACM, 357--360","author":"Masai Katsutoshi","year":"2015","unstructured":"Katsutoshi Masai , Yuta Sugiura , Katsuhiro Suzuki , Sho Shimamura , Kai Kunze , Masa Ogata , Masahiko Inami , and Maki Sugimoto . 2015 . Affective wear: towards recognizing affect in real life . In Proceedings of the 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing and the Proceedings of the 2015 ACM International Symposium on Wearable Computers. ACM, 357--360 . Katsutoshi Masai, Yuta Sugiura, Katsuhiro Suzuki, Sho Shimamura, Kai Kunze, Masa Ogata, Masahiko Inami, and Maki Sugimoto. 2015. Affective wear: towards recognizing affect in real life. In Proceedings of the 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing and the Proceedings of the 2015 ACM International Symposium on Wearable Computers. ACM, 357--360."},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1145\/2851581.2890247"},{"key":"e_1_3_2_1_59_1","volume-title":"Recurrent Convolutional Network for Video-Based Person Re-identification. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE Computer Society, 1325--1334","author":"McLaughlin Niall","unstructured":"Niall McLaughlin , Jes\u00fas Mart\u00ednez del Rinc\u00f3n , and Paul C. Miller . 2016 . Recurrent Convolutional Network for Video-Based Person Re-identification. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE Computer Society, 1325--1334 . Niall McLaughlin, Jes\u00fas Mart\u00ednez del Rinc\u00f3n, and Paul C. Miller. 2016. Recurrent Convolutional Network for Video-Based Person Re-identification. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE Computer Society, 1325--1334."},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1080\/026999397379980"},{"key":"e_1_3_2_1_61_1","volume-title":"A wearable gaze tracking system for children in unconstrained environments. Computer Vision and Image Understanding","author":"Noris Basilio","year":"2011","unstructured":"Basilio Noris , Jean-Baptiste Keller , and Aude Billard . 2011. A wearable gaze tracking system for children in unconstrained environments. Computer Vision and Image Understanding ( 2011 ), 476--486. Basilio Noris, Jean-Baptiste Keller, and Aude Billard. 2011. A wearable gaze tracking system for children in unconstrained environments. Computer Vision and Image Understanding (2011), 476--486."},{"key":"e_1_3_2_1_62_1","volume-title":"Real-time emotion recognition for gaming using deep convolutional network features. arXiv preprint arXiv:1408.3750","author":"Ouellet S\u00e9bastien","year":"2014","unstructured":"S\u00e9bastien Ouellet . 2014. Real-time emotion recognition for gaming using deep convolutional network features. arXiv preprint arXiv:1408.3750 ( 2014 ). https:\/\/arxiv.org\/abs\/1408.3750 S\u00e9bastien Ouellet. 2014. Real-time emotion recognition for gaming using deep convolutional network features. arXiv preprint arXiv:1408.3750 (2014). https:\/\/arxiv.org\/abs\/1408.3750"},{"key":"e_1_3_2_1_63_1","volume-title":"Real-time emotion recognition for gaming using deep convolutional network features. arXiv preprint arXiv:1408.3750","author":"Ouellet S\u00e9bastien","year":"2014","unstructured":"S\u00e9bastien Ouellet . 2014. Real-time emotion recognition for gaming using deep convolutional network features. arXiv preprint arXiv:1408.3750 ( 2014 ). https:\/\/arxiv.org\/abs\/1408.3750 S\u00e9bastien Ouellet. 2014. Real-time emotion recognition for gaming using deep convolutional network features. arXiv preprint arXiv:1408.3750 (2014). https:\/\/arxiv.org\/abs\/1408.3750"},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.1145\/3214281"},{"key":"e_1_3_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.1145\/1743666.1743685"},{"key":"e_1_3_2_1_66_1","volume-title":"Assessing the effectiveness of a large database of emotion-eliciting films: A new tool for emotion researchers. Cognition and Emotion","author":"Schaefer Alexandre","year":"2010","unstructured":"Alexandre Schaefer , Fr\u00e9d\u00e9ric Nils , Xavier Sanchez , and Pierre Philippot . 2010. Assessing the effectiveness of a large database of emotion-eliciting films: A new tool for emotion researchers. Cognition and Emotion ( 2010 ), 1153--1172. Alexandre Schaefer, Fr\u00e9d\u00e9ric Nils, Xavier Sanchez, and Pierre Philippot. 2010. Assessing the effectiveness of a large database of emotion-eliciting films: A new tool for emotion researchers. Cognition and Emotion (2010), 1153--1172."},{"key":"e_1_3_2_1_67_1","doi-asserted-by":"publisher","DOI":"10.1145\/632716.632878"},{"key":"e_1_3_2_1_68_1","volume-title":"McOwan","author":"Shan Caifeng","year":"2009","unstructured":"Caifeng Shan , Shaogang Gong , and Peter W . McOwan . 2009 . Facial Expression Recognition Based on Local Binary Patterns: A Comprehensive Study. Image and vision Computing ( 2009), 803--816. Caifeng Shan, Shaogang Gong, and Peter W. McOwan. 2009. Facial Expression Recognition Based on Local Binary Patterns: A Comprehensive Study. Image and vision Computing (2009), 803--816."},{"key":"e_1_3_2_1_69_1","volume-title":"Face and Gesture","author":"Shreve Matthew","unstructured":"Matthew Shreve , Sridhar Godavarthy , Dmitry Goldgof , and Sudeep Sarkar . 2011. Macro-and micro-expression spotting in long videos using spatio-temporal strain . In Face and Gesture . IEEE , 51--56. Matthew Shreve, Sridhar Godavarthy, Dmitry Goldgof, and Sudeep Sarkar. 2011. Macro-and micro-expression spotting in long videos using spatio-temporal strain. In Face and Gesture. IEEE, 51--56."},{"key":"e_1_3_2_1_70_1","volume-title":"Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556","author":"Simonyan Karen","year":"2014","unstructured":"Karen Simonyan and Andrew Zisserman . 2014. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 ( 2014 ). http:\/\/arxiv.org\/abs\/1409.1556 Karen Simonyan and Andrew Zisserman. 2014. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014). http:\/\/arxiv.org\/abs\/1409.1556"},{"key":"e_1_3_2_1_71_1","doi-asserted-by":"publisher","DOI":"10.1109\/T-AFFC.2011.37"},{"key":"e_1_3_2_1_72_1","doi-asserted-by":"publisher","DOI":"10.1145\/2750858.2807520"},{"key":"e_1_3_2_1_73_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"e_1_3_2_1_74_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.308"},{"key":"e_1_3_2_1_75_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.158"},{"key":"e_1_3_2_1_76_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46484-8_48"},{"key":"e_1_3_2_1_77_1","volume-title":"A siamese long short-term memory architecture for human re-identification","author":"Varior Rahul Rama","unstructured":"Rahul Rama Varior , Bing Shuai , Jiwen Lu , Dong Xu , and Gang Wang . 2016. A siamese long short-term memory architecture for human re-identification . In ECCV. Springer , 135--153. Rahul Rama Varior, Bing Shuai, Jiwen Lu, Dong Xu, and Gang Wang. 2016. A siamese long short-term memory architecture for human re-identification. In ECCV. Springer, 135--153."},{"key":"e_1_3_2_1_78_1","doi-asserted-by":"publisher","DOI":"10.1109\/THS.2013.6699082"},{"key":"e_1_3_2_1_79_1","volume-title":"Confusion matrix. https:\/\/en.wikipedia.org\/w\/index.php?title=Confusion_matrix&oldid=870941727. [Online","author":"Wikipedia","year":"2019","unstructured":"Wikipedia contributors. 2019. Confusion matrix. https:\/\/en.wikipedia.org\/w\/index.php?title=Confusion_matrix&oldid=870941727. [Online ; accessed 6- January - 2019 ]. Wikipedia contributors. 2019. Confusion matrix. https:\/\/en.wikipedia.org\/w\/index.php?title=Confusion_matrix&oldid=870941727. [Online; accessed 6-January-2019]."},{"key":"e_1_3_2_1_80_1","doi-asserted-by":"publisher","DOI":"10.1145\/3241539.3241563"},{"key":"e_1_3_2_1_81_1","volume-title":"VideoChef: Efficient Approximation for Streaming Video Processing Pipelines. In 2018 USENIX Annual Technical Conference (USENIX ATC 18)","author":"Xu Ran","year":"2018","unstructured":"Ran Xu , Jinkyu Koo , Rakesh Kumar , Peter Bai , Subrata Mitra , Sasa Misailovic , and Saurabh Bagchi . 2018 . VideoChef: Efficient Approximation for Streaming Video Processing Pipelines. In 2018 USENIX Annual Technical Conference (USENIX ATC 18) . USENIX Association, 43--56. Ran Xu, Jinkyu Koo, Rakesh Kumar, Peter Bai, Subrata Mitra, Sasa Misailovic, and Saurabh Bagchi. 2018. VideoChef: Efficient Approximation for Streaming Video Processing Pipelines. In 2018 USENIX Annual Technical Conference (USENIX ATC 18). USENIX Association, 43--56."},{"key":"e_1_3_2_1_82_1","volume-title":"Image Analysis and Recognition","author":"Ya\u011fi\u015f Ekin","unstructured":"Ekin Ya\u011fi\u015f and Mustafa Unel . 2018. Facial Expression Based Emotion Recognition Using Neural Networks . In Image Analysis and Recognition . Springer International Publishing , 210--217. Ekin Ya\u011fi\u015f and Mustafa Unel. 2018. Facial Expression Based Emotion Recognition Using Neural Networks. In Image Analysis and Recognition. Springer International Publishing, 210--217."},{"key":"e_1_3_2_1_83_1","volume-title":"International Conference on Mobile Computing, Applications, and Services. Springer, 149--168","author":"Yang Xiaochao","year":"2012","unstructured":"Xiaochao Yang , Chuang-Wen You , Hong Lu , Mu Lin , Nicholas D Lane , and Andrew T Campbell . 2012 . Visage: A face interpretation engine for smartphone applications . In International Conference on Mobile Computing, Applications, and Services. Springer, 149--168 . Xiaochao Yang, Chuang-Wen You, Hong Lu, Mu Lin, Nicholas D Lane, and Andrew T Campbell. 2012. Visage: A face interpretation engine for smartphone applications. In International Conference on Mobile Computing, Applications, and Services. Springer, 149--168."},{"key":"e_1_3_2_1_84_1","doi-asserted-by":"publisher","DOI":"10.1145\/2818346.2830595"},{"key":"e_1_3_2_1_85_1","volume-title":"Image Based Static Facial Expression Recognition with Multiple Deep Network Learning. In ACM on International Conference on Multimodal Interaction. ACM, 435--442","author":"Yu Zhiding","year":"2015","unstructured":"Zhiding Yu and Cha Zhang . 2015 . Image Based Static Facial Expression Recognition with Multiple Deep Network Learning. In ACM on International Conference on Multimodal Interaction. ACM, 435--442 . Zhiding Yu and Cha Zhang. 2015. Image Based Static Facial Expression Recognition with Multiple Deep Network Learning. In ACM on International Conference on Multimodal Interaction. ACM, 435--442."},{"key":"e_1_3_2_1_86_1","doi-asserted-by":"publisher","DOI":"10.1145\/3158369"},{"key":"e_1_3_2_1_87_1","volume-title":"ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE Computer Society, 6848--6856","author":"Zhang Xiangyu","year":"2018","unstructured":"Xiangyu Zhang , Xinyu Zhou , Mengxiao Lin , and Jian Sun . 2018 . ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE Computer Society, 6848--6856 . Xiangyu Zhang, Xinyu Zhou, Mengxiao Lin, and Jian Sun. 2018. ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE Computer Society, 6848--6856."},{"key":"e_1_3_2_1_88_1","doi-asserted-by":"publisher","DOI":"10.1145\/3161410"},{"key":"e_1_3_2_1_89_1","doi-asserted-by":"publisher","DOI":"10.1145\/2973750.2973762"}],"event":{"name":"MobiSys '20: The 18th Annual International Conference on Mobile Systems, Applications, and Services","location":"Toronto Ontario Canada","acronym":"MobiSys '20","sponsor":["SIGMOBILE ACM Special Interest Group on Mobility of Systems, Users, Data and Computing"]},"container-title":["Proceedings of the 18th International Conference on Mobile Systems, Applications, and Services"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3386901.3388917","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3386901.3388917","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T22:33:26Z","timestamp":1750199606000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3386901.3388917"}},"subtitle":["real-time emotion recognition from single-eye images for resource-constrained eyewear devices"],"short-title":[],"issued":{"date-parts":[[2020,6,15]]},"references-count":89,"alternative-id":["10.1145\/3386901.3388917","10.1145\/3386901"],"URL":"https:\/\/doi.org\/10.1145\/3386901.3388917","relation":{},"subject":[],"published":{"date-parts":[[2020,6,15]]},"assertion":[{"value":"2020-06-15","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}