{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,20]],"date-time":"2026-03-20T15:47:04Z","timestamp":1774021624504,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":36,"publisher":"ACM","funder":[{"name":"National Science Foundation AI Institute for Foundations of Machine Learning (IFML)","award":["Grant 2019844"],"award-info":[{"award-number":["Grant 2019844"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,8,10]]},"DOI":"10.1145\/3721238.3730653","type":"proceedings-article","created":{"date-parts":[[2025,7,23]],"date-time":"2025-07-23T08:40:47Z","timestamp":1753260047000},"page":"1-11","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["FaceExpressions-70k: A Dataset of Perceived Expression Differences"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2636-4737","authenticated-orcid":false,"given":"Avinab","family":"Saha","sequence":"first","affiliation":[{"name":"The University of Texas at Austin, Austin, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2710-083X","authenticated-orcid":false,"given":"Yu-Chih","family":"Chen","sequence":"additional","affiliation":[{"name":"The University of Texas at Austin, Austin, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7660-4802","authenticated-orcid":false,"given":"Jean-Charles","family":"Bazin","sequence":"additional","affiliation":[{"name":"Reality Labs, Meta, Sunnyvale, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8599-6681","authenticated-orcid":false,"given":"Christian","family":"H\u00e4ne","sequence":"additional","affiliation":[{"name":"Reality Labs, Meta, Sunnyvale, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9072-4250","authenticated-orcid":false,"given":"Ioannis","family":"Katsavounidis","sequence":"additional","affiliation":[{"name":"Reality Labs, Meta, Sunnyvale, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7367-0131","authenticated-orcid":false,"given":"Alexandre","family":"Chapiro","sequence":"additional","affiliation":[{"name":"Reality Labs, Meta, Sunnyvale, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6067-710X","authenticated-orcid":false,"given":"Alan","family":"Bovik","sequence":"additional","affiliation":[{"name":"The University of Texas at Austin, Austin, USA"}]}],"member":"320","published-online":{"date-parts":[[2025,7,27]]},"reference":[{"key":"e_1_3_3_3_2_1","unstructured":"Amazon. 2025. Amazon Mechanical Turk. https:\/\/www.mturk.com Accessed: 2025-01-08."},{"key":"e_1_3_3_3_3_1","doi-asserted-by":"publisher","unstructured":"Chen Cao Yanlin Weng Shun Zhou Yiying Tong and Kun Zhou. 2014. FaceWarehouse: A 3D Facial Expression Database for Visual Computing. IEEE Transactions on Visualization and Computer Graphics 20 3 (2014) 413\u2013425. 10.1109\/TVCG.2013.249","DOI":"10.1109\/TVCG.2013.249"},{"key":"e_1_3_3_3_4_1","unstructured":"Xinlei Chen Haoqi Fan Ross Girshick and Kaiming He. 2020. Improved Baselines with Momentum Contrastive Learning. arxiv:https:\/\/arXiv.org\/abs\/2003.04297\u00a0[cs.CV]"},{"key":"e_1_3_3_3_5_1","doi-asserted-by":"publisher","unstructured":"Yu-Chih Chen Avinab Saha Alexandre Chapiro Christian H\u00e4ne Jean-Charles Bazin Bo Qiu Stefano Zanetti Ioannis Katsavounidis and Alan\u00a0C. Bovik. 2024. Subjective and Objective Quality Assessment of Rendered Human Avatar Videos in Virtual Reality. IEEE Transactions on Image Processing 33 (2024) 5740\u20135754. 10.1109\/TIP.2024.3468881","DOI":"10.1109\/TIP.2024.3468881"},{"key":"e_1_3_3_3_6_1","doi-asserted-by":"publisher","unstructured":"Zeynep Cipiloglu\u00a0Yildiz. 2023. Learning a crowd-powered perceptual distance metric for facial blendshapes. J. Image Video Process. 2023 1 (May 2023) 20\u00a0pages. 10.1186\/s13640-023-00609-w","DOI":"10.1186\/s13640-023-00609-w"},{"key":"e_1_3_3_3_7_1","unstructured":"Google. 2023. MediaPipe. https:\/\/github.com\/google\/mediapipe. Accessed: 2025-01-08."},{"key":"e_1_3_3_3_8_1","unstructured":"Kaiming He Xiangyu Zhang Shaoqing Ren and Jian Sun. 2015. Deep Residual Learning for Image Recognition. arxiv:https:\/\/arXiv.org\/abs\/1512.03385\u00a0[cs.CV]"},{"key":"e_1_3_3_3_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/3680528.3687639"},{"key":"e_1_3_3_3_10_1","volume-title":"Eurographics 2014 - State of the Art Reports","author":"Lewis J.\u00a0P.","year":"2014","unstructured":"J.\u00a0P. Lewis, Ken Anjyo, Taehyun Rhee, Mengjie Zhang, Fred Pighin, and Zhigang Deng. 2014. Practice and Theory of Blendshape Facial Models. In Eurographics 2014 - State of the Art Reports, Sylvain Lefebvre and Michela Spagnuolo (Eds.). The Eurographics Association. https:\/\/doi.org\/\/10.2312\/egst.20141042"},{"key":"e_1_3_3_3_11_1","doi-asserted-by":"crossref","unstructured":"Zhi Li Christos\u00a0G. Bampis Lucjan Janowski and Ioannis Katsavounidis. 2020. A simple model for subject behavior in subjective experiments. Electronic Imaging 2020 11 (2020) 131\u20131.","DOI":"10.2352\/ISSN.2470-1173.2020.11.HVEI-131"},{"key":"e_1_3_3_3_12_1","unstructured":"Hanwei Liu Rudong An Zhimeng Zhang Bowen Ma Wei Zhang Yan Song Yujing Hu Wei Chen and Yu Ding. 2024. Norface: Improving Facial Expression Analysis by Identity Normalization. arxiv:https:\/\/arXiv.org\/abs\/2407.15617\u00a0[cs.CV] https:\/\/arxiv.org\/abs\/2407.15617"},{"key":"e_1_3_3_3_13_1","doi-asserted-by":"crossref","unstructured":"Stephen Lombardi Jason Saragih Tomas Simon and Yaser Sheikh. 2018. Deep appearance models for face rendering. ACM Transactions on Graphics (ToG) 37 4 (2018) 1\u201313.","DOI":"10.1145\/3197517.3201401"},{"key":"e_1_3_3_3_14_1","doi-asserted-by":"crossref","unstructured":"Stephen Lombardi Tomas Simon Gabriel Schwartz Michael Zollhoefer Yaser Sheikh and Jason Saragih. 2021. Mixture of volumetric primitives for efficient neural rendering. ACM Transactions on Graphics (ToG) 40 4 (2021) 1\u201313.","DOI":"10.1145\/3450626.3459863"},{"key":"e_1_3_3_3_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW59228.2023.00630"},{"key":"e_1_3_3_3_16_1","doi-asserted-by":"publisher","unstructured":"Rachel McDonnell Katja Zibrek Emma Carrigan and Rozenn Dahyot. 2021. Model for predicting perception of facial action unit activation using virtual humans. Comput. Graph. 100 C (Nov. 2021) 81\u201392. 10.1016\/j.cag.2021.07.022","DOI":"10.1016\/j.cag.2021.07.022"},{"key":"e_1_3_3_3_17_1","unstructured":"Mang Ning Albert\u00a0Ali Salah and Itir\u00a0Onal Ertugrul. 2024. Representation Learning and Identity Adversarial Training for Facial Behavior Understanding. arxiv:https:\/\/arXiv.org\/abs\/2407.11243\u00a0[cs.CV]"},{"key":"e_1_3_3_3_18_1","doi-asserted-by":"crossref","unstructured":"Kristine\u00a0L Nowak and Jesse Fox. 2018. Avatars and computer-mediated communication: a review of the definitions uses and effects of digital representations. Review of Communication Research 6 (2018) 30\u201353.","DOI":"10.12840\/issn.2255-4165.2018.06.01.015"},{"key":"e_1_3_3_3_19_1","unstructured":"Alec Radford Jong\u00a0W. Kim Chris Hallacy Aditya Ramesh Gabriel Goh Sandhini Agarwal Girish Sastry Amanda Askell Pamela Mishkin Jack Clark Gretchen Krueger and Ilya Sutskever. 2021. Learning Transferable Visual Models From Natural Language Supervision. arxiv:https:\/\/arXiv.org\/abs\/2103.00020\u00a0[cs.CV] https:\/\/arxiv.org\/abs\/2103.00020"},{"key":"e_1_3_3_3_20_1","doi-asserted-by":"publisher","DOI":"10.1111\/cgf.15064"},{"key":"e_1_3_3_3_21_1","doi-asserted-by":"publisher","unstructured":"Bryan\u00a0N. Reyes Sarah\u00a0C. Segal and Melody\u00a0C. Moulson. 2018. An investigation of the effect of race-based social categorization on adults\u2019 recognition of emotion. PLOS ONE 13 2 (2018) e0192418. 10.1371\/journal.pone.0192418","DOI":"10.1371\/journal.pone.0192418"},{"key":"e_1_3_3_3_22_1","doi-asserted-by":"publisher","unstructured":"Avinab Saha Yu-Chih Chen Chase Davis Bo Qiu Xiaoming Wang Rahul Gowda Ioannis Katsavounidis and Alan\u00a0C. Bovik. 2023. Study of Subjective and Objective Quality Assessment of Mobile Cloud Gaming Videos. IEEE Transactions on Image Processing 32 (2023) 3295\u20133310. 10.1109\/TIP.2023.3281170","DOI":"10.1109\/TIP.2023.3281170"},{"key":"e_1_3_3_3_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP46576.2022.9897940"},{"key":"e_1_3_3_3_24_1","doi-asserted-by":"crossref","unstructured":"Marie-H\u00e9l\u00e8ne Tessier Chlo\u00e9 Gingras Nicolas Robitaille and Philip\u00a0L Jackson. 2019. Toward dynamic pain expressions in avatars: perceived realism and pain level of different action unit orders. Computers in Human Behavior 96 (2019) 95\u2013109.","DOI":"10.1016\/j.chb.2019.02.001"},{"key":"e_1_3_3_3_25_1","doi-asserted-by":"crossref","unstructured":"Thomas Treal Philip\u00a0L. Jackson Jean Jeuvrey Nicolas Vignais and Aurore Meugnot. 2021. Natural human postural oscillations enhance the empathic response to a facial pain expression in a virtual character. Scientific Reports 11 1 (2021) 12493.","DOI":"10.1038\/s41598-021-91710-5"},{"key":"e_1_3_3_3_26_1","unstructured":"International\u00a0Telecommunication Union. 2004. Objective Perceptual Assessment of Video Quality: Full Reference Television. https:\/\/www.itu.int\/ITU-T\/studygroups\/com09\/docs\/tutorial_opavc.pdf Accessed: 2024-01-08."},{"key":"e_1_3_3_3_27_1","doi-asserted-by":"publisher","unstructured":"Abhinau\u00a0K. Venkataramanan and Alan\u00a0C. Bovik. 2024. Subjective Quality Assessment of Compressed Tone-Mapped High Dynamic Range Videos. IEEE Transactions on Image Processing 33 (2024) 5440\u20135455. 10.1109\/tip.2024.3463418","DOI":"10.1109\/tip.2024.3463418"},{"key":"e_1_3_3_3_28_1","doi-asserted-by":"crossref","unstructured":"Christian Wallraven Martin Breidt Douglas\u00a0W Cunningham and Heinrich\u00a0H B\u00fclthoff. 2008. Evaluating the perceptual realism of animated facial expressions. ACM Transactions on Applied Perception (TAP) 4 4 (2008) 1\u201320.","DOI":"10.1145\/1278760.1278764"},{"key":"e_1_3_3_3_29_1","doi-asserted-by":"publisher","unstructured":"Zhou Wang Alan\u00a0C. Bovik Hamid\u00a0R. Sheikh and Eero\u00a0P. Simoncelli. 2004. Image quality assessment: from error visibility to structural similarity. IEEE Transactions on Image Processing 13 4 (2004) 600\u2013612. 10.1109\/TIP.2003.819861","DOI":"10.1109\/TIP.2003.819861"},{"key":"e_1_3_3_3_30_1","doi-asserted-by":"publisher","unstructured":"Andrew\u00a0B Watson and Denis\u00a0G Pelli. 1983. QUEST: A Bayesian adaptive psychometric method. Perception & Psychophysics 33 2 (1983) 113\u2013120. 10.3758\/BF03202828","DOI":"10.3758\/BF03202828"},{"key":"e_1_3_3_3_31_1","doi-asserted-by":"crossref","unstructured":"Krzysztof Wolski Laura Trutoiu Zhao Dong Zhengyang Shen Kevin MacKenzie and Alexandre Chapiro. 2022. Geo-metric: A Perceptual Dataset of Distortions on Faces. ACM Transactions on Graphics (TOG) 41 (2022).","DOI":"10.1145\/3550454.3555475"},{"key":"e_1_3_3_3_32_1","unstructured":"Chenghsin Wuu Ningyuan Zheng Scott Ardisson Rohan Bali Danielle Belko Eric Brockmeyer Lucas Evans Timothy Godisart Hyowon Ha Xuhua Huang Alexander Hypes Taylor Koska Steven Krenn Stephen Lombardi Xiaomin Luo Kevyn McPhail Laura Millerschoen Michal Perdoch Mark Pitts Alexander Richard Jason Saragih Junko Saragih Takaaki Shiratori Tomas Simon Matt Stewart Autumn Trimble Xinshuo Weng David Whitewolf Chenglei Wu Shoou-I Yu and Yaser Sheikh. 2022. Multiface: A Dataset for Neural Face Rendering. arxiv:https:\/\/arXiv.org\/abs\/2207.11243\u00a0[cs.CV]"},{"key":"e_1_3_3_3_33_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00306"},{"key":"e_1_3_3_3_34_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00068"},{"key":"e_1_3_3_3_35_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00669"},{"key":"e_1_3_3_3_36_1","unstructured":"Yinglin Zheng Hao Yang Ting Zhang Jianmin Bao Dongdong Chen Yangyu Huang Lu Yuan Dong Chen Ming Zeng and Fang Wen. 2021. General Facial Representation Learning in a Visual-Linguistic Manner. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2112.03109 (2021)."},{"key":"e_1_3_3_3_37_1","doi-asserted-by":"crossref","unstructured":"Katja Zibrek Sean Martin and Rachel McDonnell. 2019. Is photorealism important for perception of expressive virtual humans in virtual reality? ACM Transactions on Applied Perception (TAP) 16 3 (2019) 1\u201319.","DOI":"10.1145\/3349609"}],"event":{"name":"SIGGRAPH Conference Papers '25: Special Interest Group on Computer Graphics and Interactive Techniques Conference Conference Papers","location":"Vancouver BC Canada","acronym":"SIGGRAPH Conference Papers '25","sponsor":["SIGGRAPH ACM Special Interest Group on Computer Graphics and Interactive Techniques"]},"container-title":["Proceedings of the Special Interest Group on Computer Graphics and Interactive Techniques Conference Conference Papers"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3721238.3730653","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,20]],"date-time":"2026-03-20T14:52:13Z","timestamp":1774018333000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3721238.3730653"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,27]]},"references-count":36,"alternative-id":["10.1145\/3721238.3730653","10.1145\/3721238"],"URL":"https:\/\/doi.org\/10.1145\/3721238.3730653","relation":{},"subject":[],"published":{"date-parts":[[2025,7,27]]},"assertion":[{"value":"2025-07-27","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}