{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,23]],"date-time":"2026-04-23T07:58:33Z","timestamp":1776931113449,"version":"3.51.2"},"publisher-location":"New York, NY, USA","reference-count":46,"publisher":"ACM","funder":[{"DOI":"10.13039\/501100006465","name":"Korea Creative Content Agency","doi-asserted-by":"publisher","award":["FP00022751"],"award-info":[{"award-number":["FP00022751"]}],"id":[{"id":"10.13039\/501100006465","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,11,12]]},"DOI":"10.1145\/3756884.3766013","type":"proceedings-article","created":{"date-parts":[[2025,12,5]],"date-time":"2025-12-05T07:52:48Z","timestamp":1764921168000},"page":"1-10","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Detecting Natural Emotions in Virtual Reality Through Facial Movement Analysis"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0006-7482-2228","authenticated-orcid":false,"given":"Terens","family":"Tare","sequence":"first","affiliation":[{"name":"Virginia Commonwealth University, Richmond, Virginia, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-7143-2355","authenticated-orcid":false,"given":"Rahat Rizvi","family":"Rahman","sequence":"additional","affiliation":[{"name":"Virginia Commonwealth University, Richmond, Virginia, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-3473-5757","authenticated-orcid":false,"given":"Hee Yun","family":"Choi","sequence":"additional","affiliation":[{"name":"Korea National University of Arts, Seoul, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-6479-3550","authenticated-orcid":false,"given":"Joonghyo","family":"Lim","sequence":"additional","affiliation":[{"name":"Korea National University of Arts, Seoul, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-5705-8617","authenticated-orcid":false,"given":"Go Eun","family":"Lee","sequence":"additional","affiliation":[{"name":"Korea National University of Arts, Seoul, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4958-353X","authenticated-orcid":false,"given":"Seungmoo","family":"Lee","sequence":"additional","affiliation":[{"name":"Korea National University of Arts, Seoul, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0879-2535","authenticated-orcid":false,"given":"Chungyean","family":"Cho","sequence":"additional","affiliation":[{"name":"Korea National University of Arts, Seoul, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7799-2026","authenticated-orcid":false,"given":"Kostadin","family":"Damevski","sequence":"additional","affiliation":[{"name":"Virginia Commonwealth University, Richmond, Virginia, USA"}]}],"member":"320","published-online":{"date-parts":[[2025,12,4]]},"reference":[{"key":"e_1_3_3_1_2_2","doi-asserted-by":"crossref","unstructured":"R\u00a0Rashmi Adyapady and B Annappa. 2023. A comprehensive review of facial expression recognition techniques. Multimedia Systems 29 1 (2023) 73\u2013103.","DOI":"10.1007\/s00530-022-00984-w"},{"key":"e_1_3_3_1_3_2","doi-asserted-by":"publisher","DOI":"10.1109\/FGR.2006.55"},{"key":"e_1_3_3_1_4_2","doi-asserted-by":"publisher","DOI":"10.1109\/FG57933.2023.10042673"},{"key":"e_1_3_3_1_5_2","doi-asserted-by":"crossref","unstructured":"Adil Boughida Mohamed\u00a0Nadjib Kouahla and Yacine Lafifi. 2022. A novel approach for facial expression recognition based on Gabor filters and genetic algorithm. Evolving Systems 13 2 (2022) 331\u2013345.","DOI":"10.1007\/s12530-021-09393-2"},{"key":"e_1_3_3_1_6_2","doi-asserted-by":"crossref","unstructured":"Jeffrey\u00a0F Cohn Zara Ambadar and Paul Ekman. 2007. Observer-based measurement of facial expression with the Facial Action Coding System. The handbook of emotion elicitation and assessment 1 3 (2007) 203\u2013221.","DOI":"10.1093\/oso\/9780195169157.003.0014"},{"key":"e_1_3_3_1_7_2","doi-asserted-by":"publisher","DOI":"10.1109\/CoG57401.2023.10333160"},{"key":"e_1_3_3_1_8_2","doi-asserted-by":"publisher","DOI":"10.1145\/3136755.3143004"},{"key":"e_1_3_3_1_9_2","doi-asserted-by":"publisher","DOI":"10.1145\/2993148.2997638"},{"key":"e_1_3_3_1_10_2","doi-asserted-by":"publisher","DOI":"10.1145\/2818346.2829994"},{"key":"e_1_3_3_1_11_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-33712-3_38"},{"key":"e_1_3_3_1_12_2","volume-title":"Facial action coding system: Investigator\u2019s guide","author":"Ekman Paul","year":"1978","unstructured":"Paul Ekman and Wallace\u00a0V Friesen. 1978. Facial action coding system: Investigator\u2019s guide. Consulting Psychologists Press."},{"key":"e_1_3_3_1_13_2","doi-asserted-by":"publisher","unstructured":"Paul Ekman and Wallace\u00a0V. Friesen. 1982. Felt false and miserable smiles. Journal of Nonverbal Behavior 6 4 (1982) 238\u2013252. arXiv:10.1007\/BF0098719110.1007\/BF00987191","DOI":"10.1007\/BF00987191"},{"key":"e_1_3_3_1_14_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-42051-1_16"},{"key":"e_1_3_3_1_15_2","doi-asserted-by":"crossref","unstructured":"Ursula Hess Reginald\u00a0B Adams\u00a0Jr and Robert\u00a0E Kleck. 2004. Facial appearance gender and emotion expression. Emotion 4 4 (2004) 378.","DOI":"10.1037\/1528-3542.4.4.378"},{"key":"e_1_3_3_1_16_2","doi-asserted-by":"crossref","unstructured":"Shan Jia Shuo Wang Chuanbo Hu Paula\u00a0J Webster and Xin Li. 2021. Detection of genuine and posed facial expressions of emotion: databases and methods. Frontiers in psychology 11 (2021) 580287.","DOI":"10.3389\/fpsyg.2020.580287"},{"key":"e_1_3_3_1_17_2","doi-asserted-by":"publisher","DOI":"10.1109\/ISMAR62088.2024.00030"},{"key":"e_1_3_3_1_18_2","doi-asserted-by":"crossref","unstructured":"Dae\u00a0Hoe Kim Wissam\u00a0J Baddar Jinhyeok Jang and Yong\u00a0Man Ro. 2017. Multi-objective based spatio-temporal feature representation learning robust to expression intensity variations for facial expression recognition. IEEE Transactions on Affective Computing 10 2 (2017) 223\u2013236.","DOI":"10.1109\/TAFFC.2017.2695999"},{"key":"e_1_3_3_1_19_2","doi-asserted-by":"publisher","unstructured":"Maya Lecker Scott Hallock Axel Danielson Maximilien\u00a0Van Aertrickc Merel Kindt and Hillel Aviezer. 2025. Real-life intense fear is communicated through context not facial expressions. Proceedings of the National Academy of Sciences 122 11 (2025) e2414677122. arXiv:https:\/\/www.pnas.org\/doi\/pdf\/10.1073\/pnas.241467712210.1073\/pnas.2414677122","DOI":"10.1073\/pnas.2414677122"},{"key":"e_1_3_3_1_20_2","doi-asserted-by":"crossref","unstructured":"Sze-Teng Liong John See KokSheik Wong and Raphael C-W Phan. 2018. Less is more: Micro-expression recognition from video using apex frame. Signal Processing: Image Communication 62 (2018) 82\u201392.","DOI":"10.1016\/j.image.2017.11.006"},{"key":"e_1_3_3_1_21_2","doi-asserted-by":"crossref","unstructured":"Yuanyuan Liu Xiaohui Yuan Xi Gong Zhong Xie Fang Fang and Zhongwen Luo. 2018. Conditional convolution neural network enhanced random forest for facial expression recognition. Pattern Recognition 84 (2018) 251\u2013261.","DOI":"10.1016\/j.patcog.2018.07.016"},{"key":"e_1_3_3_1_22_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2010.5543262"},{"key":"e_1_3_3_1_23_2","doi-asserted-by":"crossref","unstructured":"Brais Martinez Michel\u00a0F Valstar Bihan Jiang and Maja Pantic. 2017. Automatic analysis of facial actions: A survey. IEEE transactions on affective computing 10 3 (2017) 325\u2013347.","DOI":"10.1109\/TAFFC.2017.2731763"},{"key":"e_1_3_3_1_24_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2016.182"},{"key":"e_1_3_3_1_25_2","doi-asserted-by":"crossref","unstructured":"Shushi Namba Shoko Makihara Russell\u00a0S Kabir Makoto Miyatani and Takashi Nakao. 2017. Spontaneous facial expressions are different from posed facial expressions: Morphological properties and dynamic sequences. Current Psychology 36 3 (2017) 593\u2013605.","DOI":"10.1007\/s12144-016-9448-9"},{"key":"e_1_3_3_1_26_2","unstructured":"Thorben Ortmann Qi Wang and Larissa Putzar. 2024. EmojiHeroVR: A Study on Facial Expression Recognition under Partial Occlusion from Head-Mounted Displays. arxiv:https:\/\/arXiv.org\/abs\/2410.03331\u00a0[cs.CV] https:\/\/arxiv.org\/abs\/2410.03331"},{"key":"e_1_3_3_1_27_2","doi-asserted-by":"publisher","DOI":"10.1109\/AIxVR63409.2025.00048"},{"key":"e_1_3_3_1_28_2","doi-asserted-by":"publisher","unstructured":"Ye Pan Ruisi Zhang Shengran Cheng Shuai Tan Yu Ding Kenny Mitchell and Xubo Yang. 2023. Emotional Voice Puppetry: An Audio-based Facial Animation Approach to Portray Characters with Vivid Emotional Changes. IEEE Transactions on Visualization and Computer Graphics 29 5 (2023) 2528\u20132539. 10.1109\/TVCG.2023.3240104","DOI":"10.1109\/TVCG.2023.3240104"},{"key":"e_1_3_3_1_29_2","doi-asserted-by":"publisher","DOI":"10.1145\/3691573.3691579"},{"key":"e_1_3_3_1_30_2","unstructured":"scikit-learn developers. 2025. Feature Selection. https:\/\/scikit-learn.org\/stable\/modules\/feature_selection.html. Accessed: May 18 2025."},{"key":"e_1_3_3_1_31_2","unstructured":"scikit-learn developers. 2025. LogisticRegression Documentation. https:\/\/scikit-learn.org\/stable\/modules\/generated\/sklearn.linear_model.LogisticRegression.html. Accessed: May 18 2025."},{"key":"e_1_3_3_1_32_2","doi-asserted-by":"crossref","unstructured":"Caifeng Shan Shaogang Gong and Peter\u00a0W McOwan. 2009. Facial expression recognition based on local binary patterns: A comprehensive study. Image and vision Computing 27 6 (2009) 803\u2013816.","DOI":"10.1016\/j.imavis.2008.08.005"},{"key":"e_1_3_3_1_33_2","doi-asserted-by":"crossref","unstructured":"Mohammad Soleymani Jeroen Lichtenauer Thierry Pun and Maja Pantic. 2011. A multimodal database for affect recognition and implicit tagging. IEEE transactions on affective computing 3 1 (2011) 42\u201355.","DOI":"10.1109\/T-AFFC.2011.25"},{"key":"e_1_3_3_1_34_2","doi-asserted-by":"publisher","unstructured":"Barathi Subramanian Jeonghong Kim Mohammed Maray and Anand Paul. 2022. Digital Twin Model: A Real-Time Emotion Recognition System for Personalized Healthcare. IEEE Access 10 (2022) 81155\u201381165. 10.1109\/ACCESS.2022.3193941","DOI":"10.1109\/ACCESS.2022.3193941"},{"key":"e_1_3_3_1_35_2","doi-asserted-by":"crossref","unstructured":"Bo Sun Siming Cao Jun He and Lejun Yu. 2018. Affect recognition from facial movements and body gestures by hierarchical deep spatio-temporal features and fusion strategy. Neural Networks 105 (2018) 36\u201351.","DOI":"10.1016\/j.neunet.2017.11.021"},{"key":"e_1_3_3_1_36_2","doi-asserted-by":"crossref","unstructured":"Ning Sun Qi Li Ruizhi Huan Jixin Liu and Guang Han. 2019. Deep spatial-temporal feature fusion for facial expression recognition in static images. Pattern Recognition Letters 119 (2019) 49\u201361.","DOI":"10.1016\/j.patrec.2017.10.022"},{"key":"e_1_3_3_1_37_2","doi-asserted-by":"crossref","unstructured":"Michel\u00a0F Valstar Marc Mehu Bihan Jiang Maja Pantic and Klaus Scherer. 2012. Meta-analysis of the first facial expression recognition challenge. IEEE Transactions on Systems Man and Cybernetics Part B (Cybernetics) 42 4 (2012) 966\u2013979.","DOI":"10.1109\/TSMCB.2012.2200675"},{"key":"e_1_3_3_1_38_2","doi-asserted-by":"crossref","unstructured":"Alessandro Visconti Davide Calandra Federica Giorgione and Fabrizio Lamberti. 2025. Enhancing Social Experiences in Immersive Virtual Reality with Artificial Facial Mimicry. IEEE Transactions on Visualization and Computer Graphics (2025).","DOI":"10.1109\/TVCG.2025.3549163"},{"key":"e_1_3_3_1_39_2","unstructured":"Yan Wang Shaoqi Yan Yang Liu Wei Song Jing Liu Yang Chang Xinji Mai Xiping Hu Wenqiang Zhang and Zhongxue Gan. 2024. A survey on facial expression recognition of static and dynamic emotions. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2408.15777 (2024)."},{"key":"e_1_3_3_1_40_2","doi-asserted-by":"publisher","unstructured":"Lihang Wen Jianlong Zhou Weidong Huang and Fang Chen. 2022. A Survey of Facial Capture for Virtual Reality. IEEE Access 10 (2022) 6042\u20136052. 10.1109\/ACCESS.2021.3138200","DOI":"10.1109\/ACCESS.2021.3138200"},{"key":"e_1_3_3_1_41_2","doi-asserted-by":"crossref","unstructured":"Siyue Xie and Haifeng Hu. 2018. Facial expression recognition using hierarchical features with deep comprehensive multipatches aggregation convolutional neural networks. IEEE Transactions on Multimedia 21 1 (2018) 211\u2013220.","DOI":"10.1109\/TMM.2018.2844085"},{"key":"e_1_3_3_1_42_2","doi-asserted-by":"publisher","DOI":"10.1109\/PRMVIA58252.2023.00045"},{"key":"e_1_3_3_1_43_2","doi-asserted-by":"crossref","unstructured":"Kaihao Zhang Yongzhen Huang Yong Du and Liang Wang. 2017. Facial expression recognition based on deep evolutional spatial-temporal networks. IEEE Transactions on Image Processing 26 9 (2017) 4193\u20134203.","DOI":"10.1109\/TIP.2017.2689999"},{"key":"e_1_3_3_1_44_2","doi-asserted-by":"crossref","unstructured":"Zhihui Zhang Josep\u00a0M Fort and Lluis Gim\u00e9nez\u00a0Mateu. 2023. Facial expression recognition in virtual reality environments: challenges and opportunities. Frontiers in psychology 14 (2023) 1280136.","DOI":"10.3389\/fpsyg.2023.1280136"},{"key":"e_1_3_3_1_45_2","doi-asserted-by":"crossref","unstructured":"Guoying Zhao and Matti Pietikainen. 2007. Dynamic texture recognition using local binary patterns with an application to facial expressions. IEEE transactions on pattern analysis and machine intelligence 29 6 (2007) 915\u2013928.","DOI":"10.1109\/TPAMI.2007.1110"},{"key":"e_1_3_3_1_46_2","doi-asserted-by":"crossref","unstructured":"Ruicong Zhi Markus Flierl Qiuqi Ruan and W\u00a0Bastiaan Kleijn. 2010. Graph-preserving sparse nonnegative matrix factorization with application to facial expression recognition. IEEE Transactions on Systems Man and Cybernetics Part B (Cybernetics) 41 1 (2010) 38\u201352.","DOI":"10.1109\/TSMCB.2010.2044788"},{"key":"e_1_3_3_1_47_2","doi-asserted-by":"crossref","unstructured":"Lin Zhong Qingshan Liu Peng Yang Junzhou Huang and Dimitris\u00a0N Metaxas. 2014. Learning multiscale active facial patches for expression analysis. IEEE transactions on cybernetics 45 8 (2014) 1499\u20131510.","DOI":"10.1109\/TCYB.2014.2354351"}],"event":{"name":"VRST '25: 31st ACM Symposium on Virtual Reality Software and Technology","location":"Montreal QC Canada","acronym":"VRST '25","sponsor":["SIGCHI ACM Special Interest Group on Computer-Human Interaction","SIGGRAPH ACM Special Interest Group on Computer Graphics and Interactive Techniques"]},"container-title":["Proceedings of the 2025 31st ACM Symposium on Virtual Reality Software and Technology"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3756884.3766013","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,5]],"date-time":"2025-12-05T09:13:27Z","timestamp":1764926007000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766013"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,12]]},"references-count":46,"alternative-id":["10.1145\/3756884.3766013","10.1145\/3756884"],"URL":"https:\/\/doi.org\/10.1145\/3756884.3766013","relation":{},"subject":[],"published":{"date-parts":[[2025,11,12]]},"assertion":[{"value":"2025-12-04","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}