{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,6]],"date-time":"2026-01-06T13:07:55Z","timestamp":1767704875704,"version":"3.40.3"},"publisher-location":"Cham","reference-count":31,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030955304"},{"type":"electronic","value":"9783030955311"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-030-95531-1_5","type":"book-chapter","created":{"date-parts":[[2022,2,9]],"date-time":"2022-02-09T14:02:49Z","timestamp":1644415369000},"page":"61-76","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Reconstructing Facial Expressions of HMD Users for Avatars in VR"],"prefix":"10.1007","author":[{"given":"Christian Felix","family":"Purps","sequence":"first","affiliation":[]},{"given":"Simon","family":"Janzer","sequence":"additional","affiliation":[]},{"given":"Matthias","family":"W\u00f6lfel","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,2,10]]},"reference":[{"key":"5_CR1","first-page":"1","volume-title":"Bodily Communication","author":"M Argyle","year":"1986","unstructured":"Argyle, M.: Bodily Communication, 2nd edn., pp. 1\u2013111. Routledge, London (1986)","edition":"2"},{"key":"5_CR2","doi-asserted-by":"publisher","unstructured":"Hepperle, D., Purps, C.F., Deuchler, J., W\u00f6lfel, M.: Aspects of visual avatar appearance: self-representation, display type, and uncanny valley. Vis. Comput. (2021). https:\/\/doi.org\/10.1007\/s00371-021-02151-0","DOI":"10.1007\/s00371-021-02151-0"},{"key":"5_CR3","doi-asserted-by":"publisher","first-page":"4129","DOI":"10.1109\/TVCG.2021.3106480","volume":"27","author":"K Yu","year":"2021","unstructured":"Yu, K., Gorbachev, G., Eck, U., Pankratz, F., Navab, N., Roth, D.: Avatars for teleconsultation: effects of avatar embodiment techniques on user perception in 3D asymmetric telepresence. IEEE Trans. Vis. Comput. Graph. 27, 4129\u20134139 (2021)","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"5_CR4","doi-asserted-by":"crossref","unstructured":"Yan, Y., Lu, K., Xue, J., Gao, P., Lyu, J.: FEAFA: a well-annotated dataset for facial expression analysis and 3D facial animation, April 2019. arXiv:1904.01509 [cs, eess, stat]","DOI":"10.1109\/ICMEW.2019.0-104"},{"key":"5_CR5","unstructured":"Wei, X., Zhu, Z., Yin, L., Ji, Q.: A real time face tracking and animation system. In: 2004 Conference on Computer Vision and Pattern Recognition Workshop, pp. 71\u201371, June 2004"},{"key":"5_CR6","doi-asserted-by":"publisher","first-page":"1274","DOI":"10.1109\/TVCG.2020.3013876","volume":"28","author":"J Zhang","year":"2020","unstructured":"Zhang, J., Chen, K., Zheng, J.: Facial expression retargeting from human to avatar made easy. IEEE Trans. Vis. Comput. Graph. 28, 1274\u20131287 (2020). Conference Name: IEEE Transactions on Visualization and Computer Graphics","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"5_CR7","doi-asserted-by":"crossref","unstructured":"Brito, C.J.D.S., Mitchell, K.: Recycling a landmark dataset for real-time facial capture and animation with low cost HMD integrated cameras. In: The 17th International Conference on Virtual-Reality Continuum and its Applications in Industry, VRCAI 2019, pp. 1\u201310. Association for Computing Machinery, New York (2019)","DOI":"10.1145\/3359997.3365690"},{"key":"5_CR8","doi-asserted-by":"crossref","unstructured":"Hickson, S., Dufour, N., Sud, A., Kwatra, V., Essa, I.: Eyemotion: classifying facial expressions in VR using eye-tracking cameras. In: 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 1626\u20131635 (2019). ISSN: 1550\u20135790","DOI":"10.1109\/WACV.2019.00178"},{"issue":"3","key":"5_CR9","doi-asserted-by":"publisher","first-page":"730","DOI":"10.1109\/TMM.2019.2933338","volume":"22","author":"J Lou","year":"2020","unstructured":"Lou, J., et al.: Realistic facial expression reconstruction for VR HMD users. IEEE Trans. Multimedia 22(3), 730\u2013743 (2020). Conference Name: IEEE Transactions on Multimedia","journal-title":"IEEE Trans. Multimedia"},{"key":"5_CR10","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1016\/j.imavis.2016.01.002","volume":"47","author":"C Sagonas","year":"2016","unstructured":"Sagonas, C., Antonakos, E., Tzimiropoulos, G., Zafeiriou, S., Pantic, M.: 300 faces in-the-wild challenge: database and results. Image Vis. Comput. 47, 3\u201318 (2016)","journal-title":"Image Vis. Comput."},{"key":"5_CR11","doi-asserted-by":"crossref","unstructured":"Ekman, P., Rosenberg, E.L.: What the Face Reveals: Basic and Applied Studies of Spontaneous Expression Using the Facial Action Coding System (FACS). Oxford University Press, Oxford (1997). Google-Books-ID: KVmZKGZfmfEC","DOI":"10.1093\/oso\/9780195104462.001.0001"},{"key":"5_CR12","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"232","DOI":"10.1007\/978-3-030-34110-7_20","volume-title":"Image and Graphics","author":"V Cuculo","year":"2019","unstructured":"Cuculo, V., D\u2019Amelio, A.: OpenFACS: an open source FACS-based 3D face animation system. In: Zhao, Y., Barnes, N., Chen, B., Westermann, R., Kong, X., Lin, C. (eds.) ICIG 2019. LNCS, vol. 11902, pp. 232\u2013242. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-34110-7_20"},{"key":"5_CR13","doi-asserted-by":"crossref","unstructured":"Valstar, M.F., et al.: FERA 2015 - second facial expression recognition and analysis challenge. In: 2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), vol. 06, pp. 1\u20138, May 2015","DOI":"10.1109\/FG.2015.7284874"},{"key":"5_CR14","doi-asserted-by":"crossref","unstructured":"Mavadati, M., Sanger, P., Mahoor, M.H.: Extended DISFA dataset: investigating posed and spontaneous facial expressions, pp. 1\u20138 (2016)","DOI":"10.1109\/CVPRW.2016.182"},{"key":"5_CR15","doi-asserted-by":"crossref","unstructured":"Lucey, P., Cohn, J.F., Kanade, T., Saragih, J., Ambadar, Z., Matthews, I.: The extended cohn-kanade dataset (CK+): a complete dataset for action unit and emotion-specified expression. In: 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition - Workshops, pp. 94\u2013101, June 2010. ISSN: 2160\u20137516","DOI":"10.1109\/CVPRW.2010.5543262"},{"issue":"1","key":"5_CR16","doi-asserted-by":"publisher","first-page":"351","DOI":"10.3758\/BRM.42.1.351","volume":"42","author":"NC Ebner","year":"2010","unstructured":"Ebner, N.C., Riediger, M., Lindenberger, U.: FACES-a database of facial expressions in young, middle-aged, and older women and men: development and validation. Behav. Res. Methods 42(1), 351\u2013362 (2010). https:\/\/doi.org\/10.3758\/BRM.42.1.351","journal-title":"Behav. Res. Methods"},{"key":"5_CR17","doi-asserted-by":"crossref","unstructured":"Suresh, K., Palangappa, M., Bhuvan, S.: Face mask detection by using optimistic convolutional neural network. In: 2021 6th International Conference on Inventive Computation Technologies (ICICT), pp. 1084\u20131089 (2021)","DOI":"10.1109\/ICICT50816.2021.9358653"},{"key":"5_CR18","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition arXiv:1409.1556, April 2015"},{"key":"5_CR19","doi-asserted-by":"crossref","unstructured":"Zhihong, C., Hebin, Z., Yanbo, W., Binyan, L., Yu, L.: A vision-based robotic grasping system using deep learning for garbage sorting. In: 2017 36th Chinese Control Conference (CCC), pp. 11 223\u201311 226, July 2017. ISSN: 1934\u20131768","DOI":"10.23919\/ChiCC.2017.8029147"},{"issue":"60","key":"5_CR20","first-page":"1755","volume":"10","author":"DE King","year":"2009","unstructured":"King, D.E.: Dlib-ml: a machine learning toolkit. J. Mach. Learn. Res. 10(60), 1755\u20131758 (2009)","journal-title":"J. Mach. Learn. Res."},{"issue":"2","key":"5_CR21","doi-asserted-by":"publisher","first-page":"19","DOI":"10.1109\/34.908962","volume":"23","author":"Y-L Tian","year":"2001","unstructured":"Tian, Y.-L., Kanade, T., Cohn, J.F.: Recognizing action units for facial expression analysis. IEEE Trans. Pattern Anal. Mach. Intell. 23(2), 19 (2001)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"5_CR22","doi-asserted-by":"crossref","unstructured":"Onizuka, H., Thomas, D., Uchiyama, H., Taniguchi, R.-I.: Landmark-guided deformation transfer of template facial expressions for automatic generation of avatar blendshapes. In: 2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Korea (South), pp. 2100\u20132108. IEEE (2019)","DOI":"10.1109\/ICCVW.2019.00265"},{"issue":"6","key":"5_CR23","doi-asserted-by":"publisher","first-page":"681","DOI":"10.1109\/34.927467","volume":"23","author":"T Cootes","year":"2001","unstructured":"Cootes, T., Edwards, G., Taylor, C.: Active appearance models. IEEE Trans. Pattern Anal. Mach. Intell. 23(6), 681\u2013685 (2001). Conference Name: IEEE Transactions on Pattern Analysis and Machine Intelligence","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"4","key":"5_CR24","doi-asserted-by":"publisher","first-page":"153:1","DOI":"10.1145\/3072959.3073664","volume":"36","author":"A-E Ichim","year":"2017","unstructured":"Ichim, A.-E., Kadle\u010dek, P., Kavan, L., Pauly, M.: Phace: physics-based face modeling and animation. ACM Trans. Graph. 36(4), 153:1-153:14 (2017)","journal-title":"ACM Trans. Graph."},{"key":"5_CR25","unstructured":"Lewis, J.P., Anjyo, K., Rhee, T., Zhang, M., Pighin, F., Deng, Z.: Practice and Theory of Blendshape Facial Models, p. 23 (2014)"},{"issue":"4","key":"5_CR26","doi-asserted-by":"publisher","first-page":"1181","DOI":"10.1111\/j.1467-8659.2011.01976.x","volume":"30","author":"E d\u2019Eon","year":"2011","unstructured":"d\u2019Eon, E., Francois, G., Hill, M., Letteri, J., Aubry, J.-M.: An energy-conserving hair reflectance model. Comput. Graph. Forum 30(4), 1181\u20131187 (2011)","journal-title":"Comput. Graph. Forum"},{"issue":"12","key":"5_CR27","doi-asserted-by":"publisher","first-page":"2830","DOI":"10.1016\/j.neuropsychologia.2012.08.010","volume":"50","author":"C Blais","year":"2012","unstructured":"Blais, C., Roy, C., Fiset, D., Arguin, M., Gosselin, F.: The eyes are not the window to basic emotions. Neuropsychologia 50(12), 2830\u20132838 (2012)","journal-title":"Neuropsychologia"},{"issue":"6","key":"5_CR28","doi-asserted-by":"publisher","first-page":"309","DOI":"10.1080\/00221325.2017.1361377","volume":"178","author":"M Guarnera","year":"2017","unstructured":"Guarnera, M., Hichy, Z., Cascio, M., Carrubba, S., Buccheri, S.L.: Facial expressions and the ability to recognize emotions from the eyes or mouth: a comparison between children and adults. J. Genet. Psychol. 178(6), 309\u2013318 (2017). https:\/\/doi.org\/10.1080\/00221325.2017.1361377","journal-title":"J. Genet. Psychol."},{"key":"5_CR29","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"649","DOI":"10.1007\/978-3-030-24289-3_48","volume-title":"Computational Science and Its Applications","author":"G Biondi","year":"2019","unstructured":"Biondi, G., Franzoni, V., Gervasi, O., Perri, D.: An approach for improving automatic mouth emotion recognition. In: Misra, S., et al. (eds.) ICCSA 2019. LNCS, vol. 11619, pp. 649\u2013664. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-24289-3_48"},{"key":"5_CR30","doi-asserted-by":"crossref","unstructured":"Dinculescu, A.: Automatic identification of anthropological face landmarks for emotion detection. In: 2019 9th International Conference on Recent Advances in Space Technologies (RAST), pp. 585\u2013590 (2019)","DOI":"10.1109\/RAST.2019.8767835"},{"key":"5_CR31","doi-asserted-by":"crossref","unstructured":"W\u00f6lfel, M., Hepperle, D., Purps, C.F., Deuchler, J., Hettmann, W.: Entering a new dimension in virtual reality research: an overview of existing toolkits, their features and challenges. In: International Conference on Cyberworlds (CW) (2021)","DOI":"10.1109\/CW52790.2021.00038"}],"container-title":["Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering","ArtsIT, Interactivity and Game Creation"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-95531-1_5","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,11,17]],"date-time":"2023-11-17T06:49:54Z","timestamp":1700203794000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-95531-1_5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783030955304","9783030955311"],"references-count":31,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-95531-1_5","relation":{},"ISSN":["1867-8211","1867-822X"],"issn-type":[{"type":"print","value":"1867-8211"},{"type":"electronic","value":"1867-822X"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"10 February 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ArtsIT","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on ArtsIT, Interactivity and Game Creation","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2 December 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"3 December 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"10","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"artsit2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/artsit.org","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Confy +","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"57","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"31","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"54% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}