{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,26]],"date-time":"2026-03-26T15:58:42Z","timestamp":1774540722266,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":64,"publisher":"ACM","license":[{"start":{"date-parts":[[2017,7,28]],"date-time":"2017-07-28T00:00:00Z","timestamp":1501200000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2017,7,28]]},"DOI":"10.1145\/3099564.3099581","type":"proceedings-article","created":{"date-parts":[[2017,7,26]],"date-time":"2017-07-26T20:05:41Z","timestamp":1501099541000},"page":"1-10","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":75,"title":["Production-level facial performance capture using deep convolutional neural networks"],"prefix":"10.1145","author":[{"given":"Samuli","family":"Laine","sequence":"first","affiliation":[{"name":"NVIDIA"}]},{"given":"Tero","family":"Karras","sequence":"additional","affiliation":[{"name":"NVIDIA"}]},{"given":"Timo","family":"Aila","sequence":"additional","affiliation":[{"name":"NVIDIA"}]},{"given":"Antti","family":"Herva","sequence":"additional","affiliation":[{"name":"Remedy Entertainment"}]},{"given":"Shunsuke","family":"Saito","sequence":"additional","affiliation":[{"name":"University of Southern California"}]},{"given":"Ronald","family":"Yu","sequence":"additional","affiliation":[{"name":"University of Southern California"}]},{"given":"Hao","family":"Li","sequence":"additional","affiliation":[{"name":"University of Southern California"}]},{"given":"Jaakko","family":"Lehtinen","sequence":"additional","affiliation":[{"name":"Aalto University"}]}],"member":"320","published-online":{"date-parts":[[2017,7,28]]},"reference":[{"key":"e_1_3_2_2_1_1","doi-asserted-by":"publisher","DOI":"10.1145\/1667239.1667251"},{"key":"e_1_3_2_2_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/2010324.1964970"},{"key":"e_1_3_2_2_3_1","doi-asserted-by":"publisher","DOI":"10.1145\/1553374.1553380"},{"key":"e_1_3_2_2_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/2766924"},{"key":"e_1_3_2_2_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/2485895.2485915"},{"key":"e_1_3_2_2_6_1","unstructured":"Bernd Bickel Manuel Lang Mario Botsch Miguel A. Otaduy and Markus Gross. 2008. Pose-space animation and transfer of facial details. In ACM SCA. 57--66.  Bernd Bickel Manuel Lang Mario Botsch Miguel A. Otaduy and Markus Gross. 2008. Pose-space animation and transfer of facial details. In ACM SCA . 57--66."},{"key":"e_1_3_2_2_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/311535.311556"},{"key":"e_1_3_2_2_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/1198555.1198596"},{"key":"e_1_3_2_2_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/2461912.2461976"},{"key":"e_1_3_2_2_10_1","doi-asserted-by":"publisher","DOI":"10.1145\/1778765.1778778"},{"key":"e_1_3_2_2_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/311535.311537"},{"key":"e_1_3_2_2_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/2766943"},{"key":"e_1_3_2_2_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/2601097.2601204"},{"key":"e_1_3_2_2_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/2461912.2462012"},{"key":"e_1_3_2_2_15_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-4-431-66911-1_13"},{"key":"e_1_3_2_2_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/34.927467"},{"key":"e_1_3_2_2_17_1","volume-title":"S\u00f8ren Kaae S\u00f8nderby, et al","author":"Dieleman Sander","year":"2015","unstructured":"Sander Dieleman , Jan Schl\u00fcter , Colin Raffel , Eben Olson , S\u00f8ren Kaae S\u00f8nderby, et al . 2015 . Lasagne : First release. (2015). Sander Dieleman, Jan Schl\u00fcter, Colin Raffel, Eben Olson, S\u00f8ren Kaae S\u00f8nderby, et al. 2015. Lasagne: First release. (2015)."},{"key":"e_1_3_2_2_18_1","unstructured":"Dimensional Imaging. 2016. DI4D PRO System. http:\/\/www.di4d.com\/systems\/di4d-pro-system\/. (2016).  Dimensional Imaging. 2016. DI4D PRO System. http:\/\/www.di4d.com\/systems\/di4d-pro-system\/. (2016)."},{"key":"e_1_3_2_2_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/2897824.2925984"},{"key":"e_1_3_2_2_20_1","doi-asserted-by":"crossref","unstructured":"P. Ekman and W. Friesen. 1978. The Facial Action Coding System: A Technique for the Measurement of Facial Movement. Consulting Psychologists Press Palo Alto.  P. Ekman and W. Friesen. 1978. The Facial Action Coding System: A Technique for the Measurement of Facial Movement . Consulting Psychologists Press Palo Alto.","DOI":"10.1037\/t27734-000"},{"key":"e_1_3_2_2_21_1","volume-title":"Proc. Computer Vision and Pattern Recognition (CVPR).","author":"Furukawa Yasutaka","year":"2009","unstructured":"Yasutaka Furukawa and Jean Ponce . 2009 . Dense 3D motion capture for human faces . In Proc. Computer Vision and Pattern Recognition (CVPR). Yasutaka Furukawa and Jean Ponce. 2009. Dense 3D motion capture for human faces. In Proc. Computer Vision and Pattern Recognition (CVPR)."},{"key":"e_1_3_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.1111\/j.1467-8659.2011.01888.x"},{"key":"e_1_3_2_2_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/2638549"},{"key":"e_1_3_2_2_24_1","article-title":"Corrective 3D Reconstruction of Lips from Monocular Video","volume":"35","author":"Garrido Pablo","year":"2016","unstructured":"Pablo Garrido , Michael Zollh\u00f6fer , Chenglei Wu , Derek Bradley , Patrick P\u00e9rez , Thabo Beeler , and Christian Theobalt . 2016 . Corrective 3D Reconstruction of Lips from Monocular Video . ACM Trans. Graph. 35 , 6 (2016), 219:1--219:11. Pablo Garrido, Michael Zollh\u00f6fer, Chenglei Wu, Derek Bradley, Patrick P\u00e9rez, Thabo Beeler, and Christian Theobalt. 2016. Corrective 3D Reconstruction of Lips from Monocular Video. ACM Trans. Graph. 35, 6 (2016), 219:1--219:11.","journal-title":"ACM Trans. Graph."},{"key":"e_1_3_2_2_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/280814.280822"},{"key":"e_1_3_2_2_26_1","volume-title":"Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification. CoRR abs\/1502.01852","author":"He Kaiming","year":"2015","unstructured":"Kaiming He , Xiangyu Zhang , Shaoqing Ren , and Jian Sun . 2015. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification. CoRR abs\/1502.01852 ( 2015 ). http:\/\/arxiv.org\/abs\/1502.01852 Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2015. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification. CoRR abs\/1502.01852 (2015). http:\/\/arxiv.org\/abs\/1502.01852"},{"key":"e_1_3_2_2_27_1","volume-title":"Unconstrained Realtime Facial Performance Capture","author":"Hsieh Pei-Lun","unstructured":"Pei-Lun Hsieh , Chongyang Ma , Jihun Yu , and Hao Li. 2015a. Unconstrained Realtime Facial Performance Capture . In IEEE CVPR. 1675--1683. Pei-Lun Hsieh, Chongyang Ma, Jihun Yu, and Hao Li. 2015a. Unconstrained Realtime Facial Performance Capture. In IEEE CVPR. 1675--1683."},{"key":"e_1_3_2_2_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298776"},{"key":"e_1_3_2_2_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/2766974"},{"key":"e_1_3_2_2_30_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.241"},{"key":"e_1_3_2_2_31_1","volume-title":"Kingma and Jimmy Ba","author":"Diederik","year":"2014","unstructured":"Diederik P. Kingma and Jimmy Ba . 2014 . Adam : A Method for Stochastic Optimization. CoRR abs\/1412.6980 (2014). http:\/\/arxiv.org\/abs\/1412.6980 Diederik P. Kingma and Jimmy Ba. 2014. Adam: A Method for Stochastic Optimization. CoRR abs\/1412.6980 (2014). http:\/\/arxiv.org\/abs\/1412.6980"},{"key":"e_1_3_2_2_32_1","volume-title":"Hinton","author":"Krizhevsky Alex","year":"2012","unstructured":"Alex Krizhevsky , Ilya Sutskever , and Geoffrey E . Hinton . 2012 . ImageNet Classification with Deep Convolutional Neural Networks. In Proc. NIPS. 1097--1105. Alex Krizhevsky, Ilya Sutskever, and Geoffrey E. Hinton. 2012. ImageNet Classification with Deep Convolutional Neural Networks. In Proc. NIPS. 1097--1105."},{"key":"e_1_3_2_2_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/1833349.1778769"},{"key":"e_1_3_2_2_34_1","doi-asserted-by":"publisher","DOI":"10.1145\/2461912.2462019"},{"key":"e_1_3_2_2_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/2816795.2818122"},{"key":"e_1_3_2_2_36_1","volume-title":"Audiovisual speech synthesis: An overview of the state-of-the-art. Speech Communication 66 (2","author":"Mattheyses Wesley","year":"2015","unstructured":"Wesley Mattheyses and Werner Verhelst . 2015. Audiovisual speech synthesis: An overview of the state-of-the-art. Speech Communication 66 (2 2015 ), 182--217. Wesley Mattheyses and Werner Verhelst. 2015. Audiovisual speech synthesis: An overview of the state-of-the-art. Speech Communication 66 (2 2015), 182--217."},{"key":"e_1_3_2_2_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/2980179.2980252"},{"key":"e_1_3_2_2_38_1","unstructured":"Photoscan. 2014. Agisoft. (2014). http:\/\/www.agisoft.com\/  Photoscan. 2014. Agisoft. (2014). http:\/\/www.agisoft.com\/"},{"key":"e_1_3_2_2_39_1","unstructured":"Fred Pighin and J. P. Lewis. 2006. Performance-driven facial animation. In ACM SIGGRAPH Courses (SIGGRAPH '06).  Fred Pighin and J. P. Lewis. 2006. Performance-driven facial animation. In ACM SIGGRAPH Courses (SIGGRAPH '06)."},{"key":"e_1_3_2_2_40_1","volume-title":"ACM SIGGRAPH 2006 Courses.","author":"Pughin F.","unstructured":"F. Pughin and J. P. Lewis . 2006. Performance-driven facial animation . In ACM SIGGRAPH 2006 Courses. F. Pughin and J. P. Lewis. 2006. Performance-driven facial animation. In ACM SIGGRAPH 2006 Courses."},{"key":"e_1_3_2_2_41_1","volume-title":"Real-Time Facial Segmentation and Performance Capture from RGB Input. CoRR abs\/1604.02647","author":"Saito Shunsuke","year":"2016","unstructured":"Shunsuke Saito , Tianye Li , and Hao Li. 2016. Real-Time Facial Segmentation and Performance Capture from RGB Input. CoRR abs\/1604.02647 ( 2016 ). http:\/\/arxiv.org\/abs\/1604.02647 Shunsuke Saito, Tianye Li, and Hao Li. 2016. Real-Time Facial Segmentation and Performance Capture from RGB Input. CoRR abs\/1604.02647 (2016). http:\/\/arxiv.org\/abs\/1604.02647"},{"key":"e_1_3_2_2_42_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-010-0380-4"},{"key":"e_1_3_2_2_43_1","doi-asserted-by":"publisher","DOI":"10.1145\/2661229.2661290"},{"key":"e_1_3_2_2_44_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDAR.2003.1227801"},{"key":"e_1_3_2_2_45_1","volume-title":"Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv preprint arXiv: 1409.1556","author":"Simonyan Karen","year":"2014","unstructured":"Karen Simonyan and Andrew Zisserman . 2014. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv preprint arXiv: 1409.1556 ( 2014 ). Karen Simonyan and Andrew Zisserman. 2014. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv preprint arXiv: 1409.1556 (2014)."},{"key":"e_1_3_2_2_46_1","volume-title":"Ried-miller","author":"Springenberg Jost Tobias","year":"2014","unstructured":"Jost Tobias Springenberg , Alexey Dosovitskiy , Thomas Brox , and Martin A . Ried-miller . 2014 . Striving for Simplicity: The All Convolutional Net . arXiv preprint arXiv:1412.6806 (2014). Jost Tobias Springenberg, Alexey Dosovitskiy, Thomas Brox, and Martin A. Ried-miller. 2014. Striving for Simplicity: The All Convolutional Net. arXiv preprint arXiv:1412.6806 (2014)."},{"key":"e_1_3_2_2_47_1","doi-asserted-by":"publisher","DOI":"10.5555\/2627435.2670313"},{"key":"e_1_3_2_2_48_1","volume-title":"Proc. SCA. 275--284","author":"Taylor Sarah L.","year":"2012","unstructured":"Sarah L. Taylor , Moshe Mahler , Barry-John Theobald , and Iain Matthews . 2012 . Dynamic Units of Visual Speech . In Proc. SCA. 275--284 . Sarah L. Taylor, Moshe Mahler, Barry-John Theobald, and Iain Matthews. 2012. Dynamic Units of Visual Speech. In Proc. SCA. 275--284."},{"key":"e_1_3_2_2_49_1","volume-title":"Theano: A Python framework for fast computation of mathematical expressions. arXiv e-prints abs\/1605.02688 (May","author":"Team Theano Development","year":"2016","unstructured":"Theano Development Team . 2016 . Theano: A Python framework for fast computation of mathematical expressions. arXiv e-prints abs\/1605.02688 (May 2016). Theano Development Team. 2016. Theano: A Python framework for fast computation of mathematical expressions. arXiv e-prints abs\/1605.02688 (May 2016)."},{"key":"e_1_3_2_2_50_1","doi-asserted-by":"publisher","DOI":"10.1145\/2816795.2818056"},{"key":"e_1_3_2_2_51_1","volume-title":"Proc. Computer Vision and Pattern Recognition (CVPR).","author":"Thies J.","unstructured":"J. Thies , M. Zollh\u00f6fer , M. Stamminger , C. Theobalt , and M. Nie\u00dfner . 2016. Face2Face: Real-time Face Capture and Reenactment of RGB Videos . In Proc. Computer Vision and Pattern Recognition (CVPR). J. Thies, M. Zollh\u00f6fer, M. Stamminger, C. Theobalt, and M. Nie\u00dfner. 2016. Face2Face: Real-time Face Capture and Reenactment of RGB Videos. In Proc. Computer Vision and Pattern Recognition (CVPR)."},{"key":"e_1_3_2_2_52_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-011-0464-9"},{"key":"e_1_3_2_2_53_1","doi-asserted-by":"publisher","DOI":"10.1145\/2366145.2366206"},{"key":"e_1_3_2_2_54_1","doi-asserted-by":"publisher","DOI":"10.1145\/1186822.1073209"},{"key":"e_1_3_2_2_55_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2003.819861"},{"key":"e_1_3_2_2_56_1","doi-asserted-by":"publisher","DOI":"10.1145\/2010324.1964972"},{"key":"e_1_3_2_2_57_1","doi-asserted-by":"publisher","DOI":"10.1145\/1599470.1599472"},{"key":"e_1_3_2_2_58_1","doi-asserted-by":"publisher","DOI":"10.1145\/1599470.1599472"},{"key":"e_1_3_2_2_59_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.gmod.2013.10.002"},{"key":"e_1_3_2_2_60_1","doi-asserted-by":"publisher","DOI":"10.1145\/97880.97906"},{"key":"e_1_3_2_2_61_1","doi-asserted-by":"publisher","DOI":"10.1145\/97880.97906"},{"key":"e_1_3_2_2_62_1","doi-asserted-by":"publisher","DOI":"10.1145\/2897824.2925882"},{"key":"e_1_3_2_2_63_1","doi-asserted-by":"publisher","DOI":"10.1145\/1015706.1015759"},{"key":"e_1_3_2_2_64_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7299081"}],"event":{"name":"SCA '17: The ACM SIGGRAPH \/ Eurographics Symposium on Computer Animation","location":"Los Angeles California","acronym":"SCA '17","sponsor":["SIGGRAPH ACM Special Interest Group on Computer Graphics and Interactive Techniques","EUROGRAPHICS The European Association for Computer Graphics"]},"container-title":["Proceedings of the ACM SIGGRAPH \/ Eurographics Symposium on Computer Animation"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3099564.3099581","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3099564.3099581","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T03:37:18Z","timestamp":1750217838000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3099564.3099581"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2017,7,28]]},"references-count":64,"alternative-id":["10.1145\/3099564.3099581","10.1145\/3099564"],"URL":"https:\/\/doi.org\/10.1145\/3099564.3099581","relation":{},"subject":[],"published":{"date-parts":[[2017,7,28]]},"assertion":[{"value":"2017-07-28","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}