{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T12:26:36Z","timestamp":1776083196652,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":43,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,8,7]],"date-time":"2022-08-07T00:00:00Z","timestamp":1659830400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022,8,7]]},"DOI":"10.1145\/3528233.3530701","type":"proceedings-article","created":{"date-parts":[[2022,7,20]],"date-time":"2022-07-20T13:56:43Z","timestamp":1658325403000},"page":"1-9","source":"Crossref","is-referenced-by-count":24,"title":["NeuralPassthrough: Learned Real-Time View Synthesis for VR"],"prefix":"10.1145","author":[{"given":"Lei","family":"Xiao","sequence":"first","affiliation":[{"name":"Reality Labs Research, Meta, United States of America"}]},{"given":"Salah","family":"Nouri","sequence":"additional","affiliation":[{"name":"Reality Labs Research, Meta, United States of America"}]},{"given":"Joel","family":"Hegland","sequence":"additional","affiliation":[{"name":"Reality Labs Research, Meta, United States of America"}]},{"given":"Alberto Garcia","family":"Garcia","sequence":"additional","affiliation":[{"name":"Reality Labs, Meta, Switzerland"}]},{"given":"Douglas","family":"Lanman","sequence":"additional","affiliation":[{"name":"Reality Labs Research, Meta, United States of America"}]}],"member":"320","published-online":{"date-parts":[[2022,8,7]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"2021. Azure Kinect DK. https:\/\/azure.microsoft.com\/en-us\/services\/kinect-dk\/  2021. Azure Kinect DK. https:\/\/azure.microsoft.com\/en-us\/services\/kinect-dk\/"},{"key":"e_1_3_2_1_2_1","unstructured":"2021. Rift-S VR. https:\/\/www.oculus.com\/rift-s\/features\/  2021. Rift-S VR. https:\/\/www.oculus.com\/rift-s\/features\/"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58542-6_42"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/3386569.3392485"},{"key":"e_1_3_2_1_5_1","volume-title":"Computer graphics forum, Vol.\u00a033","author":"Chapiro Alexandre","unstructured":"Alexandre Chapiro , Simon Heinzle , Tun\u00e7\u00a0Ozan Ayd\u0131n , Steven Poulakos , Matthias Zwicker , Aljosa Smolic , and Markus Gross . 2014. Optimizing stereo-to-multiview conversion for autostereoscopic displays . In Computer graphics forum, Vol.\u00a033 . Wiley Online Library , 63\u201372. Alexandre Chapiro, Simon Heinzle, Tun\u00e7\u00a0Ozan Ayd\u0131n, Steven Poulakos, Matthias Zwicker, Aljosa Smolic, and Markus Gross. 2014. Optimizing stereo-to-multiview conversion for autostereoscopic displays. In Computer graphics forum, Vol.\u00a033. Wiley Online Library, 63\u201372."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/3384540"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/166117.166153"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/2508363.2508376"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00247"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"crossref","unstructured":"Chen Gao Ayush Saraf Johannes Kopf and Jia-Bin Huang. 2021. Dynamic View Synthesis from Dynamic Monocular Video. arXiv preprint arXiv:2105.06468(2021).  Chen Gao Ayush Saraf Johannes Kopf and Jia-Bin Huang. 2021. Dynamic View Synthesis from Dynamic Monocular Video. arXiv preprint arXiv:2105.06468(2021).","DOI":"10.1109\/ICCV48922.2021.00566"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01408"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/237170.237200"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/WACV51458.2022.00009"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/3130800.3130828"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00582"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00583"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/2980179.2980251"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/3386569.3392420"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3414685.3417820"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/237170.237199"},{"key":"e_1_3_2_1_21_1","volume-title":"Raft-stereo: Multilevel recurrent field transforms for stereo matching. arXiv preprint arXiv:2109.07547(2021).","author":"Lipson Lahav","year":"2021","unstructured":"Lahav Lipson , Zachary Teed , and Jia Deng . 2021 . Raft-stereo: Multilevel recurrent field transforms for stereo matching. arXiv preprint arXiv:2109.07547(2021). Lahav Lipson, Zachary Teed, and Jia Deng. 2021. Raft-stereo: Multilevel recurrent field transforms for stereo matching. arXiv preprint arXiv:2109.07547(2021)."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01252-6_6"},{"key":"e_1_3_2_1_23_1","unstructured":"Lingjie Liu Jiatao Gu Kyaw\u00a0Zaw Lin Tat-Seng Chua and Christian Theobalt. 2020. Neural sparse voxel fields. arXiv preprint arXiv:2007.11571(2020).  Lingjie Liu Jiatao Gu Kyaw\u00a0Zaw Lin Tat-Seng Chua and Christian Theobalt. 2020. Neural sparse voxel fields. arXiv preprint arXiv:2007.11571(2020)."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/3386569.3392416"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3272127.3275099"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1145\/3306346.3322980"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58452-8_24"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00548"},{"key":"e_1_3_2_1_29_1","unstructured":"Oculus. 2016. asynchronous spacewarp. https:\/\/www.oculus.com\/blog\/introducing-asw-2-point-0-better-accuracy-lower-latency\/  Oculus. 2016. asynchronous spacewarp. https:\/\/www.oculus.com\/blog\/introducing-asw-2-point-0-better-accuracy-lower-latency\/"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01018"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"crossref","unstructured":"Christian Reiser Songyou Peng Yiyi Liao and Andreas Geiger. 2021. KiloNeRF: Speeding up Neural Radiance Fields with Thousands of Tiny MLPs. arXiv preprint arXiv:2103.13744(2021).  Christian Reiser Songyou Peng Yiyi Liao and Andreas Geiger. 2021. KiloNeRF: Speeding up Neural Radiance Fields with Thousands of Tiny MLPs. arXiv preprint arXiv:2103.13744(2021).","DOI":"10.1109\/ICCV48922.2021.01407"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/280814.280882"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00805"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2012.2214933"},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00026"},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2003.819861"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00749"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3272127.3275032","article-title":"DeepFocus: learned image synthesis for computational displays","volume":"37","author":"Xiao Lei","year":"2018","unstructured":"Lei Xiao , Anton Kaplanyan , Alexander Fix , Matthew Chapman , and Douglas Lanman . 2018 . DeepFocus: learned image synthesis for computational displays . ACM Transactions on Graphics (TOG) 37 , 6 (2018), 1 \u2013 13 . Lei Xiao, Anton Kaplanyan, Alexander Fix, Matthew Chapman, and Douglas Lanman. 2018. DeepFocus: learned image synthesis for computational displays. ACM Transactions on Graphics (TOG) 37, 6 (2018), 1\u201313.","journal-title":"ACM Transactions on Graphics (TOG)"},{"key":"e_1_3_2_1_39_1","unstructured":"Alex Yu Ruilong Li Matthew Tancik Hao Li Ren Ng and Angjoo Kanazawa. 2021a. Plenoctrees for real-time rendering of neural radiance fields. arXiv preprint arXiv:2103.14024(2021).  Alex Yu Ruilong Li Matthew Tancik Hao Li Ren Ng and Angjoo Kanazawa. 2021a. Plenoctrees for real-time rendering of neural radiance fields. arXiv preprint arXiv:2103.14024(2021)."},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00455"},{"key":"e_1_3_2_1_41_1","unstructured":"Kai Zhang Gernot Riegler Noah Snavely and Vladlen Koltun. 2020. Nerf++: Analyzing and improving neural radiance fields. arXiv preprint arXiv:2010.07492(2020).  Kai Zhang Gernot Riegler Noah Snavely and Vladlen Koltun. 2020. Nerf++: Analyzing and improving neural radiance fields. arXiv preprint arXiv:2010.07492(2020)."},{"key":"e_1_3_2_1_42_1","unstructured":"Tinghui Zhou Richard Tucker John Flynn Graham Fyffe and Noah Snavely. 2018. Stereo magnification: Learning view synthesis using multiplane images. arXiv preprint arXiv:1805.09817(2018).  Tinghui Zhou Richard Tucker John Flynn Graham Fyffe and Noah Snavely. 2018. Stereo magnification: Learning view synthesis using multiplane images. arXiv preprint arXiv:1805.09817(2018)."},{"key":"e_1_3_2_1_43_1","volume-title":"High-quality video view interpolation using a layered representation. ACM transactions on graphics (TOG) 23, 3","author":"Zitnick C\u00a0Lawrence","year":"2004","unstructured":"C\u00a0Lawrence Zitnick , Sing\u00a0Bing Kang , Matthew Uyttendaele , Simon Winder , and Richard Szeliski . 2004. High-quality video view interpolation using a layered representation. ACM transactions on graphics (TOG) 23, 3 ( 2004 ), 600\u2013608. C\u00a0Lawrence Zitnick, Sing\u00a0Bing Kang, Matthew Uyttendaele, Simon Winder, and Richard Szeliski. 2004. High-quality video view interpolation using a layered representation. ACM transactions on graphics (TOG) 23, 3 (2004), 600\u2013608."}],"event":{"name":"SIGGRAPH '22: Special Interest Group on Computer Graphics and Interactive Techniques Conference","location":"Vancouver BC Canada","acronym":"SIGGRAPH '22","sponsor":["SIGGRAPH ACM Special Interest Group on Computer Graphics and Interactive Techniques"]},"container-title":["Special Interest Group on Computer Graphics and Interactive Techniques Conference Proceedings"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3528233.3530701","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:02:49Z","timestamp":1750186969000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3528233.3530701"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,8,7]]},"references-count":43,"alternative-id":["10.1145\/3528233.3530701","10.1145\/3528233"],"URL":"https:\/\/doi.org\/10.1145\/3528233.3530701","relation":{},"subject":[],"published":{"date-parts":[[2022,8,7]]}}}