{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,28]],"date-time":"2025-08-28T12:28:35Z","timestamp":1756384115410,"version":"3.40.3"},"publisher-location":"Cham","reference-count":87,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031726576"},{"type":"electronic","value":"9783031726583"}],"license":[{"start":{"date-parts":[[2024,10,2]],"date-time":"2024-10-02T00:00:00Z","timestamp":1727827200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,10,2]],"date-time":"2024-10-02T00:00:00Z","timestamp":1727827200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-72658-3_18","type":"book-chapter","created":{"date-parts":[[2024,10,2]],"date-time":"2024-10-02T03:32:37Z","timestamp":1727839957000},"page":"305-323","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["G3R: Gradient Guided Generalizable Reconstruction"],"prefix":"10.1007","author":[{"given":"Yun","family":"Chen","sequence":"first","affiliation":[]},{"given":"Jingkang","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Ze","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Sivabalan","family":"Manivasagam","sequence":"additional","affiliation":[]},{"given":"Raquel","family":"Urtasun","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,10,2]]},"reference":[{"key":"18_CR1","unstructured":"Altizure: Mapping the world in 3D. https:\/\/www.altizure.com"},{"key":"18_CR2","doi-asserted-by":"crossref","unstructured":"Adler, J., \u00d6ktem, O.: Solving ill-posed inverse problems using iterative deep neural networks. arXiv (2017)","DOI":"10.1088\/1361-6420\/aa9581"},{"key":"18_CR3","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"696","DOI":"10.1007\/978-3-030-58542-6_42","volume-title":"Computer Vision \u2013 ECCV 2020","author":"K-A Aliev","year":"2020","unstructured":"Aliev, K.-A., Sevastopolsky, A., Kolos, M., Ulyanov, D., Lempitsky, V.: Neural point-based graphics. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12367, pp. 696\u2013712. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58542-6_42"},{"key":"18_CR4","unstructured":"Andrychowicz, M., et al.: Learning to learn by gradient descent by gradient descent. In: NeurIPS (2016)"},{"key":"18_CR5","doi-asserted-by":"crossref","unstructured":"Carreira, J., Agrawal, P., Fragkiadaki, K., Malik, J.: Human pose estimation with iterative error feedback. In: CVPR (2015)","DOI":"10.1109\/CVPR.2016.512"},{"key":"18_CR6","doi-asserted-by":"crossref","unstructured":"Charatan, D., Li, S., Tagliasacchi, A., Sitzmann, V.: pixelsplat: 3D gaussian splats from image pairs for scalable generalizable 3D reconstruction. arXiv (2023)","DOI":"10.1109\/CVPR52733.2024.01840"},{"key":"18_CR7","doi-asserted-by":"crossref","unstructured":"Chen, A., et al.: MVSNeRF: fast generalizable radiance field reconstruction from multi-view stereo. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.01386"},{"key":"18_CR8","unstructured":"Cheng, K., et al.: GaussianPro: 3D gaussian splatting with progressive propagation. arXiv (2024)"},{"key":"18_CR9","doi-asserted-by":"crossref","unstructured":"Chibane, J., Bansal, A., Lazova, V., Pons-Moll, G.: Stereo radiance fields (SRF): Learning view synthesis for sparse views of novel scenes. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00782"},{"key":"18_CR10","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"424","DOI":"10.1007\/978-3-319-46723-8_49","volume-title":"Medical Image Computing and Computer-Assisted Intervention \u2013 MICCAI 2016","author":"\u00d6 \u00c7i\u00e7ek","year":"2016","unstructured":"\u00c7i\u00e7ek, \u00d6., Abdulkadir, A., Lienkamp, S.S., Brox, T., Ronneberger, O.: 3D U-Net: learning dense volumetric segmentation from sparse annotation. In: Ourselin, S., Joskowicz, L., Sabuncu, M.R., Unal, G., Wells, W. (eds.) MICCAI 2016. LNCS, vol. 9901, pp. 424\u2013432. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46723-8_49"},{"key":"18_CR11","doi-asserted-by":"crossref","unstructured":"Cong, W., et al.: Enhancing nerf akin to enhancing LLMS: generalizable nerf transformer with mixture-of-view-experts. In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.00296"},{"key":"18_CR12","doi-asserted-by":"crossref","unstructured":"Flynn, J., et al.: DeepView: view synthesis with learned gradient descent. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00247"},{"key":"18_CR13","doi-asserted-by":"crossref","unstructured":"Gu\u00e9don, A., Lepetit, V.: Sugar: surface-aligned gaussian splatting for efficient 3D mesh reconstruction and high-quality mesh rendering. arXiv (2023)","DOI":"10.1109\/CVPR52733.2024.00512"},{"key":"18_CR14","unstructured":"Hong, Y., et al.: LRM: large reconstruction model for single image to 3D. In: The Twelfth International Conference on Learning Representations (2024). https:\/\/openreview.net\/forum?id=sllU8vvsFF"},{"key":"18_CR15","doi-asserted-by":"crossref","unstructured":"Huang, J., Gojcic, Z., Atzmon, M., Litany, O., Fidler, S., Williams, F.: Neural kernel surface reconstruction. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.00425"},{"key":"18_CR16","doi-asserted-by":"crossref","unstructured":"Huang, S., et al.: Neural lidar fields for novel view synthesis. arXiv (2023)","DOI":"10.1109\/ICCV51070.2023.01672"},{"key":"18_CR17","doi-asserted-by":"crossref","unstructured":"Johari, M.M., Lepoittevin, Y., Fleuret, F.: GeoNeRF: generalizing nerf with geometry priors. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01782"},{"key":"18_CR18","doi-asserted-by":"crossref","unstructured":"Kerbl, B., Kopanas, G., Leimk\u00fchler, T., Drettakis, G.: 3D gaussian splatting for real-time radiance field rendering. TOG (2023)","DOI":"10.1145\/3592433"},{"key":"18_CR19","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. In: ICLR (2015)"},{"key":"18_CR20","doi-asserted-by":"crossref","unstructured":"Kopanas, G., Philip, J., Leimk\u00fchler, T., Drettakis, G.: Point-based neural rendering with per-view optimization. Computer graphics forum (Print) (2021)","DOI":"10.1111\/cgf.14339"},{"key":"18_CR21","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"198","DOI":"10.1007\/978-3-031-19784-0_12","volume-title":"ECCV 2022","author":"J Kulh\u2019anek","year":"2022","unstructured":"Kulh\u2019anek, J., Derner, E., Sattler, T., Babuvska, R.: ViewFormer: NeRF-free neural rendering from few images using transformers. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13675, pp. 198\u2013216. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19784-0_12"},{"key":"18_CR22","unstructured":"Li, J., et al.: Instant3D: fast text-to-3D with sparse-view generation and large reconstruction model. In: The Twelfth International Conference on Learning Representations (2024). https:\/\/openreview.net\/forum?id=2lDQLiH1W4"},{"key":"18_CR23","unstructured":"Li, K., Malik, J.: Learning to optimize. In: ICLR (2016)"},{"key":"18_CR24","doi-asserted-by":"crossref","unstructured":"Li, Y., Wang, G., Ji, X., Xiang, Y., Fox, D.: DeepIM: deep iterative matching for 6D pose estimation. IJCV (2018)","DOI":"10.1007\/978-3-030-01231-1_42"},{"key":"18_CR25","unstructured":"Liang, Y., He, H., Chen, Y.: RETR: modeling rendering via transformer for generalizable neural surface reconstruction. In: NeurIPS (2023)"},{"key":"18_CR26","doi-asserted-by":"crossref","unstructured":"Lin, H., et al.: Efficient neural radiance fields for interactive free-viewpoint video. In: SIGGRAPH Asia 2022 Conference Papers (2022)","DOI":"10.1145\/3550469.3555376"},{"key":"18_CR27","doi-asserted-by":"crossref","unstructured":"Lin, J., et\u00a0al.: VastGaussian: vast 3D gaussians for large scene reconstruction. arXiv (2024)","DOI":"10.1109\/CVPR52733.2024.00494"},{"key":"18_CR28","unstructured":"Lin, Z.H.,et al.: UrbaNIR: large-scale urban scene inverse rendering from a single video. arXiv (2023)"},{"key":"18_CR29","doi-asserted-by":"crossref","unstructured":"Liu, J.Y., Chen, Y., Yang, Z., Wang, J., Manivasagam, S., Urtasun, R.: Real-time neural rasterization for large scenes. In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.00773"},{"key":"18_CR30","doi-asserted-by":"crossref","unstructured":"Liu, R., Wu, R., Van\u00a0Hoorick, B., Tokmakov, P., Zakharov, S., Vondrick, C.: Zero-1-to-3: zero-shot one image to 3D object. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 9298\u20139309 (2023)","DOI":"10.1109\/ICCV51070.2023.00853"},{"key":"18_CR31","doi-asserted-by":"crossref","unstructured":"Liu, Y., et al.: Neural rays for occlusion-aware image-based rendering. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.00767"},{"key":"18_CR32","series-title":"LNCS","first-page":"210","volume-title":"ECCV 2022","author":"X Long","year":"2022","unstructured":"Long, X., Lin, C., Wang, P., Komura, T., Wang, W.: Sparseneus: fast generalizable neural surface reconstruction from sparse views. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13692, pp. 210\u2013227. Springer, Cham (2022)"},{"key":"18_CR33","doi-asserted-by":"crossref","unstructured":"Luiten, J., Kopanas, G., Leibe, B., Ramanan, D.: Dynamic 3D gaussians: tracking by persistent dynamic view synthesis. arXiv (2023)","DOI":"10.1109\/3DV62453.2024.00044"},{"key":"18_CR34","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"229","DOI":"10.1007\/978-3-030-58558-7_14","volume-title":"Computer Vision \u2013 ECCV 2020","author":"W-C Ma","year":"2020","unstructured":"Ma, W.-C., Wang, S., Gu, J., Manivasagam, S., Torralba, A., Urtasun, R.: Deep feedback inverse problem solver. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12350, pp. 229\u2013246. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58558-7_14"},{"key":"18_CR35","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"833","DOI":"10.1007\/978-3-030-01264-9_49","volume-title":"Computer Vision \u2013 ECCV 2018","author":"F Manhardt","year":"2018","unstructured":"Manhardt, F., Kehl, W., Navab, N., Tombari, F.: Deep model-based 6D pose refinement in RGB. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) Computer Vision \u2013 ECCV 2018. LNCS, vol. 11218, pp. 833\u2013849. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01264-9_49"},{"key":"18_CR36","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"405","DOI":"10.1007\/978-3-030-58452-8_24","volume-title":"Computer Vision \u2013 ECCV 2020","author":"B Mildenhall","year":"2020","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 405\u2013421. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_24"},{"key":"18_CR37","doi-asserted-by":"crossref","unstructured":"M\u00fcller, T., Evans, A., Schied, C., Keller, A.: Instant neural graphics primitives with a multiresolution hash encoding (2022)","DOI":"10.1145\/3528223.3530127"},{"key":"18_CR38","doi-asserted-by":"crossref","unstructured":"M\u00fcller, N., et al.: AutoRF: learning 3D object radiance fields from single view observations. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.00394"},{"key":"18_CR39","doi-asserted-by":"crossref","unstructured":"Niemeyer, M., Mescheder, L., Oechsle, M., Geiger, A.: Differentiable volumetric rendering: learning implicit 3D representations without 3D supervision. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00356"},{"key":"18_CR40","doi-asserted-by":"crossref","unstructured":"Ost, J., Mannan, F., Thuerey, N., Knodt, J., Heide, F.: Neural scene graphs for dynamic scenes. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00288"},{"key":"18_CR41","unstructured":"Pun, A., et al.: Neural lighting simulation for urban scenes. In: NeurIPS (2023)"},{"key":"18_CR42","doi-asserted-by":"crossref","unstructured":"Reizenstein, J., Shapovalov, R., Henzler, P., Sbordone, L., Labatut, P., Novotn\u00fd, D.: Common objects in 3D: large-scale learning and evaluation of real-life 3D category reconstruction. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.01072"},{"key":"18_CR43","doi-asserted-by":"crossref","unstructured":"Ren, Y., Wang, F., Zhang, T., Pollefeys, M., Susstrunk, S.E.: VolreCon: volume rendering of signed ray distance functions for generalizable multi-view reconstruction. In: CVPR (2022)","DOI":"10.1109\/CVPR52729.2023.01601"},{"key":"18_CR44","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"623","DOI":"10.1007\/978-3-030-58529-7_37","volume-title":"Computer Vision \u2013 ECCV 2020","author":"G Riegler","year":"2020","unstructured":"Riegler, G., Koltun, V.: Free view synthesis. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12364, pp. 623\u2013640. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58529-7_37"},{"key":"18_CR45","doi-asserted-by":"crossref","unstructured":"Riegler, G., Koltun, V.: Stable view synthesis. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.01204"},{"key":"18_CR46","doi-asserted-by":"crossref","unstructured":"Roessle, B., M\u00fcller, N., Porzi, L., Bul\u00f2, S.R., Kontschieder, P., Nie\u00dfner, M.: GANERF: leveraging discriminators to optimize neural radiance fields. ACM Trans. Graph (2023)","DOI":"10.1145\/3618402"},{"key":"18_CR47","doi-asserted-by":"crossref","unstructured":"Rombach, R., Esser, P., Ommer, B.: Geometry-free view synthesis: transformers and no 3D priors. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.01409"},{"key":"18_CR48","doi-asserted-by":"crossref","unstructured":"Sajjadi, M.S.M., et al.: Rust: latent neural scene representations from unposed imagery. In: CVPR (2022)","DOI":"10.1109\/CVPR52729.2023.01659"},{"key":"18_CR49","doi-asserted-by":"crossref","unstructured":"Sajjadi, M.S.M., et al.: Scene representation transformer: geometry-free novel view synthesis through set-latent scene representations. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.00613"},{"key":"18_CR50","unstructured":"Seitzer, M., van Steenkiste, S., Kipf, T., Greff, K., Sajjadi, M.S.M.: DYST: towards dynamic neural scene representations on real-world videos. arXiv (2023)"},{"key":"18_CR51","unstructured":"Sitzmann, V., Rezchikov, S., Freeman, W., Tenenbaum, J., Durand, F.: Light field networks: neural scene representations with single-evaluation rendering. NeurIPS (2021)"},{"key":"18_CR52","unstructured":"Sitzmann, V., Zollh\u00f6fer, M., Wetzstein, G.: Scene representation networks: continuous 3D-structure-aware neural scene representations. In: NeurIPS (2019)"},{"key":"18_CR53","doi-asserted-by":"crossref","unstructured":"Snavely, N., Seitz, S.M., Szeliski, R.: Photo tourism: exploring photo collections in 3D. In:SIGGRAPH (2006)","DOI":"10.1145\/1141911.1141964"},{"key":"18_CR54","unstructured":"Song, J., Meng, C., Ermon, S.: Denoising diffusion implicit models. In: ICLR (2020)"},{"key":"18_CR55","doi-asserted-by":"crossref","unstructured":"Srinivasan, P.P., Tucker, R., Barron, J.T., Ramamoorthi, R., Ng, R., Snavely, N.: Pushing the boundaries of view extrapolation with multiplane images. arXiv (2019)","DOI":"10.1109\/CVPR.2019.00026"},{"key":"18_CR56","doi-asserted-by":"crossref","unstructured":"Suhail, M., Esteves, C., Sigal, L., Makadia, A.: Light field neural rendering. In: CVPR (2021)","DOI":"10.1109\/CVPR52688.2022.00809"},{"key":"18_CR57","series-title":"LNCS","first-page":"156","volume-title":"ECCV 2022","author":"M Suhail","year":"2022","unstructured":"Suhail, M., Esteves, C., Sigal, L., Makadia, A.: Generalizable patch-based neural rendering. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13692, pp. 156\u2013174. Springer, Cham (2022)"},{"key":"18_CR58","doi-asserted-by":"crossref","unstructured":"Sun, P., et al.: Scalability in perception for autonomous driving: waymo open dataset. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00252"},{"key":"18_CR59","doi-asserted-by":"crossref","unstructured":"Tancik, M., et al.: Block-nerf: scalable large scene neural view synthesis. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.00807"},{"key":"18_CR60","doi-asserted-by":"crossref","unstructured":"Tang, H., et al.: Torchsparse++: efficient training and inference framework for sparse convolution on GPUs. In: IEEE\/ACM International Symposium on Microarchitecture (MICRO) (2023)","DOI":"10.1145\/3613424.3614303"},{"key":"18_CR61","doi-asserted-by":"crossref","unstructured":"Tonderski, A., Lindstr\u00f6m, C., Hess, G., Ljungbergh, W., Svensson, L., Petersson, C.: Neurad: neural rendering for autonomous driving. arXiv (2023)","DOI":"10.1109\/CVPR52733.2024.01411"},{"key":"18_CR62","doi-asserted-by":"crossref","unstructured":"Trevithick, A., Yang, B.: GRF: learning a general radiance field for 3D representation and rendering. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.01490"},{"key":"18_CR63","doi-asserted-by":"crossref","unstructured":"Turki, H., Ramanan, D., Satyanarayanan, M.: Mega-nerf: scalable construction of large-scale nerfs for virtual fly-throughs. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01258"},{"key":"18_CR64","unstructured":"Wang, J., et al.: CADSim: robust and scalable in-the-wild 3D reconstruction for controllable sensor simulation. In: 6th Annual Conference on Robot Learning (2022)"},{"key":"18_CR65","unstructured":"Wang, P., Chen, X., Chen, T., Venugopalan, S., Wang, Z., et\u00a0al.: Is attention all nerf needs? arXiv (2022)"},{"key":"18_CR66","doi-asserted-by":"crossref","unstructured":"Wang, Q., et al.: IBRNet: learning multi-view image-based rendering. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00466"},{"key":"18_CR67","doi-asserted-by":"crossref","unstructured":"Wang, Z., et al.: Neural fields meet explicit geometric representations for inverse rendering of urban scenes. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.00809"},{"key":"18_CR68","unstructured":"Wei, X., et al.: MesHLRM: large reconstruction model for high-quality mesh. arXiv preprint arXiv:2404.12385 (2024)"},{"key":"18_CR69","unstructured":"Wichrowska, O., et al.: Learned optimizers that scale and generalize. In: ICML (2017)"},{"key":"18_CR70","doi-asserted-by":"crossref","unstructured":"Wiles, O., Gkioxari, G., Szeliski, R., Johnson, J.: Synsin: End-to-end view synthesis from a single image. arXiv (2019)","DOI":"10.1109\/CVPR42600.2020.00749"},{"key":"18_CR71","doi-asserted-by":"crossref","unstructured":"Wu, J., et al.: GOMVS: geometrically consistent cost aggregation for multi-view stereo. In: CVPR (2024)","DOI":"10.1109\/CVPR52733.2024.01910"},{"key":"18_CR72","doi-asserted-by":"crossref","unstructured":"Wu, Z., et\u00a0al.: Mars: An instance-aware, modular and realistic simulator for autonomous driving. arXiv (2023)","DOI":"10.1007\/978-981-99-8850-1_1"},{"key":"18_CR73","doi-asserted-by":"crossref","unstructured":"Xiao, P., et\u00a0al.: Pandaset: advanced sensor suite dataset for autonomous driving. In: ITSC (2021)","DOI":"10.1109\/ITSC48978.2021.9565009"},{"key":"18_CR74","unstructured":"Yan, Y., et al.: Street gaussians for modeling dynamic urban scenes. arXiv (2024)"},{"key":"18_CR75","doi-asserted-by":"crossref","unstructured":"Yang, H., Hong, L., Li, A., Hu, T., Li, Z., Lee, G.H., Wang, L.: Contranerf: Generalizable neural radiance fields for synthetic-to-real novel view synthesis via contrastive learning. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.01584"},{"key":"18_CR76","unstructured":"Yang, J., et\u00a0al.: EmerNeRF: emergent spatial-temporal scene decomposition via self-supervision. arXiv (2023)"},{"key":"18_CR77","doi-asserted-by":"crossref","unstructured":"Yang, Z., et al.: UNISIM: a neural closed-loop sensor simulator. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.00140"},{"key":"18_CR78","doi-asserted-by":"crossref","unstructured":"Yang, Z., Manivasagam, S., Chen, Y., Wang, J., Hu, R., Urtasun, R.: Reconstructing objects in-the-wild for realistic sensor simulation. In: ICRA (2023)","DOI":"10.1109\/ICRA48891.2023.10160535"},{"key":"18_CR79","unstructured":"Yang, Z., Manivasagam, S., Liang, M., Yang, B., Ma, W.C., Urtasun, R.: Recovering and simulating pedestrians in the wild. In: Conference on Robot Learning, pp. 419\u2013431. PMLR (2021)"},{"key":"18_CR80","doi-asserted-by":"crossref","unstructured":"Yang, Z., et al.: S3: neural shape, skeleton, and skinning fields for 3D human modeling. In: CVPR, pp. 13284\u201313293 (2021)","DOI":"10.1109\/CVPR46437.2021.01308"},{"key":"18_CR81","doi-asserted-by":"crossref","unstructured":"Yao, Y., et al.: BlendedMVS: a large-scale dataset for generalized multi-view stereo networks. CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00186"},{"key":"18_CR82","doi-asserted-by":"crossref","unstructured":"Yu, A., Ye, V., Tancik, M., Kanazawa, A.: PixelNeRF: neural radiance fields from one or few images. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00455"},{"key":"18_CR83","doi-asserted-by":"crossref","unstructured":"Zhang, K., et al.: GS-LRM: large reconstruction model for 3D gaussian splatting. arXiv preprint arXiv:2404.19702 (2024)","DOI":"10.1007\/978-3-031-72670-5_1"},{"key":"18_CR84","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"18_CR85","unstructured":"Zhenxing, M., Xu, D.: Switch-NeRF: learning scene decomposition with mixture of experts for large-scale neural radiance fields. In: ICLR (2022)"},{"key":"18_CR86","doi-asserted-by":"crossref","unstructured":"Zhou, T., Tucker, R., Flynn, J., Fyffe, G., Snavely, N.: Stereo magnification: learning view synthesis using multiplane images. In: SIGGRAPH (2018)","DOI":"10.1145\/3197517.3201323"},{"key":"18_CR87","doi-asserted-by":"crossref","unstructured":"Zhou, X., Lin, Z., Shan, X., Wang, Y., Sun, D., Yang, M.H.: DrivingGaussian: composite gaussian splatting for surrounding dynamic autonomous driving scenes. arXiv (2023)","DOI":"10.1109\/CVPR52733.2024.02044"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-72658-3_18","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,2]],"date-time":"2024-10-02T03:40:07Z","timestamp":1727840407000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-72658-3_18"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,2]]},"ISBN":["9783031726576","9783031726583"],"references-count":87,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-72658-3_18","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,10,2]]},"assertion":[{"value":"2 October 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}