{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,3]],"date-time":"2026-01-03T06:45:07Z","timestamp":1767422707839,"version":"3.40.3"},"publisher-location":"Cham","reference-count":69,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031197802"},{"type":"electronic","value":"9783031197819"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19781-9_42","type":"book-chapter","created":{"date-parts":[[2022,10,22]],"date-time":"2022-10-22T12:12:59Z","timestamp":1666440779000},"page":"730-748","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":26,"title":["Sem2NeRF: Converting Single-View Semantic Masks to\u00a0Neural Radiance Fields"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-0943-1512","authenticated-orcid":false,"given":"Yuedong","family":"Chen","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8764-6178","authenticated-orcid":false,"given":"Qianyi","family":"Wu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3584-9640","authenticated-orcid":false,"given":"Chuanxia","family":"Zheng","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5264-2572","authenticated-orcid":false,"given":"Tat-Jen","family":"Cham","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9444-3763","authenticated-orcid":false,"given":"Jianfei","family":"Cai","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,23]]},"reference":[{"key":"42_CR1","doi-asserted-by":"crossref","unstructured":"Barron, J.T., Mildenhall, B., Tancik, M., Hedman, P., Martin-Brualla, R., Srinivasan, P.P.: Mip-NeRF: a multiscale representation for anti-aliasing neural radiance fields. In: IEEE International Conference on Computer Vision, pp. 5855\u20135864 (2021)","DOI":"10.1109\/ICCV48922.2021.00580"},{"key":"42_CR2","doi-asserted-by":"crossref","unstructured":"Chan, E.R., et al.: Efficient geometry-aware 3D generative adversarial networks. IEEE Conference on Computer Vision and Pattern Recognition (2022)","DOI":"10.1109\/CVPR52688.2022.01565"},{"key":"42_CR3","doi-asserted-by":"crossref","unstructured":"Chan, E.R., Monteiro, M., Kellnhofer, P., Wu, J., Wetzstein, G.: pi-GAN: Periodic implicit generative adversarial networks for 3D-aware image synthesis. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 5799\u20135809 (2021)","DOI":"10.1109\/CVPR46437.2021.00574"},{"issue":"1","key":"42_CR4","first-page":"1","volume":"41","author":"A Chen","year":"2022","unstructured":"Chen, A., Liu, R., Xie, L., Chen, Z., Su, H., Yu, J.: SofGAN: a portrait image generator with dynamic styling. ACM Trans. Graph. 41(1), 1\u201326 (2022)","journal-title":"ACM Trans. Graph."},{"key":"42_CR5","doi-asserted-by":"crossref","unstructured":"Chen, W., Hays, J.: Sketchygan: Towards diverse and realistic sketch to image synthesis. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 9416\u20139425 (2018)","DOI":"10.1109\/CVPR.2018.00981"},{"key":"42_CR6","unstructured":"Chen, Y., Huang, J., Wang, J., Xie, X.: Edge prior augmented networks for motion deblurring on naturally blurry images. arXiv preprint arXiv:2109.08915 (2021)"},{"key":"42_CR7","doi-asserted-by":"crossref","unstructured":"Chen, Z., Zhang, H.: Learning implicit fields for generative shape modeling. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 5939\u20135948 (2019)","DOI":"10.1109\/CVPR.2019.00609"},{"key":"42_CR8","doi-asserted-by":"crossref","unstructured":"Choi, Y., Uh, Y., Yoo, J., Ha, J.W.: StarGAN v2: diverse image synthesis for multiple domains. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 8188\u20138197 (2020)","DOI":"10.1109\/CVPR42600.2020.00821"},{"key":"42_CR9","doi-asserted-by":"crossref","unstructured":"Collins, E., Bala, R., Price, B., Susstrunk, S.: Editing in style: uncovering the local semantics of GANs. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 5771\u20135780 (2020)","DOI":"10.1109\/CVPR42600.2020.00581"},{"key":"42_CR10","unstructured":"Dosovitskiy, A., et al.: an image is worth 16x16 words: transformers for image recognition at scale. International Conference on Learning Representations (2021)"},{"key":"42_CR11","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"88","DOI":"10.1007\/978-3-030-58555-6_6","volume-title":"Computer Vision \u2013 ECCV 2020","author":"S Goel","year":"2020","unstructured":"Goel, S., Kanazawa, A., Malik, J.: Shape and viewpoint without keypoints. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12360, pp. 88\u2013104. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58555-6_6"},{"key":"42_CR12","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. Advances in Neural Information Processing Systems 27 (2014)"},{"key":"42_CR13","unstructured":"Gu, J., Liu, L., Wang, P., Theobalt, C.: StyleNeRF: a style-based 3D-aware generator for high-resolution image synthesis. In: International Conference on Learning Representations (2022)"},{"key":"42_CR14","doi-asserted-by":"crossref","unstructured":"Hao, Z., Mallya, A., Belongie, S., Liu, M.Y.: GANcraft: unsupervised 3D neural rendering of minecraft worlds. In: IEEE International Conference on Computer Vision, pp. 14072\u201314082 (2021)","DOI":"10.1109\/ICCV48922.2021.01381"},{"key":"42_CR15","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"42_CR16","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANs trained by a two time-scale update rule converge to a local nash equilibrium. In: Advances in Neural Information Processing Systems 30 (2017)"},{"key":"42_CR17","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"592","DOI":"10.1007\/978-3-030-58610-2_35","volume-title":"Computer Vision \u2013 ECCV 2020","author":"H-P Huang","year":"2020","unstructured":"Huang, H.-P., Tseng, H.-Y., Lee, H.-Y., Huang, J.-B.: Semantic view synthesis. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12357, pp. 592\u2013608. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58610-2_35"},{"key":"42_CR18","doi-asserted-by":"crossref","unstructured":"Isola, P., Zhu, J.Y., Zhou, T., Efros, A.A.: Image-to-image translation with conditional adversarial networks. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1125\u20131134 (2017)","DOI":"10.1109\/CVPR.2017.632"},{"key":"42_CR19","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"694","DOI":"10.1007\/978-3-319-46475-6_43","volume-title":"Computer Vision \u2013 ECCV 2016","author":"J Johnson","year":"2016","unstructured":"Johnson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time style transfer and super-resolution. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9906, pp. 694\u2013711. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46475-6_43"},{"key":"42_CR20","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"386","DOI":"10.1007\/978-3-030-01267-0_23","volume-title":"Computer Vision \u2013 ECCV 2018","author":"A Kanazawa","year":"2018","unstructured":"Kanazawa, A., Tulsiani, S., Efros, A.A., Malik, J.: Learning category-specific mesh reconstruction from image collections. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11219, pp. 386\u2013402. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01267-0_23"},{"key":"42_CR21","unstructured":"Karras, T., Aila, T., Laine, S., Lehtinen, J.: Progressive growing of GANs for improved quality, stability, and variation. In: International Conference on Learning Representations (2018)"},{"key":"42_CR22","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 4401\u20134410 (2019)","DOI":"10.1109\/CVPR.2019.00453"},{"key":"42_CR23","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of styleGAN. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 8110\u20138119 (2020)","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"42_CR24","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. In: International Conference on Learning Representations (2014)"},{"key":"42_CR25","doi-asserted-by":"crossref","unstructured":"Lee, C.H., Liu, Z., Wu, L., Luo, P.: MaskGAN: towards diverse and interactive facial image manipulation. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 5549\u20135558 (2020)","DOI":"10.1109\/CVPR42600.2020.00559"},{"issue":"3","key":"42_CR26","doi-asserted-by":"publisher","first-page":"245","DOI":"10.1145\/78964.78965","volume":"9","author":"M Levoy","year":"1990","unstructured":"Levoy, M.: Efficient ray tracing of volume data. ACM Trans. Graph. 9(3), 245\u2013261 (1990)","journal-title":"ACM Trans. Graph."},{"key":"42_CR27","unstructured":"Ling, H., Kreis, K., Li, D., Kim, S.W., Torralba, A., Fidler, S.: EditGAN: high-precision semantic image editing. In: Advances in Neural Information Processing Systems (2021)"},{"key":"42_CR28","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"363","DOI":"10.1007\/978-3-030-58574-7_22","volume-title":"Computer Vision \u2013 ECCV 2020","author":"W Lira","year":"2020","unstructured":"Lira, W., Merz, J., Ritchie, D., Cohen-Or, D., Zhang, H.: GANHopper: multi-hop GAN for unsupervised image-to-image translation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12371, pp. 363\u2013379. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58574-7_22"},{"key":"42_CR29","first-page":"15651","volume":"33","author":"L Liu","year":"2020","unstructured":"Liu, L., Gu, J., Zaw Lin, K., Chua, T.S., Theobalt, C.: Neural sparse voxel fields. Adv. Neural Inform. Process. Syst. 33, 15651\u201315663 (2020)","journal-title":"Adv. Neural Inform. Process. Syst."},{"key":"42_CR30","doi-asserted-by":"crossref","unstructured":"Liu, X., Xu, Y., Wu, Q., Zhou, H., Wu, W., Zhou, B.: Semantic-aware implicit neural audio-driven video portrait generation. arXiv preprint arXiv:2201.07786 (2022)","DOI":"10.1007\/978-3-031-19836-6_7"},{"key":"42_CR31","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: IEEE Conference on Computer Vision, pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"42_CR32","doi-asserted-by":"crossref","unstructured":"Liu, Z., Luo, P., Wang, X., Tang, X.: Deep learning face attributes in the wild. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 3730\u20133738 (2015)","DOI":"10.1109\/ICCV.2015.425"},{"issue":"4","key":"42_CR33","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459863","volume":"40","author":"S Lombardi","year":"2021","unstructured":"Lombardi, S., Simon, T., Schwartz, G., Zollhoefer, M., Sheikh, Y., Saragih, J.: Mixture of volumetric primitives for efficient neural rendering. ACM Trans. Graph. 40(4), 1\u201313 (2021)","journal-title":"ACM Trans. Graph."},{"key":"42_CR34","doi-asserted-by":"crossref","unstructured":"Luo, S., Hu, W.: Diffusion probabilistic models for 3d point cloud generation. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 2837\u20132845 (2021)","DOI":"10.1109\/CVPR46437.2021.00286"},{"key":"42_CR35","unstructured":"Mescheder, L., Geiger, A., Nowozin, S.: Which training methods for GANs do actually converge? In: International Conference on Machine Learning, pp. 3481\u20133490. PMLR (2018)"},{"key":"42_CR36","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"405","DOI":"10.1007\/978-3-030-58452-8_24","volume-title":"Computer Vision \u2013 ECCV 2020","author":"B Mildenhall","year":"2020","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: NeRF: representing scenes as neural radiance fields for view synthesis. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 405\u2013421. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_24"},{"key":"42_CR37","doi-asserted-by":"crossref","unstructured":"Niemeyer, M., Geiger, A.: Giraffe: Representing scenes as compositional generative neural feature fields. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 11453\u201311464 (2021)","DOI":"10.1109\/CVPR46437.2021.01129"},{"key":"42_CR38","doi-asserted-by":"crossref","unstructured":"Park, T., Liu, M.Y., Wang, T.C., Zhu, J.Y.: Semantic image synthesis with spatially-adaptive normalization. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 2337\u20132346 (2019)","DOI":"10.1109\/CVPR.2019.00244"},{"key":"42_CR39","unstructured":"Paszke, A., et al.: PyTorch: an imperative style, high-performance deep learning library. In: Advances in Neural Information Processing Systems 32 (2019)"},{"key":"42_CR40","doi-asserted-by":"crossref","unstructured":"Perez, E., Strub, F., De Vries, H., Dumoulin, V., Courville, A.: Film: visual reasoning with a general conditioning layer. In: AAAI, vol. 32 (2018)","DOI":"10.1609\/aaai.v32i1.11671"},{"key":"42_CR41","doi-asserted-by":"crossref","unstructured":"Pumarola, A., Corona, E., Pons-Moll, G., Moreno-Noguer, F.: D-NeRF: neural radiance fields for dynamic scenes. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 10318\u201310327 (2021)","DOI":"10.1109\/CVPR46437.2021.01018"},{"key":"42_CR42","doi-asserted-by":"crossref","unstructured":"Richardson, E., et al.: Encoding in style: a styleGAN encoder for image-to-image translation. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 2287\u20132296 (2021)","DOI":"10.1109\/CVPR46437.2021.00232"},{"key":"42_CR43","unstructured":"Salimans, T., Goodfellow, I., Zaremba, W., Cheung, V., Radford, A., Chen, X.: Improved techniques for training GANs. Advances in Neural Information Processing Systems 29 (2016)"},{"key":"42_CR44","first-page":"20154","volume":"33","author":"K Schwarz","year":"2020","unstructured":"Schwarz, K., Liao, Y., Niemeyer, M., Geiger, A.: GRAF: generative radiance fields for 3D-aware image synthesis. Adv. Neural Inform. Process. Syst. 33, 20154\u201320166 (2020)","journal-title":"Adv. Neural Inform. Process. Syst."},{"key":"42_CR45","doi-asserted-by":"crossref","unstructured":"Shi, Y., Yang, X., Wan, Y., Shen, X.: SemanticStyleGAN: learning compositional generative priors for controllable image synthesis and editing. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 11254\u201311264 (2022)","DOI":"10.1109\/CVPR52688.2022.01097"},{"key":"42_CR46","first-page":"7462","volume":"33","author":"V Sitzmann","year":"2020","unstructured":"Sitzmann, V., Martel, J., Bergman, A., Lindell, D., Wetzstein, G.: Implicit neural representations with periodic activation functions. Adv. Neural Inform. Process. Syst. 33, 7462\u20137473 (2020)","journal-title":"Adv. Neural Inform. Process. Syst."},{"issue":"4","key":"42_CR47","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459771","volume":"40","author":"G Song","year":"2021","unstructured":"Song, G., et al.: AgileGAN: stylizing portraits by inversion-consistent transfer learning. ACM Trans. Graph. 40(4), 1\u201313 (2021)","journal-title":"ACM Trans. Graph."},{"key":"42_CR48","doi-asserted-by":"crossref","unstructured":"Sun, C., Sun, M., Chen, H.T.: Direct voxel grid optimization: super-fast convergence for radiance fields reconstruction. In: IEEE Conference on Computer Vision and Pattern Recognition (2022)","DOI":"10.1109\/CVPR52688.2022.00538"},{"key":"42_CR49","doi-asserted-by":"crossref","unstructured":"Sun, J., et al.: FENeRF: face editing in neural radiance fields. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 7672\u20137682 (2022)","DOI":"10.1109\/CVPR52688.2022.00752"},{"key":"42_CR50","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z.: Rethinking the inception architecture for computer vision. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 2818\u20132826 (2016)","DOI":"10.1109\/CVPR.2016.308"},{"key":"42_CR51","doi-asserted-by":"crossref","unstructured":"Wang, T.C., Liu, M.Y., Zhu, J.Y., Tao, A., Kautz, J., Catanzaro, B.: High-resolution image synthesis and semantic manipulation with conditional GANs. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 8798\u20138807 (2018)","DOI":"10.1109\/CVPR.2018.00917"},{"key":"42_CR52","unstructured":"Wu, J., Zhang, C., Xue, T., Freeman, B., Tenenbaum, J.: Learning a probabilistic latent space of object shapes via 3D generative-adversarial modeling. In: Advances in Neural Information Processing Systems 29 (2016)"},{"key":"42_CR53","doi-asserted-by":"crossref","unstructured":"Wu, Q., Liu, X., Chen, Y., Li, K., Zheng, C., Cai, J., Zheng, J.: Object-compositional neural implicit surfaces. arXiv preprint arXiv:2207.09686 (2022)","DOI":"10.1109\/ICCV51070.2023.01989"},{"key":"42_CR54","doi-asserted-by":"crossref","unstructured":"Wu, Z., Lischinski, D., Shechtman, E.: StyleSpace Analysis: disentangled controls for stylegan image generation. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 12863\u201312872 (2021)","DOI":"10.1109\/CVPR46437.2021.01267"},{"key":"42_CR55","unstructured":"Wu, Z., Nitzan, Y., Shechtman, E., Lischinski, D.: StyleAlign: analysis and applications of aligned styleGAN models. In: International Conference on Learning Representations (2022)"},{"key":"42_CR56","doi-asserted-by":"crossref","unstructured":"Xu, Y., et al.: TransEditor: transformer-based dual-space GAN for highly controllable facial editing. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 7683\u20137692 (2022)","DOI":"10.1109\/CVPR52688.2022.00753"},{"key":"42_CR57","doi-asserted-by":"crossref","unstructured":"Xu, Y., Peng, S., Yang, C., Shen, Y., Zhou, B.: 3D-aware image synthesis via learning structural and textural representations. In: IEEE Conference on Computer Vision and Pattern Recognition (2022)","DOI":"10.1109\/CVPR52688.2022.01788"},{"key":"42_CR58","doi-asserted-by":"crossref","unstructured":"Yen-Chen, L., Florence, P., Barron, J.T., Rodriguez, A., Isola, P., Lin, T.Y.: iNeRF: inverting neural radiance fields for pose estimation. In: IEEE International Conference on Intelligent Robots and Systems, pp. 1323\u20131330. IEEE (2021)","DOI":"10.1109\/IROS51168.2021.9636708"},{"key":"42_CR59","doi-asserted-by":"crossref","unstructured":"Zhang, J., Sangineto, E., Tang, H., Siarohin, A., Zhong, Z., Sebe, N., Wang, W.: 3D-aware semantic-guided generative model for human synthesis. arXiv preprint arXiv:2112.01422 (2021)","DOI":"10.1007\/978-3-031-19784-0_20"},{"key":"42_CR60","unstructured":"Zhang, R.: Making convolutional networks shift-invariant again. In: International Conference on Machine Learning, pp. 7324\u20137334. PMLR (2019)"},{"key":"42_CR61","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 586\u2013595 (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"42_CR62","doi-asserted-by":"crossref","unstructured":"Zhang, X., Zheng, Z., Gao, D., Zhang, B., Pan, P., Yang, Y.: Multi-view consistent generative adversarial networks for 3D-aware image synthesis. In: IEEE IEEE Conference on Computer Vision and Pattern Recognition, pp. 18450\u201318459 (2022)","DOI":"10.1109\/CVPR52688.2022.01790"},{"key":"42_CR63","doi-asserted-by":"crossref","unstructured":"Zhang, Y., et al.: DatasetGAN: efficient labeled data factory with minimal human effort. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 10145\u201310155 (2021)","DOI":"10.1109\/CVPR46437.2021.01001"},{"key":"42_CR64","doi-asserted-by":"crossref","unstructured":"Zheng, C., Cham, T.J., Cai, J.: Pluralistic image completion. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1438\u20131447 (2019)","DOI":"10.1109\/CVPR.2019.00153"},{"key":"42_CR65","unstructured":"Zheng, C., Cham, T.J., Cai, J.: TFill: image completion via a transformer-based architecture. In: IEEE Conference on Computer Vision and Pattern Recognition (2022)"},{"key":"42_CR66","doi-asserted-by":"crossref","unstructured":"Zhi, S., Laidlow, T., Leutenegger, S., Davison, A.J.: In-place scene labelling and understanding with implicit scene representation. In: International Conference on Computer Vision, pp. 15838\u201315847 (2021)","DOI":"10.1109\/ICCV48922.2021.01554"},{"key":"42_CR67","doi-asserted-by":"crossref","unstructured":"Zhou, T., Tucker, R., Flynn, J., Fyffe, G., Snavely, N.: Stereo magnification: learning view synthesis using multiplane images. ACM Transactions on Graphics (2018)","DOI":"10.1145\/3197517.3201323"},{"key":"42_CR68","doi-asserted-by":"crossref","unstructured":"Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: International Conference on Computer Vision, pp. 2223\u20132232 (2017)","DOI":"10.1109\/ICCV.2017.244"},{"key":"42_CR69","unstructured":"Zhu, J.Y., et al.: Toward multimodal image-to-image translation. In: Advances in Neural Information Processing Systems 30 (2017)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19781-9_42","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,12]],"date-time":"2024-03-12T16:46:32Z","timestamp":1710261992000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19781-9_42"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031197802","9783031197819"],"references-count":69,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19781-9_42","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"23 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}