{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T06:08:24Z","timestamp":1742969304659,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":34,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819988495"},{"type":"electronic","value":"9789819988501"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-99-8850-1_7","type":"book-chapter","created":{"date-parts":[[2024,2,3]],"date-time":"2024-02-03T18:02:05Z","timestamp":1706983325000},"page":"78-90","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Explicit Composition of\u00a0Neural Radiance Fields by\u00a0Learning an\u00a0Occlusion Field"],"prefix":"10.1007","author":[{"given":"Xunsen","family":"Sun","sequence":"first","affiliation":[]},{"given":"Hao","family":"Zhu","sequence":"additional","affiliation":[]},{"given":"Yuanxun","family":"Lu","sequence":"additional","affiliation":[]},{"given":"Xun","family":"Cao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,2,4]]},"reference":[{"key":"7_CR1","doi-asserted-by":"crossref","unstructured":"Barron, J.T., Mildenhall, B., Tancik, M., Hedman, P., Martin-Brualla, R., Srinivasan, P.P.: MIP-NERF: a multiscale representation for anti-aliasing neural radiance fields. In: CVPR, pp. 5855\u20135864 (2021)","DOI":"10.1109\/ICCV48922.2021.00580"},{"key":"7_CR2","doi-asserted-by":"crossref","unstructured":"Chan, E.R., et al.: Efficient geometry-aware 3d generative adversarial networks. In: CVPR, pp. 16123\u201316133 (2022)","DOI":"10.1109\/CVPR52688.2022.01565"},{"key":"7_CR3","doi-asserted-by":"crossref","unstructured":"Deng, K., Liu, A., Zhu, J.Y., Ramanan, D.: Depth-supervised nerf: fewer views and faster training for free. In: CVPR, pp. 12882\u201312891 (2022)","DOI":"10.1109\/CVPR52688.2022.01254"},{"key":"7_CR4","doi-asserted-by":"crossref","unstructured":"Gafni, G., Thies, J., Zollhofer, M., Nie\u00dfner, M.: Dynamic neural radiance fields for monocular 4d facial avatar reconstruction. In: CVPR, pp. 8649\u20138658 (2021)","DOI":"10.1109\/CVPR46437.2021.00854"},{"key":"7_CR5","doi-asserted-by":"crossref","unstructured":"Genova, K., Cole, F., Sud, A., Sarna, A., Funkhouser, T.: Local deep implicit functions for 3d shape. In: CVPR, pp. 4857\u20134866 (2020)","DOI":"10.1109\/CVPR42600.2020.00491"},{"key":"7_CR6","unstructured":"Guo, M., Fathi, A., Wu, J., Funkhouser, T.: Object-centric neural scene rendering. arXiv preprint arXiv:2012.08503 (2020)"},{"key":"7_CR7","doi-asserted-by":"crossref","unstructured":"Guo, Y.C., Kang, D., Bao, L., He, Y., Zhang, S.H.: Nerfren: neural radiance fields with reflections. In: CVPR, pp. 18409\u201318418 (2022)","DOI":"10.1109\/CVPR52688.2022.01786"},{"key":"7_CR8","doi-asserted-by":"crossref","unstructured":"Jiang, C., Sud, A., Makadia, A., Huang, J., Nie\u00dfner, M., Funkhouser, T., et al.: Local implicit grid representations for 3d scenes. In: CVPR, pp. 6001\u20136010 (2020)","DOI":"10.1109\/CVPR42600.2020.00604"},{"key":"7_CR9","doi-asserted-by":"crossref","unstructured":"Kim, M., Seo, S., Han, B.: Infonerf: ray entropy minimization for few-shot neural volume rendering. In: CVPR, pp. 12912\u201312921 (2022)","DOI":"10.1109\/CVPR52688.2022.01257"},{"key":"7_CR10","first-page":"15651","volume":"33","author":"L Liu","year":"2020","unstructured":"Liu, L., Gu, J., Zaw Lin, K., Chua, T.S., Theobalt, C.: Neural sparse voxel fields. NIPS 33, 15651\u201315663 (2020)","journal-title":"NIPS"},{"key":"7_CR11","doi-asserted-by":"crossref","unstructured":"Martin-Brualla, R., Radwan, N., Sajjadi, M.S., Barron, J.T., Dosovitskiy, A., Duckworth, D.: Nerf in the wild: neural radiance fields for unconstrained photo collections. In: CVPR, pp. 7210\u20137219 (2021)","DOI":"10.1109\/CVPR46437.2021.00713"},{"issue":"2","key":"7_CR12","first-page":"99","volume":"1","author":"N Max","year":"1995","unstructured":"Max, N.: Optical models for direct volume rendering. TVCG 1(2), 99\u2013108 (1995)","journal-title":"TVCG"},{"key":"7_CR13","doi-asserted-by":"crossref","unstructured":"Mescheder, L., Oechsle, M., Niemeyer, M., Nowozin, S., Geiger, A.: Occupancy networks: learning 3d reconstruction in function space. In: CVPR, pp. 4460\u20134470 (2019)","DOI":"10.1109\/CVPR.2019.00459"},{"issue":"1","key":"7_CR14","first-page":"99","volume":"65","author":"B Mildenhall","year":"2021","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: Nerf: representing scenes as neural radiance fields for view synthesis. ECCV 65(1), 99\u2013106 (2021)","journal-title":"ECCV"},{"issue":"4","key":"7_CR15","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3528223.3530127","volume":"41","author":"T M\u00fcller","year":"2022","unstructured":"M\u00fcller, T., Evans, A., Schied, C., Keller, A.: Instant neural graphics primitives with a multiresolution hash encoding. ToG 41(4), 1\u201315 (2022)","journal-title":"ToG"},{"key":"7_CR16","doi-asserted-by":"crossref","unstructured":"Niemeyer, M., Geiger, A.: Giraffe: representing scenes as compositional generative neural feature fields. In: CVPR, pp. 11453\u201311464 (2021)","DOI":"10.1109\/CVPR46437.2021.01129"},{"key":"7_CR17","doi-asserted-by":"crossref","unstructured":"Ost, J., Mannan, F., Thuerey, N., Knodt, J., Heide, F.: Neural scene graphs for dynamic scenes. In: CVPR, pp. 2856\u20132865 (2021)","DOI":"10.1109\/CVPR46437.2021.00288"},{"key":"7_CR18","doi-asserted-by":"crossref","unstructured":"Park, J.J., Florence, P., Straub, J., Newcombe, R., Lovegrove, S.: Deepsdf: learning continuous signed distance functions for shape representation. In: CVPR, pp. 165\u2013174 (2019)","DOI":"10.1109\/CVPR.2019.00025"},{"key":"7_CR19","doi-asserted-by":"crossref","unstructured":"Pumarola, A., Corona, E., Pons-Moll, G., Moreno-Noguer, F.: D-nerf: neural radiance fields for dynamic scenes. In: CVPR, pp. 10318\u201310327 (2021)","DOI":"10.1109\/CVPR46437.2021.01018"},{"key":"7_CR20","doi-asserted-by":"crossref","unstructured":"Rebain, D., Jiang, W., Yazdani, S., Li, K., Yi, K.M., Tagliasacchi, A.: Derf: decomposed radiance fields. In: CVPR, pp. 14153\u201314161 (2021)","DOI":"10.1109\/CVPR46437.2021.01393"},{"key":"7_CR21","doi-asserted-by":"crossref","unstructured":"Reiser, C., Peng, S., Liao, Y., Geiger, A.: Kilonerf: speeding up neural radiance fields with thousands of tiny MLPS. In: CVPR, pp. 14335\u201314345 (2021)","DOI":"10.1109\/ICCV48922.2021.01407"},{"key":"7_CR22","doi-asserted-by":"crossref","unstructured":"Roessle, B., Barron, J.T., Mildenhall, B., Srinivasan, P.P., Nie\u00dfner, M.: Dense depth priors for neural radiance fields from sparse input views. In: CVPR, pp. 12892\u201312901 (2022)","DOI":"10.1109\/CVPR52688.2022.01255"},{"key":"7_CR23","unstructured":"Smith, C., et al.: Unsupervised discovery and composition of object light fields. arXiv preprint arXiv:2205.03923 (2022)"},{"key":"7_CR24","doi-asserted-by":"crossref","unstructured":"Tancik, M., et al.: Block-nerf: scalable large scene neural view synthesis. In: CVPR, pp. 8248\u20138258 (2022)","DOI":"10.1109\/CVPR52688.2022.00807"},{"key":"7_CR25","doi-asserted-by":"crossref","unstructured":"Verbin, D., Hedman, P., Mildenhall, B., Zickler, T., Barron, J.T., Srinivasan, P.P.: Ref-nerf: structured view-dependent appearance for neural radiance fields. In: CVPR, pp. 5481\u20135490. IEEE (2022)","DOI":"10.1109\/CVPR52688.2022.00541"},{"key":"7_CR26","doi-asserted-by":"crossref","unstructured":"Wang, Z., et al.: Learning compositional radiance fields of dynamic human heads. In: CVPR, pp. 5704\u20135713 (2021)","DOI":"10.1109\/CVPR46437.2021.00565"},{"key":"7_CR27","unstructured":"Wu, T., Zhong, F., Tagliasacchi, A., Cole, F., Oztireli, C.: D2nerf: self-supervised decoupling of dynamic and static objects from a monocular video. In: NIPS (2022)"},{"key":"7_CR28","doi-asserted-by":"crossref","unstructured":"Yang, B., Zhang, Y., Xu, Y., Li, Y., Zhou, H., Bao, H., Zhang, G., Cui, Z.: Learning object-compositional neural radiance field for editable scene rendering. In: ICCV, pp. 13779\u201313788 (2021)","DOI":"10.1109\/ICCV48922.2021.01352"},{"key":"7_CR29","doi-asserted-by":"crossref","unstructured":"Yuan, W., Lv, Z., Schmidt, T., Lovegrove, S.: Star: self-supervised tracking and reconstruction of rigid objects in motion with neural rendering. In: CVPR, pp. 13144\u201313152 (2021)","DOI":"10.1109\/CVPR46437.2021.01294"},{"key":"7_CR30","unstructured":"Zhang, K., Riegler, G., Snavely, N., Koltun, V.: Nerf++: analyzing and improving neural radiance fields. arXiv preprint arXiv:2010.07492 (2020)"},{"key":"7_CR31","doi-asserted-by":"crossref","unstructured":"Zhang, X., Bi, S., Sunkavalli, K., Su, H., Xu, Z.: Nerfusion: fusing radiance fields for large-scale scene reconstruction. In: CVPR, pp. 5449\u20135458 (2022)","DOI":"10.1109\/CVPR52688.2022.00537"},{"issue":"6","key":"7_CR32","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3478513.3480500","volume":"40","author":"X Zhang","year":"2021","unstructured":"Zhang, X., Srinivasan, P.P., Deng, B., Debevec, P., Freeman, W.T., Barron, J.T.: Nerfactor: neural factorization of shape and reflectance under an unknown illumination. TOG 40(6), 1\u201318 (2021)","journal-title":"TOG"},{"key":"7_CR33","unstructured":"Zhuang, Y., et al.: Neai: pre-convoluted representation for plug-and-play neural ambient illumination. arXiv preprint arXiv:2304.08757 (2023)"},{"key":"7_CR34","doi-asserted-by":"publisher","unstructured":"Zhuang, Y., Zhu, H., Sun, X., Cao, X.: MoFaNeRF: morphable facial neural radiance field. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13663, pp. 268\u2013285. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-20062-5_16","DOI":"10.1007\/978-3-031-20062-5_16"}],"container-title":["Lecture Notes in Computer Science","Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-99-8850-1_7","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,2,3]],"date-time":"2024-02-03T18:03:15Z","timestamp":1706983395000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-99-8850-1_7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9789819988495","9789819988501"],"references-count":34,"URL":"https:\/\/doi.org\/10.1007\/978-981-99-8850-1_7","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"4 February 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"CICAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"CAAI International Conference on Artificial Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Fuzhou","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 July 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 July 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"3","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"cicai2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/cicai.caai.cn\/#\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"376","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"101","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"16","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"27% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.9","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1.9","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}