{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T17:28:08Z","timestamp":1777656488550,"version":"3.51.4"},"publisher-location":"Cham","reference-count":77,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031730290","type":"print"},{"value":"9783031730306","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,11,24]],"date-time":"2024-11-24T00:00:00Z","timestamp":1732406400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,24]],"date-time":"2024-11-24T00:00:00Z","timestamp":1732406400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-73030-6_25","type":"book-chapter","created":{"date-parts":[[2024,11,25]],"date-time":"2024-11-25T16:57:42Z","timestamp":1732553862000},"page":"446-465","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":13,"title":["Photorealistic Object Insertion with\u00a0Diffusion-Guided Inverse Rendering"],"prefix":"10.1007","author":[{"given":"Ruofan","family":"Liang","sequence":"first","affiliation":[]},{"given":"Zan","family":"Gojcic","sequence":"additional","affiliation":[]},{"given":"Merlin","family":"Nimier-David","sequence":"additional","affiliation":[]},{"given":"David","family":"Acuna","sequence":"additional","affiliation":[]},{"given":"Nandita","family":"Vijaykumar","sequence":"additional","affiliation":[]},{"given":"Sanja","family":"Fidler","sequence":"additional","affiliation":[]},{"given":"Zian","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,24]]},"reference":[{"key":"25_CR1","unstructured":"Poly Haven - The Public 3D Asset Library. https:\/\/polyhaven.com"},{"key":"25_CR2","unstructured":"Balaji, Y., et al.: eDiff-i: text-to-image diffusion models with ensemble of expert denoisers. arXiv preprint arXiv:2211.01324 (2022)"},{"key":"25_CR3","doi-asserted-by":"publisher","unstructured":"Bangaru, S.P., Li, T.M., Durand, F.: Unbiased warped-area sampling for differentiable rendering. ACM Trans. Graph. 39(6) (2020). https:\/\/doi.org\/10.1145\/3414685.3417833","DOI":"10.1145\/3414685.3417833"},{"issue":"8","key":"25_CR4","doi-asserted-by":"publisher","first-page":"1670","DOI":"10.1109\/TPAMI.2014.2377712","volume":"37","author":"JT Barron","year":"2014","unstructured":"Barron, J.T., Malik, J.: Shape, illumination, and reflectance from shading. IEEE Trans. Pattern Anal. Mach. Intell. 37(8), 1670\u20131687 (2014)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"25_CR5","first-page":"3","volume":"2","author":"H Barrow","year":"1978","unstructured":"Barrow, H., Tenenbaum, J., Hanson, A., Riseman, E.: Recovering intrinsic scene characteristics. Comput. Vis. Syst 2, 3\u201326 (1978)","journal-title":"Comput. Vis. Syst"},{"issue":"4","key":"25_CR6","doi-asserted-by":"publisher","first-page":"159","DOI":"10.1145\/2601097.2601206","volume":"33","author":"S Bell","year":"2014","unstructured":"Bell, S., Bala, K., Snavely, N.: Intrinsic images in the wild. ACM Trans. Graph. (TOG) 33(4), 159 (2014)","journal-title":"ACM Trans. Graph. (TOG)"},{"issue":"1","key":"25_CR7","doi-asserted-by":"publisher","first-page":"75","DOI":"10.1006\/cviu.1996.0006","volume":"63","author":"MJ Black","year":"1996","unstructured":"Black, M.J., Anandan, P.: The robust estimation of multiple motions: parametric and piecewise-smooth flow fields. Comput. Vis. Image Underst. 63(1), 75\u2013104 (1996)","journal-title":"Comput. Vis. Image Underst."},{"key":"25_CR8","doi-asserted-by":"crossref","unstructured":"Boss, M., Jampani, V., Kim, K., Lensch, H.P., Kautz, J.: Two-shot spatially-varying BRDF and shape estimation. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00404"},{"key":"25_CR9","doi-asserted-by":"publisher","first-page":"130","DOI":"10.1145\/1618452.1618476","volume":"28","author":"A Bousseau","year":"2009","unstructured":"Bousseau, A., Paris, S., Durand, F.: User-assisted intrinsic images. ACM Trans. Graph. (TOG) 28, 130 (2009)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"25_CR10","unstructured":"Chari, P., et al.: Personalized restoration via dual-pivot tuning. arXiv preprint arXiv:2312.17234 (2023)"},{"key":"25_CR11","unstructured":"Dai, X., et\u00a0al.: Emu: enhancing image generation models using photogenic needles in a haystack. arXiv preprint arXiv:2309.15807 (2023)"},{"key":"25_CR12","unstructured":"Dastjerdi, M.R.K., Eisenmann, J., Hold-Geoffroy, Y., Lalonde, J.F.: EverLight: indoor-outdoor editable HDR lighting estimation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 7420\u20137429 (2023)"},{"key":"25_CR13","unstructured":"Durkan, C., Bekasov, A., Murray, I., Papamakarios, G.: Neural spline flows. In: Advances in Neural Information Processing Systems, vol. 32 (2019)"},{"key":"25_CR14","doi-asserted-by":"publisher","unstructured":"Gal, R., Alaluf, Y., Atzmon, Y., Patashnik, O., Bermano, A.H., Chechik, G., Cohen-Or, D.: An image is worth one word: personalizing text-to-image generation using textual inversion (2022). https:\/\/doi.org\/10.48550\/ARXIV.2208.01618","DOI":"10.48550\/ARXIV.2208.01618"},{"key":"25_CR15","doi-asserted-by":"crossref","unstructured":"Gardner, M.A., Hold-Geoffroy, Y., Sunkavalli, K., Gagn\u00e9, C., Lalonde, J.F.: Deep parametric indoor lighting estimation. In: ICCV, pp. 7175\u20137183 (2019)","DOI":"10.1109\/ICCV.2019.00727"},{"key":"25_CR16","doi-asserted-by":"crossref","unstructured":"Gardner, M.A., et al.: Learning to predict indoor illumination from a single image. arXiv preprint arXiv:1704.00090 (2017)","DOI":"10.1145\/3130800.3130891"},{"key":"25_CR17","doi-asserted-by":"crossref","unstructured":"Garon, M., Sunkavalli, K., Hadap, S., Carr, N., Lalonde, J.F.: Fast spatially-varying indoor lighting estimation. In: CVPR, pp. 6908\u20136917 (2019)","DOI":"10.1109\/CVPR.2019.00707"},{"key":"25_CR18","doi-asserted-by":"crossref","unstructured":"Grosse, R., Johnson, M.K., Adelson, E.H., Freeman, W.T.: Ground truth dataset and baseline evaluations for intrinsic image algorithms. In: ICCV, pp. 2335\u20132342. IEEE (2009)","DOI":"10.1109\/ICCV.2009.5459428"},{"key":"25_CR19","doi-asserted-by":"crossref","unstructured":"Haque, A., Tancik, M., Efros, A., Holynski, A., Kanazawa, A.: Instruct-NeRF2NeRF: editing 3D scenes with instructions. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (2023)","DOI":"10.1109\/ICCV51070.2023.01808"},{"key":"25_CR20","unstructured":"Ho, J., Jain, A., Abbeel, P.: Denoising diffusion probabilistic models. arXiv preprint arxiv:2006.11239 (2020)"},{"key":"25_CR21","unstructured":"Ho, J., Salimans, T.: Classifier-free diffusion guidance. arXiv preprint arXiv:2207.12598 (2022)"},{"key":"25_CR22","doi-asserted-by":"crossref","unstructured":"Hold-Geoffroy, Y., Athawale, A., Lalonde, J.F.: Deep sky modeling for single image outdoor lighting estimation. In: CVPR, pp. 6927\u20136935 (2019)","DOI":"10.1109\/CVPR.2019.00709"},{"key":"25_CR23","doi-asserted-by":"crossref","unstructured":"Hold-Geoffroy, Y., Sunkavalli, K., Hadap, S., Gambaretto, E., Lalonde, J.F.: Deep outdoor illumination estimation. In: CVPR, pp. 7312\u20137321 (2017)","DOI":"10.1109\/CVPR.2017.255"},{"key":"25_CR24","unstructured":"Hu, E.J., et al.: LoRA: low-rank adaptation of large language models. In: International Conference on Learning Representations (2022). https:\/\/openreview.net\/forum?id=nZeVKeeFYf9"},{"key":"25_CR25","unstructured":"Jakob, W., et al.: Mitsuba 3 renderer (2022). https:\/\/mitsuba-renderer.org"},{"key":"25_CR26","doi-asserted-by":"crossref","unstructured":"Kajiya, J.T.: The rendering equation. In: Proceedings of the 13th Annual Conference on Computer Graphics and Interactive Techniques, pp. 143\u2013150 (1986)","DOI":"10.1145\/15922.15902"},{"key":"25_CR27","doi-asserted-by":"crossref","unstructured":"Karimi\u00a0Dastjerdi, M.R., Hold-Geoffroy, Y., Eisenmann, J., Khodadadeh, S., Lalonde, J.F.: Guided co-modulated GAN for 360 field of view extrapolation. In: International Conference on 3D Vision (3DV) (2022)","DOI":"10.1109\/3DV57658.2022.00059"},{"key":"25_CR28","doi-asserted-by":"crossref","unstructured":"Ke, B., Obukhov, A., Huang, S., Metzger, N., Daudt, R.C., Schindler, K.: Repurposing diffusion-based image generators for monocular depth estimation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2024)","DOI":"10.1109\/CVPR52733.2024.00907"},{"key":"25_CR29","doi-asserted-by":"crossref","unstructured":"Kocsis, P., Sitzmann, V., Nie\u00dfner, M.: Intrinsic image diffusion for single-view material estimation. In: arxiv (2023)","DOI":"10.1109\/CVPR52733.2024.00497"},{"key":"25_CR30","doi-asserted-by":"crossref","unstructured":"Kovacs, B., Bell, S., Snavely, N., Bala, K.: Shading annotations in the wild. In: CVPR, pp. 6998\u20137007 (2017)","DOI":"10.1109\/CVPR.2017.97"},{"issue":"1","key":"25_CR31","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1364\/JOSA.61.000001","volume":"61","author":"EH Land","year":"1971","unstructured":"Land, E.H., McCann, J.J.: Lightness and retinex theory. Josa 61(1), 1\u201311 (1971)","journal-title":"Josa"},{"key":"25_CR32","doi-asserted-by":"crossref","unstructured":"LeGendre, C., et al.: DeepLight: learning illumination for unconstrained mobile mixed reality. In: CVPR, pp. 5918\u20135928 (2019)","DOI":"10.1109\/CVPR.2019.00607"},{"key":"25_CR33","doi-asserted-by":"publisher","unstructured":"Li, T.M., Aittala, M., Durand, F., Lehtinen, J.: Differentiable monte Carlo ray tracing through edge sampling. ACM Trans. Graph. 37(6) (2018). https:\/\/doi.org\/10.1145\/3272127.3275109","DOI":"10.1145\/3272127.3275109"},{"key":"25_CR34","doi-asserted-by":"crossref","unstructured":"Li, Z., Snavely, N.: CGintrinsics: better intrinsic image decomposition through physically-based rendering. In: ECCV, pp. 371\u2013387 (2018)","DOI":"10.1007\/978-3-030-01219-9_23"},{"key":"25_CR35","doi-asserted-by":"crossref","unstructured":"Li, Z., Shafiei, M., Ramamoorthi, R., Sunkavalli, K., Chandraker, M.: Inverse rendering for complex indoor scenes: shape, spatially-varying lighting and SVBRDF from a single image. In: CVPR, pp. 2475\u20132484 (2020)","DOI":"10.1109\/CVPR42600.2020.00255"},{"issue":"6","key":"25_CR36","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3272127.3275055","volume":"37","author":"Z Li","year":"2018","unstructured":"Li, Z., Xu, Z., Ramamoorthi, R., Sunkavalli, K., Chandraker, M.: Learning to reconstruct shape and spatially-varying reflectance from a single image. ACM Trans. Graph. (TOG) 37(6), 1\u201311 (2018)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"25_CR37","doi-asserted-by":"publisher","unstructured":"Li, Z., Yu, L., Okunev, M., Chandraker, M., Dong, Z.: Spatiotemporally consistent HDR indoor lighting estimation. ACM Trans. Graph. 42(3) (2023). https:\/\/doi.org\/10.1145\/3595921","DOI":"10.1145\/3595921"},{"key":"25_CR38","unstructured":"Li, Z., et\u00a0al.: OpenRooms: an end-to-end open framework for photorealistic indoor scene datasets. arXiv preprint arXiv:2007.12868 (2020)"},{"key":"25_CR39","doi-asserted-by":"publisher","unstructured":"Loubet, G., Holzschuch, N., Jakob, W.: Reparameterizing discontinuous integrands for differentiable rendering. ACM Trans. Graph. 38(6) (2019). https:\/\/doi.org\/10.1145\/3355089.3356510","DOI":"10.1145\/3355089.3356510"},{"key":"25_CR40","doi-asserted-by":"crossref","unstructured":"Lyu, L., et al.: Diffusion posterior illumination for ambiguity-aware inverse rendering. ACM Trans. Graph. 42(6) (2023)","DOI":"10.1145\/3618357"},{"key":"25_CR41","unstructured":"Meng, C., et al.: SDEdit: guided image synthesis and editing with stochastic differential equations. In: International Conference on Learning Representations (2022)"},{"key":"25_CR42","doi-asserted-by":"publisher","unstructured":"Nimier-David, M., Speierer, S., Ruiz, B., Jakob, W.: Radiative backpropagation: an adjoint method for lightning-fast differentiable rendering. ACM Trans. Graph. 39(4) (2020). https:\/\/doi.org\/10.1145\/3386569.3392406","DOI":"10.1145\/3386569.3392406"},{"key":"25_CR43","doi-asserted-by":"publisher","unstructured":"Nimier-David, M., Vicini, D., Zeltner, T., Jakob, W.: Mitsuba 2: a retargetable forward and inverse renderer. ACM Trans. Graph. 38(6) (2019). https:\/\/doi.org\/10.1145\/3355089.3356498","DOI":"10.1145\/3355089.3356498"},{"key":"25_CR44","doi-asserted-by":"crossref","unstructured":"Phongthawee, P., et al.: DiffusionLight: light probes for free by painting a chrome ball. In: ArXiv (2023)","DOI":"10.1109\/CVPR52733.2024.00018"},{"key":"25_CR45","unstructured":"Poole, B., Jain, A., Barron, J.T., Mildenhall, B.: DreamFusion: Text-to-3D using 2D diffusion. arXiv (2022)"},{"key":"25_CR46","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models (2021)","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"25_CR47","doi-asserted-by":"crossref","unstructured":"Ruiz, N., Li, Y., Jampani, V., Pritch, Y., Rubinstein, M., Aberman, K.: DreamBooth: fine tuning text-to-image diffusion models for subject-driven generation (2022)","DOI":"10.1109\/CVPR52729.2023.02155"},{"key":"25_CR48","first-page":"36479","volume":"35","author":"C Saharia","year":"2022","unstructured":"Saharia, C., et al.: Photorealistic text-to-image diffusion models with deep language understanding. Adv. Neural. Inf. Process. Syst. 35, 36479\u201336494 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"25_CR49","doi-asserted-by":"crossref","unstructured":"Sarkar, A., Mai, H., Mahapatra, A., Lazebnik, S., Forsyth, D., Bhattad, A.: Shadows don\u2019t lie and lines can\u2019t bend! generative models don\u2019t know projective geometry...for now (2023)","DOI":"10.1109\/CVPR52733.2024.02658"},{"key":"25_CR50","doi-asserted-by":"crossref","unstructured":"Sengupta, S., Gu, J., Kim, K., Liu, G., Jacobs, D.W., Kautz, J.: Neural inverse rendering of an indoor scene from a single image. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00869"},{"key":"25_CR51","doi-asserted-by":"crossref","unstructured":"Shah, V., et al.: ZipLoRA: any subject in any style by effectively merging LoRAs (2023)","DOI":"10.1007\/978-3-031-73232-4_24"},{"key":"25_CR52","doi-asserted-by":"crossref","unstructured":"Shi, J., Xiong, W., Lin, Z., Jung, H.J.: InstantBooth: personalized text-to-image generation without test-time finetuning. arXiv preprint arXiv:2304.03411 (2023)","DOI":"10.1109\/CVPR52733.2024.00816"},{"key":"25_CR53","doi-asserted-by":"crossref","unstructured":"Song, S., Funkhouser, T.: Neural Illumination: lighting prediction for indoor environments. In: CVPR, pp. 6918\u20136926 (2019)","DOI":"10.1109\/CVPR.2019.00708"},{"key":"25_CR54","doi-asserted-by":"crossref","unstructured":"Srinivasan, P.P., Mildenhall, B., Tancik, M., Barron, J.T., Tucker, R., Snavely, N.: Lighthouse: predicting lighting volumes for spatially-coherent illumination. In: CVPR, pp. 8080\u20138089 (2020)","DOI":"10.1109\/CVPR42600.2020.00810"},{"key":"25_CR55","doi-asserted-by":"crossref","unstructured":"Sun, P., et\u00a0al.: Scalability in perception for autonomous driving: Waymo open dataset. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00252"},{"key":"25_CR56","doi-asserted-by":"crossref","unstructured":"Tang, J., Zhu, Y., Wang, H., Chan, J.H., Li, S., Shi, B.: Estimating spatially-varying lighting in urban scenes with disentangled representation. In: ECCV (2022)","DOI":"10.1007\/978-3-031-20068-7_26"},{"key":"25_CR57","unstructured":"Tang, L., et al.: RealFill: reference-driven generation for authentic image completion. arXiv preprint arXiv:2309.16668 (2023)"},{"key":"25_CR58","unstructured":"Tao, A., Sapra, K., Catanzaro, B.: Hierarchical multi-scale attention for semantic segmentation. arXiv preprint arXiv:2005.10821 (2020)"},{"key":"25_CR59","doi-asserted-by":"crossref","unstructured":"Veach, E., Guibas, L.J.: Optimally combining sampling techniques for monte Carlo rendering. In: Proceedings of the 22nd Annual Conference on Computer Graphics and Interactive Techniques, pp. 419\u2013428 (1995)","DOI":"10.1145\/218380.218498"},{"key":"25_CR60","doi-asserted-by":"publisher","unstructured":"Vicini, D., Speierer, S., Jakob, W.: Path replay backpropagation: differentiating light paths using constant memory and linear time. ACM Trans. Graph. 40(4) (2021). https:\/\/doi.org\/10.1145\/3450626.3459804","DOI":"10.1145\/3450626.3459804"},{"key":"25_CR61","doi-asserted-by":"crossref","unstructured":"Wang, G., Yang, Y., Loy, C.C., Liu, Z.: StyleLight: HDR panorama generation for lighting estimation and editing. In: European Conference on Computer Vision (ECCV) (2022)","DOI":"10.1007\/978-3-031-19784-0_28"},{"key":"25_CR62","doi-asserted-by":"crossref","unstructured":"Wang, Z., Chen, W., Acuna, D., Kautz, J., Fidler, S.: Neural light field estimation for street scenes with differentiable virtual object insertion. In: ECCV (2022)","DOI":"10.1007\/978-3-031-20086-1_22"},{"key":"25_CR63","doi-asserted-by":"crossref","unstructured":"Wang, Z., Philion, J., Fidler, S., Kautz, J.: Learning indoor inverse rendering with 3D spatially-varying lighting. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.01231"},{"key":"25_CR64","doi-asserted-by":"crossref","unstructured":"Wimbauer, F., Wu, S., Rupprecht, C.: De-rendering 3D objects in the wild. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01794"},{"key":"25_CR65","doi-asserted-by":"publisher","unstructured":"Yan, K., Lassner, C., Budge, B., Dong, Z., Zhao, S.: Efficient estimation of boundary integrals for path-space differentiable rendering. ACM Trans. Graph. 41(4) (2022). https:\/\/doi.org\/10.1145\/3528223.3530080","DOI":"10.1145\/3528223.3530080"},{"key":"25_CR66","unstructured":"Yang, J., et al.: EmerNeRF: emergent spatial-temporal scene decomposition via self-supervision. arXiv preprint arXiv:2311.02077 (2023)"},{"key":"25_CR67","doi-asserted-by":"crossref","unstructured":"Yu, H.X., et al.: Accidental light probes. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12521\u201312530 (2023)","DOI":"10.1109\/CVPR52729.2023.01205"},{"key":"25_CR68","unstructured":"Yu, X., Guo, Y.C., Li, Y., Liang, D., Zhang, S.H., Qi, X.: Text-to-3D with classifier score distillation (2023)"},{"key":"25_CR69","doi-asserted-by":"crossref","unstructured":"Yu, Y., Smith, W.A.: InverseRenderNet: learning single image inverse rendering. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00327"},{"key":"25_CR70","doi-asserted-by":"crossref","unstructured":"Zhan, F., et al.: EMLight: lighting estimation via spherical distribution approximation. In: Proceedings of the AAAI Conference on Artificial Intelligence (2021)","DOI":"10.1609\/aaai.v35i4.16440"},{"key":"25_CR71","doi-asserted-by":"publisher","unstructured":"Zhang, C., Miller, B., Yan, K., Gkioulekas, I., Zhao, S.: Path-space differentiable rendering. ACM Trans. Graph. 39(4) (2020). https:\/\/doi.org\/10.1145\/3386569.3392383","DOI":"10.1145\/3386569.3392383"},{"key":"25_CR72","doi-asserted-by":"publisher","unstructured":"Zhang, C., Yu, Z., Zhao, S.: Path-space differentiable rendering of participating media. ACM Trans. Graph. 40(4) (2021). https:\/\/doi.org\/10.1145\/3450626.3459782","DOI":"10.1145\/3450626.3459782"},{"key":"25_CR73","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 586\u2013595 (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"25_CR74","doi-asserted-by":"publisher","unstructured":"Zhang, Z., Roussel, N., Jakob, W.: Projective sampling for differentiable rendering of geometry. ACM Trans. Graph. 42(6) (2023). https:\/\/doi.org\/10.1145\/3618385","DOI":"10.1145\/3618385"},{"key":"25_CR75","doi-asserted-by":"crossref","unstructured":"Zhao, Q., Tan, P., Dai, Q., Shen, L., Wu, E., Lin, S.: A closed-form solution to retinex with nonlocal texture constraints. IEEE Trans. Pattern Anal. Mach. Intell. 34(7), 1437\u20131444 (2012)","DOI":"10.1109\/TPAMI.2012.77"},{"key":"25_CR76","doi-asserted-by":"crossref","unstructured":"Zhao, Y., Guo, T.: POINTAR: efficient lighting estimation for mobile augmented reality. arXiv preprint arXiv:2004.00006 (2020)","DOI":"10.1007\/978-3-030-58592-1_40"},{"key":"25_CR77","doi-asserted-by":"crossref","unstructured":"Zhu, Y., Zhang, Y., Li, S., Shi, B.: Spatially-varying outdoor lighting estimation from intrinsics. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.01264"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-73030-6_25","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,25]],"date-time":"2024-11-25T17:19:12Z","timestamp":1732555152000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-73030-6_25"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,24]]},"ISBN":["9783031730290","9783031730306"],"references-count":77,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-73030-6_25","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,24]]},"assertion":[{"value":"24 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}