{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,25]],"date-time":"2026-03-25T18:19:47Z","timestamp":1774462787236,"version":"3.50.1"},"reference-count":54,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T00:00:00Z","timestamp":1772064000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T00:00:00Z","timestamp":1772064000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62206071"],"award-info":[{"award-number":["62206071"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100007847","name":"Natural Science Foundation of Jilin Province","doi-asserted-by":"publisher","award":["YDZJ202501ZYTS586"],"award-info":[{"award-number":["YDZJ202501ZYTS586"]}],"id":[{"id":"10.13039\/100007847","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","award":["3072025ZX0402"],"award-info":[{"award-number":["3072025ZX0402"]}],"id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2026,3]]},"DOI":"10.1007\/s00371-026-04392-3","type":"journal-article","created":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T05:45:22Z","timestamp":1772084722000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Enhancing novel view synthesis with random patch radiance fields"],"prefix":"10.1007","volume":"42","author":[{"given":"Weijie","family":"Ren","sequence":"first","affiliation":[]},{"given":"Guohui","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Kang","family":"Zou","sequence":"additional","affiliation":[]},{"given":"Zhiyu","family":"Lyu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,2,26]]},"reference":[{"issue":"1","key":"4392_CR1","doi-asserted-by":"publisher","first-page":"532","DOI":"10.1109\/TNNLS.2022.3175775","volume":"35","author":"A Karambakhsh","year":"2024","unstructured":"Karambakhsh, A., Sheng, B., Li, P., Li, H., Kim, J., Jung, Y., Chen, C.L.P.: SparseVoxNet: 3-D object recognition with sparsely aggregation of 3-D dense blocks. IEEE Trans. Neural Netw. Learn. Syst. 35(1), 532\u2013546 (2024). https:\/\/doi.org\/10.1109\/TNNLS.2022.3175775","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"issue":"1","key":"4392_CR2","doi-asserted-by":"publisher","first-page":"2201","DOI":"10.1002\/cav.2201","volume":"35","author":"X Zhu","year":"2024","unstructured":"Zhu, X., Yao, X., Zhang, J., Zhu, M., You, L., Yang, X., Zhang, J., Zhao, H., Zeng, D.: TMSDNet: transformer with multi-scale dense network for single and multi-view 3D reconstruction. Comput. Anim. Virtual Worlds 35(1), 2201 (2024). https:\/\/doi.org\/10.1002\/cav.2201","journal-title":"Comput. Anim. Virtual Worlds"},{"issue":"1","key":"4392_CR3","doi-asserted-by":"publisher","first-page":"243","DOI":"10.1007\/s00371-024-03321-6","volume":"41","author":"F Bian","year":"2025","unstructured":"Bian, F., Xiong, S., Yi, R., Ma, L.: Multi-view stereo-regulated NeRF for urban scene novel view synthesis. Vis. Comput. 41(1), 243\u2013255 (2025). https:\/\/doi.org\/10.1007\/s00371-024-03321-6","journal-title":"Vis. Comput."},{"issue":"1","key":"4392_CR4","doi-asserted-by":"publisher","first-page":"2203","DOI":"10.1002\/cav.2203","volume":"35","author":"Y Liu","year":"2024","unstructured":"Liu, Y., Huang, E., Zhou, Z., Wang, K., Liu, S.: 3D facial attractiveness prediction based on deep feature fusion. Comput. Anim. Virtual Worlds 35(1), 2203 (2024). https:\/\/doi.org\/10.1002\/cav.2203","journal-title":"Comput. Anim. Virtual Worlds"},{"issue":"1","key":"4392_CR5","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1145\/3503250","volume":"65","author":"B Mildenhall","year":"2021","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: Nerf: representing scenes as neural radiance fields for view synthesis. Commun. ACM 65(1), 99\u2013106 (2021). https:\/\/doi.org\/10.1145\/3503250","journal-title":"Commun. ACM"},{"key":"4392_CR6","doi-asserted-by":"crossref","unstructured":"Mescheder, L., Oechsle, M., Niemeyer, M., Nowozin, S., Geiger, A.: Occupancy networks: learning 3D reconstruction in function space. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4460\u20134470 (2019)","DOI":"10.1109\/CVPR.2019.00459"},{"issue":"5","key":"4392_CR7","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s00371-025-03974-x","volume":"41","author":"T Bao","year":"2025","unstructured":"Bao, T., Ye, J., Bao, Z., Leow, C.S., Hu, H., Lu, J., Fujishiro, I., Xu, J.: L2H-NeRF: low- to high-frequency-guided NeRF for 3D reconstruction with a few input scenes. Vis. Comput. 41(5), 1\u201312 (2025). https:\/\/doi.org\/10.1007\/s00371-025-03974-x","journal-title":"Vis. Comput."},{"issue":"7","key":"4392_CR8","doi-asserted-by":"publisher","first-page":"4851","DOI":"10.1007\/s00371-024-03694-8","volume":"41","author":"J Wang","year":"2025","unstructured":"Wang, J., Li, H., Chen, Y.: Seg-invRender: fusing semantic segmentation based on NeRF for inverse rendering considering shadows. Vis. Comput. 41(7), 4851\u20134864 (2025). https:\/\/doi.org\/10.1007\/s00371-024-03694-8","journal-title":"Vis. Comput."},{"issue":"1","key":"4392_CR9","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.vrih.2022.11.001","volume":"5","author":"C Hoshizawa","year":"2023","unstructured":"Hoshizawa, C., Komuro, T.: View interpolation networks for reproducing the material appearance of specular objects. Virtual Real. Intel. Hardw. 5(1), 1\u201310 (2023). https:\/\/doi.org\/10.1016\/j.vrih.2022.11.001","journal-title":"Virtual Real. Intel. Hardw."},{"key":"4392_CR10","doi-asserted-by":"publisher","first-page":"50","DOI":"10.1109\/TMM.2021.3120873","volume":"25","author":"X Lin","year":"2023","unstructured":"Lin, X., Sun, S., Huang, W., Sheng, B., Li, P., Feng, D.D.: EAPT: efficient attention pyramid transformer for image processing. IEEE Trans. Multimed. 25, 50\u201361 (2023). https:\/\/doi.org\/10.1109\/TMM.2021.3120873","journal-title":"IEEE Trans. Multimed."},{"key":"4392_CR11","doi-asserted-by":"crossref","unstructured":"Jain, A., Tancik, M., Abbeel, P.: Putting NeRF on a Diet: Semantically consistent Few-Shot view synthesis. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 5885\u20135894 (2021)","DOI":"10.1109\/ICCV48922.2021.00583"},{"issue":"2","key":"4392_CR12","doi-asserted-by":"publisher","first-page":"991","DOI":"10.1007\/s00371-024-03379-2","volume":"41","author":"J Yuan","year":"2025","unstructured":"Yuan, J., Fan, M., Liu, Z., Han, T., Kuang, Z., Pan, C., Ding, J.: Collaborative neural radiance fields for novel view synthesis. Vis. Comput. 41(2), 991\u20131006 (2025). https:\/\/doi.org\/10.1007\/s00371-024-03379-2","journal-title":"Vis. Comput."},{"key":"4392_CR13","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2024.110290","volume":"150","author":"Z Lyu","year":"2024","unstructured":"Lyu, Z., Chen, Y., Hou, Y.: MCPNet: Multi-space color correction and features prior fusion for single-image dehazing in non-homogeneous haze scenarios. Pattern Recogn. 150, 110290 (2024). https:\/\/doi.org\/10.1016\/j.patcog.2024.110290","journal-title":"Pattern Recogn."},{"key":"4392_CR14","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TGRS.2024.3357800","volume":"62","author":"J Fan","year":"2024","unstructured":"Fan, J., Sui, Z., Wang, X.: Multiphysical interpretable deep learning network for oil spill identification based on SAR images. IEEE Trans. Geosci. Remote Sens. 62, 1\u201315 (2024). https:\/\/doi.org\/10.1109\/TGRS.2024.3357800","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"4392_CR15","doi-asserted-by":"publisher","first-page":"34","DOI":"10.1007\/s11554-025-01833-5","volume":"23","author":"W Ren","year":"2026","unstructured":"Ren, W., Zhang, S., Zhou, Z.: Yolo11-mgnb: lightweight real-time small object detection algorithm for uav remote sensing images. J. Real-Time Image Proc. 23, 34 (2026)","journal-title":"J. Real-Time Image Proc."},{"key":"4392_CR16","doi-asserted-by":"publisher","first-page":"2271","DOI":"10.7717\/peerj-cs.2271","volume":"10","author":"Y Tian","year":"2024","unstructured":"Tian, Y., Liu, Y.W., Lin, B.H., Li, P.: Research on marine flexible biological target detection based on improved yolov8 algorithm. PeerJ Comput. Sci. 10, 2271 (2024). https:\/\/doi.org\/10.7717\/peerj-cs.2271","journal-title":"PeerJ Comput. Sci."},{"key":"4392_CR17","first-page":"15651","volume":"33","author":"L Liu","year":"2020","unstructured":"Liu, L., Gu, J., Zaw Lin, K., Chua, T.-S., Theobalt, C.: Neural sparse voxel fields. Adv. Neural. Inf. Process. Syst. 33, 15651\u201315663 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"issue":"4","key":"4392_CR18","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3072959.3073599","volume":"36","author":"A Knapitsch","year":"2017","unstructured":"Knapitsch, A., Park, J., Zhou, Q.-Y., Koltun, V.: Tanks and temples: benchmarking large-scale scene reconstruction. ACM Trans. Graph. 36(4), 1\u201313 (2017). https:\/\/doi.org\/10.1145\/3072959.3073599","journal-title":"ACM Trans. Graph."},{"key":"4392_CR19","doi-asserted-by":"publisher","DOI":"10.1016\/j.rse.2020.111666","volume":"239","author":"BU Meinen","year":"2020","unstructured":"Meinen, B.U., Robinson, D.T.: Mapping erosion and deposition in an agricultural landscape: optimization of UAV image acquisition schemes for SfM-MVS. Remote Sens. Environ. 239, 111666 (2020). https:\/\/doi.org\/10.1016\/j.rse.2020.111666","journal-title":"Remote Sens. Environ."},{"key":"4392_CR20","first-page":"31","volume":"42","author":"A Mahapatra","year":"2016","unstructured":"Mahapatra, A., Sa, P.K., Majhi, B., Padhy, S.: MVS: a multi-view video synopsis framework. Signal Proces.: Image Commun. 42, 31\u201344 (2016)","journal-title":"Signal Proces.: Image Commun."},{"key":"4392_CR21","doi-asserted-by":"crossref","unstructured":"Rematas, K., Liu, A., Srinivasan, P.P., Barron, J.T., Tagliasacchi, A., Funkhouser, T., Ferrari, V.: Urban radiance fields. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12932\u201312942 (2022)","DOI":"10.1109\/CVPR52688.2022.01259"},{"key":"4392_CR22","doi-asserted-by":"crossref","unstructured":"Reiser, C., Peng, S., Liao, Y., Geiger, A.: KiloNeRF: speeding up neural radiance fields with thousands of tiny MLPs. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 14335\u201314345 (2021)","DOI":"10.1109\/ICCV48922.2021.01407"},{"key":"4392_CR23","doi-asserted-by":"crossref","unstructured":"Hu, T., Liu, S., Chen, Y., Shen, T., Jia, J.: EfficientNeRF efficient neural radiance fields. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12902\u201312911 (2022)","DOI":"10.1109\/CVPR52688.2022.01256"},{"key":"4392_CR24","doi-asserted-by":"crossref","unstructured":"Ma, L., Li, X., Liao, J., Zhang, Q., Wang, X., Wang, J., Sander, P.V.: Deblur-NeRF: neural radiance fields from blurry images. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12861\u201312870 (2022)","DOI":"10.1109\/CVPR52688.2022.01252"},{"key":"4392_CR25","doi-asserted-by":"crossref","unstructured":"Deng, K., Liu, A., Zhu, J.-Y., Ramanan, D.: Depth-Supervised NeRF: fewer views and faster training for free. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12882\u201312891 (2022)","DOI":"10.1109\/CVPR52688.2022.01254"},{"key":"4392_CR26","doi-asserted-by":"publisher","first-page":"623","DOI":"10.1007\/978-3-030-58529-7_37","volume-title":"Computer Vision - ECCV 2020","author":"G Riegler","year":"2020","unstructured":"Riegler, G., Koltun, V.: Free View Synthesis. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) Computer Vision - ECCV 2020, pp. 623\u2013640. Springer, Cham (2020)"},{"key":"4392_CR27","doi-asserted-by":"crossref","unstructured":"Sarlin, P.-E., DeTone, D., Malisiewicz, T., Rabinovich, A.: SuperGlue: learning feature matching with graph neural networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4938\u20134947 (2020)","DOI":"10.1109\/CVPR42600.2020.00499"},{"key":"4392_CR28","doi-asserted-by":"crossref","unstructured":"Sun, J., Shen, Z., Wang, Y., Bao, H., Zhou, X.: LoFTR: detector-free local feature matching with transformers. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8922\u20138931 (2021)","DOI":"10.1109\/CVPR46437.2021.00881"},{"issue":"4","key":"4392_CR29","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3528223.3530127","volume":"41","author":"T M\u00fcller","year":"2022","unstructured":"M\u00fcller, T., Evans, A., Schied, C., Keller, A.: Instant neural graphics primitives with a multiresolution hash encoding. ACM Trans. Graph. 41(4), 1\u201315 (2022). https:\/\/doi.org\/10.1145\/3528223.3530127","journal-title":"ACM Trans. Graph."},{"key":"4392_CR30","doi-asserted-by":"crossref","unstructured":"Barron, J.T., Mildenhall, B., Tancik, M., Hedman, P., Martin-Brualla, R., Srinivasan, P.P.: Mip-NeRF: a multiscale representation for Anti-aliasing neural radiance fields. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 5855\u20135864 (2021)","DOI":"10.1109\/ICCV48922.2021.00580"},{"key":"4392_CR31","doi-asserted-by":"crossref","unstructured":"Barron, J.T., Mildenhall, B., Verbin, D., Srinivasan, P.P., Hedman, P.: Mip-NeRF 360: unbounded Anti-aliased neural radiance fields. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5470\u20135479 (2022)","DOI":"10.1109\/CVPR52688.2022.00539"},{"key":"4392_CR32","doi-asserted-by":"crossref","unstructured":"Niemeyer, M., Barron, J.T., Mildenhall, B., Sajjadi, M.S.M., Geiger, A., Radwan, N.: RegNeRF: regularizing neural radiance fields for view synthesis from sparse inputs. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5480\u20135490 (2022)","DOI":"10.1109\/CVPR52688.2022.00540"},{"key":"4392_CR33","doi-asserted-by":"crossref","unstructured":"Yu, A., Ye, V., Tancik, M., Kanazawa, A.: pixelNeRF: neural radiance fields from one or few images. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4578\u20134587 (2021)","DOI":"10.1109\/CVPR46437.2021.00455"},{"key":"4392_CR34","doi-asserted-by":"publisher","unstructured":"Li, J., Shang, S., Shao, L.: MetaNER: Named Entity Recognition with Meta-Learning. In: Proceedings of The Web Conference 2020. WWW \u201920, pp. 429\u2013440. Association for Computing Machinery, New York, NY, USA (2020). https:\/\/doi.org\/10.1145\/3366423.3380127","DOI":"10.1145\/3366423.3380127"},{"key":"4392_CR35","doi-asserted-by":"publisher","unstructured":"Chen, Y., Chen, Z., Zhang, C., Wang, F., Yang, X., Wang, Y., Cai, Z., Yang, L., Liu, H., Lin, G.: GaussianEditor: swift and controllable 3D editing with Gaussian splatting. In: 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 21476\u201321485 (2024). https:\/\/doi.org\/10.1109\/CVPR52733.2024.02029","DOI":"10.1109\/CVPR52733.2024.02029"},{"key":"4392_CR36","doi-asserted-by":"publisher","unstructured":"Zou, C., Ma, Q., Wang, J., Lu, M., Zhang, S., He, Z.: GaussianEnhancer: a general rendering enhancer for Gaussian splatting. In: ICASSP 2025 - 2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1\u20135 (2025). https:\/\/doi.org\/10.1109\/ICASSP49660.2025.10888633","DOI":"10.1109\/ICASSP49660.2025.10888633"},{"key":"4392_CR37","doi-asserted-by":"crossref","unstructured":"Martin-Brualla, R., Radwan, N., Sajjadi, M.S.M., Barron, J.T., Dosovitskiy, A., Duckworth, D.: NeRF in the wild: neural radiance fields for unconstrained photo collections. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7210\u20137219 (2021)","DOI":"10.1109\/CVPR46437.2021.00713"},{"key":"4392_CR38","doi-asserted-by":"crossref","unstructured":"Kim, M., Seo, S., Han, B.: InfoNeRF: ray entropy minimization for few-shot neural volume rendering. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12912\u201312921 (2022)","DOI":"10.1109\/CVPR52688.2022.01257"},{"key":"4392_CR39","doi-asserted-by":"crossref","unstructured":"Chen, A., Xu, Z., Zhao, F., Zhang, X., Xiang, F., Yu, J., Su, H.: MVSNeRF: fast generalizable radiance field reconstruction from Multi-view stereo. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 14124\u201314133 (2021)","DOI":"10.1109\/ICCV48922.2021.01386"},{"key":"4392_CR40","doi-asserted-by":"publisher","unstructured":"Tang et al., PaReNeRF: Toward Fast Large-Scale Dynamic NeRF with Patch-Based Reference, IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, pp. 5428\u20135438 (2024), https:\/\/doi.org\/10.1109\/CVPR52733.2024.00519","DOI":"10.1109\/CVPR52733.2024.00519"},{"key":"4392_CR41","doi-asserted-by":"publisher","DOI":"10.1117\/12.3026753","author":"L Swartz","year":"2024","unstructured":"Swartz, L., Kucukpinar, T., Collins, J., Massaro, R., Palaniappan, K.: PatchNeRF: localized neural radiance field training for city-scale aerial images. Int. Soc. Opt. Photonics (2024). https:\/\/doi.org\/10.1117\/12.3026753","journal-title":"Int. Soc. Opt. Photonics"},{"issue":"3","key":"4392_CR42","doi-asserted-by":"publisher","first-page":"245","DOI":"10.1145\/78964.78965","volume":"9","author":"M Levoy","year":"1990","unstructured":"Levoy, M.: Efficient ray tracing of volume data. ACM Trans. Graph. 9(3), 245\u2013261 (1990). https:\/\/doi.org\/10.1145\/78964.78965","journal-title":"ACM Trans. Graph."},{"key":"4392_CR43","doi-asserted-by":"publisher","unstructured":"Hor\u00e9, A., Ziou, D.: Image quality metrics: PSNR vs. SSIM. In: 2010 20th International Conference on Pattern Recognition, pp. 2366\u20132369 (2010). https:\/\/doi.org\/10.1109\/ICPR.2010.579","DOI":"10.1109\/ICPR.2010.579"},{"key":"4392_CR44","doi-asserted-by":"publisher","first-page":"35","DOI":"10.1007\/s11235-010-9351-x","volume":"49","author":"Q Huynh-Thu","year":"2012","unstructured":"Huynh-Thu, Q., Ghanbari, M.: The accuracy of PSNR in predicting video quality for different video scenes and frame rates. Telecommun. Syst. 49, 35\u201348 (2012)","journal-title":"Telecommun. Syst."},{"key":"4392_CR45","doi-asserted-by":"publisher","first-page":"623","DOI":"10.7717\/peerj-cs.623","volume":"7","author":"D Chicco","year":"2021","unstructured":"Chicco, D., Warrens, M.J., Jurman, G.: The coefficient of determination R-squared is more informative than SMAPE, MAE, MAPE, MSE and RMSE in regression analysis evaluation. Peerj Comput. Sci. 7, 623 (2021)","journal-title":"Peerj Comput. Sci."},{"key":"4392_CR46","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 586\u2013595 (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"4392_CR47","doi-asserted-by":"publisher","unstructured":"Sun, C., Sun, M. and Chen, H.T.: Direct voxel grid optimization: Super-fast convergence for radiance fields reconstruction. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition pp. 5459\u20135469 (2022). https:\/\/doi.org\/10.1109\/CVPR52688.2022.00538","DOI":"10.1109\/CVPR52688.2022.00538"},{"key":"4392_CR48","doi-asserted-by":"crossref","unstructured":"Lin, C.-Y., Fu, Q., Merth, T., Yang, K., Ranjan, A.: FastSR-NeRF: Improving NeRF efficiency on consumer devices with a simple super-resolution pipeline. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision (WACV), pp. 6036\u20136045 (2024)","DOI":"10.1109\/WACV57701.2024.00593"},{"issue":"6","key":"4392_CR49","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3618402","volume":"42","author":"B Roessle","year":"2023","unstructured":"Roessle, B., M\u00fcller, N., Porzi, L., Bul\u00f2, S.R., Kontschieder, P., Niessner, M.: GANeRF: leveraging discriminators to optimize neural radiance fields. ACM Trans. Graph. 42(6), 1\u201314 (2023). https:\/\/doi.org\/10.1145\/3618402","journal-title":"ACM Trans. Graph."},{"key":"4392_CR50","doi-asserted-by":"publisher","first-page":"333","DOI":"10.1007\/978-3-031-19824-3_20","volume-title":"Computer Vision - ECCV 2022","author":"A Chen","year":"2022","unstructured":"Chen, A., Xu, Z., Geiger, A., Yu, J., Su, H.: TensoRF: Tensorial Radiance Fields. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) Computer Vision - ECCV 2022, pp. 333\u2013350. Springer, Cham (2022)"},{"issue":"6","key":"4392_CR51","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3687937","volume":"43","author":"Z Yu","year":"2024","unstructured":"Yu, Z., Sattler, T., Geiger, A.: Gaussian opacity fields: efficient adaptive surface reconstruction in unbounded scenes. ACM Trans. Graph. 43(6), 1\u201313 (2024). https:\/\/doi.org\/10.1145\/3687937","journal-title":"ACM Trans. Graph."},{"key":"4392_CR52","doi-asserted-by":"publisher","unstructured":"Tancik, M., Weber, E., Ng, E., Li, R., Yi, B., Wang, T., Kristoffersen, A., Austin, J., Salahi, K., Ahuja, A., Mcallister, D., Kerr, J., Kanazawa, A.: Nerfstudio: A modular framework for neural radiance field development. In: ACM SIGGRAPH 2023 Conference Proceedings. SIGGRAPH \u201923. Association for Computing Machinery, New York, NY, USA (2023). https:\/\/doi.org\/10.1145\/3588432.3591516","DOI":"10.1145\/3588432.3591516"},{"key":"4392_CR53","doi-asserted-by":"publisher","unstructured":"Qiu, Z., Ren, C., Song, K., Zeng, X., Yang, L., Zhang, J.: Deformable NeRF using recursively subdivided Tetrahedra. In: Proceedings of the 32nd ACM International Conference on Multimedia. MM \u201924, pp. 6424\u20136432. Association for Computing Machinery, New York, NY, USA (2024). https:\/\/doi.org\/10.1145\/3664647.3681019","DOI":"10.1145\/3664647.3681019"},{"key":"4392_CR54","doi-asserted-by":"publisher","unstructured":"Xie, Z., Yang, X., Yang, Y., Sun, Q., Jiang, Y., Wang, H., Cai, Y., Sun, M.: S3im: stochastic structural similarity and its unreasonable effectiveness for neural fields. In: 2023 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 17978\u201317988 (2023). https:\/\/doi.org\/10.1109\/ICCV51070.2023.01652","DOI":"10.1109\/ICCV51070.2023.01652"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-026-04392-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-026-04392-3","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-026-04392-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,25]],"date-time":"2026-03-25T16:21:37Z","timestamp":1774455697000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-026-04392-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2,26]]},"references-count":54,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2026,3]]}},"alternative-id":["4392"],"URL":"https:\/\/doi.org\/10.1007\/s00371-026-04392-3","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,2,26]]},"assertion":[{"value":"16 June 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 January 2026","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 February 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"180"}}