{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T09:36:35Z","timestamp":1775208995748,"version":"3.50.1"},"reference-count":83,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2026,1,28]],"date-time":"2026-01-28T00:00:00Z","timestamp":1769558400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T00:00:00Z","timestamp":1775174400000},"content-version":"vor","delay-in-days":65,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U19A2063"],"award-info":[{"award-number":["U19A2063"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100013061","name":"Jilin Provincial Scientific and Technological Development Program","doi-asserted-by":"publisher","award":["20230201080GX"],"award-info":[{"award-number":["20230201080GX"]}],"id":[{"id":"10.13039\/501100013061","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J. King Saud Univ. Comput. Inf. Sci."],"published-print":{"date-parts":[[2026,4]]},"DOI":"10.1007\/s44443-026-00491-2","type":"journal-article","created":{"date-parts":[[2026,1,28]],"date-time":"2026-01-28T05:19:00Z","timestamp":1769577540000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["F2Plenoxels: fast voxel radiance fields without neural networks for few-shot view synthesis"],"prefix":"10.1007","volume":"38","author":[{"given":"Jun","family":"Peng","sequence":"first","affiliation":[]},{"given":"Chunyi","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Yunbiao","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,28]]},"reference":[{"issue":"7","key":"491_CR1","doi-asserted-by":"publisher","first-page":"6832","DOI":"10.1109\/TCSVT.2025.3538684","volume":"35","author":"Y Bao","year":"2025","unstructured":"Bao Y, Ding T, Huo J et al (2025) 3D Gaussian splatting: Survey, technologies, challenges, and opportunities. IEEE Trans Circuits Syst Video Technol 35(7):6832\u20136852. https:\/\/doi.org\/10.1109\/TCSVT.2025.3538684","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"491_CR2","doi-asserted-by":"publisher","unstructured":"Barron JT, Mildenhall B, Tancik M et\u00a0al (2021) Mip-NeRF: A multiscale representation for anti-aliasing neural radiance fields. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), pp 5835\u20135844. https:\/\/doi.org\/10.1109\/ICCV48922.2021.00580","DOI":"10.1109\/ICCV48922.2021.00580"},{"key":"491_CR3","doi-asserted-by":"publisher","unstructured":"Barron JT, Mildenhall B, Verbin D et\u00a0al (2022) Mip-NeRF 360: Unbounded anti-aliased neural radiance fields. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 5460\u20135469. https:\/\/doi.org\/10.1109\/CVPR52688.2022.00539","DOI":"10.1109\/CVPR52688.2022.00539"},{"key":"491_CR4","doi-asserted-by":"publisher","unstructured":"Chan ER, Lin CZ, Chan MA et\u00a0al (2022) Efficient geometry-aware 3D generative adversarial networks. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 16102\u201316112. https:\/\/doi.org\/10.1109\/CVPR52688.2022.01565","DOI":"10.1109\/CVPR52688.2022.01565"},{"key":"491_CR5","doi-asserted-by":"publisher","unstructured":"Charatan D, Li SL, Tagliasacchi A et\u00a0al (2024) PixelSplat: 3D Gaussian splats from image pairs for scalable generalizable 3D reconstruction. In: 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 19457\u201319467. https:\/\/doi.org\/10.1109\/CVPR52733.2024.01840","DOI":"10.1109\/CVPR52733.2024.01840"},{"key":"491_CR6","doi-asserted-by":"publisher","unstructured":"Chen A, Xu Z, Zhao F et\u00a0al (2021) MVSNeRF: Fast generalizable radiance field reconstruction from multi-view stereo. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), pp 14104\u201314113. https:\/\/doi.org\/10.1109\/ICCV48922.2021.01386","DOI":"10.1109\/ICCV48922.2021.01386"},{"key":"491_CR7","doi-asserted-by":"publisher","unstructured":"Chen A, Xu Z, Geiger A et\u00a0al (2022) TensoRF: Tensorial radiance fields. In: Avidan S, Brostow G, Ciss\u00e9 M, et\u00a0al (eds) Computer Vision \u2013 ECCV 2022. Springer Nature Switzerland, Cham, pp 333\u2013350. https:\/\/doi.org\/10.1007\/978-3-031-19824-3_20","DOI":"10.1007\/978-3-031-19824-3_20"},{"key":"491_CR8","doi-asserted-by":"publisher","unstructured":"Chen S, Yan B, Sang X et\u00a0al (2023) Bidirectional optical flow NeRF: High accuracy and high quality under fewer views. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp 359\u2013368. https:\/\/doi.org\/10.1609\/aaai.v37i1.25109","DOI":"10.1609\/aaai.v37i1.25109"},{"key":"491_CR9","doi-asserted-by":"publisher","unstructured":"Chen Y, Xu H, Zheng C et\u00a0al (2025a) MVSplat: Efficient 3D Gaussian splatting from\u00a0sparse multi-view images. In: Leonardis A, Ricci E, Roth S et\u00a0al (eds) Computer Vision \u2013 ECCV 2024. Springer Nature Switzerland, Cham, pp 370\u2013386. https:\/\/doi.org\/10.1007\/978-3-031-72664-4_21","DOI":"10.1007\/978-3-031-72664-4_21"},{"issue":"6","key":"491_CR10","doi-asserted-by":"publisher","first-page":"4358","DOI":"10.1109\/TPAMI.2025.3543072","volume":"47","author":"Y Chen","year":"2025","unstructured":"Chen Y, Zhang J, Xie Z et al (2025) S-NeRF++: Autonomous driving simulation via neural reconstruction and generation. IEEE Trans Pattern Anal Mach Intell 47(6):4358\u20134376. https:\/\/doi.org\/10.1109\/TPAMI.2025.3543072","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"491_CR11","doi-asserted-by":"publisher","unstructured":"Deng K, Liu A, Zhu JY et\u00a0al (2022) Depth-supervised NeRF: Fewer views and faster training for free. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 12872\u201312881. https:\/\/doi.org\/10.1109\/CVPR52688.2022.01254","DOI":"10.1109\/CVPR52688.2022.01254"},{"key":"491_CR12","doi-asserted-by":"publisher","unstructured":"Edstedt J, Athanasiadis I, Wadenb\u00e4ck M et\u00a0al (2023) DKM: Dense kernelized feature matching for geometry estimation. In: 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 17765\u201317775. https:\/\/doi.org\/10.1109\/CVPR52729.2023.01704","DOI":"10.1109\/CVPR52729.2023.01704"},{"key":"491_CR13","doi-asserted-by":"publisher","unstructured":"Fridovich-Keil S, Yu A, Tancik M et\u00a0al (2022) Plenoxels: Radiance fields without neural networks. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, June 18-24, 2022. IEEE, pp 5491\u20135500. https:\/\/doi.org\/10.1109\/CVPR52688.2022.00542","DOI":"10.1109\/CVPR52688.2022.00542"},{"key":"491_CR14","doi-asserted-by":"publisher","unstructured":"Fridovich-Keil S, Meanti G, Warburg FR et\u00a0al (2023) K-Planes: Explicit radiance fields in space, time, and appearance. In: 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 12479\u201312488. https:\/\/doi.org\/10.1109\/CVPR52729.2023.01201","DOI":"10.1109\/CVPR52729.2023.01201"},{"key":"491_CR15","doi-asserted-by":"crossref","unstructured":"Gao R, Ho\u0142y\u0144ski A, Henzler P et\u00a0al (2024) CAT3D: Create anything in 3D with multi-view diffusion models. In: Globerson A, Mackey L, Belgrave D et\u00a0al (eds) Advances in Neural Information Processing Systems, vol\u00a037. Curran Associates, Inc., pp 75468\u201375494","DOI":"10.52202\/079017-2403"},{"key":"491_CR16","doi-asserted-by":"publisher","unstructured":"Gomez D, Gong B, Ovsjanikov M (2025) FourieRF: Few-shot NeRFs via progressive Fourier frequency control. https:\/\/doi.org\/10.48550\/arXiv.2502.01405. arXiv:2502.01405","DOI":"10.48550\/arXiv.2502.01405"},{"key":"491_CR17","unstructured":"Ho J, Jain A, Abbeel P (2020) Denoising diffusion probabilistic models. In: Larochelle H, Ranzato M, Hadsell R et\u00a0al (eds) Advances in Neural Information Processing Systems, vol\u00a033. Curran Associates, Inc., pp 6840\u20136851"},{"key":"491_CR18","doi-asserted-by":"publisher","unstructured":"Huo Y, Gang S, Guan C (2023) FCIHMRT: Feature cross-layer interaction hybrid method based on Res2Net and transformer for remote sensing scene classification. Electronics 12(4362). https:\/\/doi.org\/10.3390\/electronics12204362","DOI":"10.3390\/electronics12204362"},{"key":"491_CR19","doi-asserted-by":"publisher","unstructured":"Jain A, Tancik M, Abbeel P (2021) Putting NeRF on a diet: Semantically consistent few-shot view synthesis. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), pp 5865\u20135874. https:\/\/doi.org\/10.1109\/ICCV48922.2021.00583","DOI":"10.1109\/ICCV48922.2021.00583"},{"key":"491_CR20","doi-asserted-by":"publisher","unstructured":"Jang Y, P\u00e9rez-Pellitero E (2025) CoMapGS: Covisibility map-based Gaussian splatting for sparse novel view synthesis. In: 2025 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 26779\u201326788. https:\/\/doi.org\/10.1109\/CVPR52734.2025.02494","DOI":"10.1109\/CVPR52734.2025.02494"},{"key":"491_CR21","doi-asserted-by":"publisher","unstructured":"Jeong Y, Ahn S, Choy C et\u00a0al (2021) Self-calibrating neural radiance fields. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), pp 5826\u20135834. https:\/\/doi.org\/10.1109\/ICCV48922.2021.00579","DOI":"10.1109\/ICCV48922.2021.00579"},{"key":"491_CR22","doi-asserted-by":"publisher","unstructured":"Kajiya JT, Herzen BV (1984) Ray tracing volume densities. In: Christiansen H (ed) Proceedings of the 11th Annual Conference on Computer Graphics and Interactive Techniques, SIGGRAPH 1984, Minneapolis, Minnesota, USA, July 23-27, 1984. ACM, pp 165\u2013174. https:\/\/doi.org\/10.1145\/800031.808594","DOI":"10.1145\/800031.808594"},{"key":"491_CR23","doi-asserted-by":"publisher","unstructured":"Karras T, Laine S, Aittala M et\u00a0al (2020) Analyzing and improving the image quality of StyleGAN. In: 2020 IEEE\/CVF Conference on computer vision and pattern recognition (CVPR), pp 8107\u20138116. https:\/\/doi.org\/10.1109\/CVPR42600.2020.00813","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"491_CR24","doi-asserted-by":"publisher","unstructured":"Kerbl B, Kopanas G, Leimkuehler T et\u00a0al (2023) 3D Gaussian splatting for real-time radiance field rendering. ACM Trans Graph 42(4):139:1\u2013139:14. https:\/\/doi.org\/10.1145\/3592433","DOI":"10.1145\/3592433"},{"key":"491_CR25","doi-asserted-by":"publisher","unstructured":"Kim M, Seo S, Han B (2022) InfoNeRF: Ray entropy minimization for few-shot neural volume rendering. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 12902\u201312911. https:\/\/doi.org\/10.1109\/CVPR52688.2022.01257","DOI":"10.1109\/CVPR52688.2022.01257"},{"key":"491_CR26","unstructured":"Kim M, Kim JS, Yun SY et\u00a0al (2024) Synergistic integration of coordinate network and tensorial feature for improving neural radiance fields from sparse inputs. In: Forty-First International Conference on Machine Learning, ICML 2024, Vienna, Austria, July 21-27, 2024. OpenReview.net"},{"key":"491_CR27","doi-asserted-by":"publisher","unstructured":"Kondo N, Ikeda Y, Tagliasacchi A et\u00a0al (2021) VaxNeRF: Revisiting the classic for voxel-accelerated neural radiance field. https:\/\/doi.org\/10.48550\/arXiv.2111.13112. arXiv:2111.13112","DOI":"10.48550\/arXiv.2111.13112"},{"key":"491_CR28","doi-asserted-by":"publisher","first-page":"110863","DOI":"10.1016\/j.patcog.2024.110863","volume":"157","author":"S Lai","year":"2025","unstructured":"Lai S, Cui L, Yin J (2025) Fast radiance field reconstruction from sparse inputs. Pattern Recognit 157:110863. https:\/\/doi.org\/10.1016\/j.patcog.2024.110863","journal-title":"Pattern Recognit"},{"key":"491_CR29","doi-asserted-by":"crossref","unstructured":"Lao Y, Xu X, Cai Z et\u00a0al (2023) CorresNeRF: Image correspondence priors for neural radiance fields. In: Proceedings of the 37th International Conference on Neural Information Processing Systems. Curran Associates Inc., Red Hook, NY, USA, NIPS \u201923, pp 40504\u201340520","DOI":"10.52202\/075280-1762"},{"issue":"2","key":"491_CR30","doi-asserted-by":"publisher","first-page":"150","DOI":"10.1109\/34.273735","volume":"16","author":"A Laurentini","year":"1994","unstructured":"Laurentini A (1994) The visual hull concept for silhouette-based image understanding. IEEE Trans Pattern Anal Mach Intell 16(2):150\u2013162. https:\/\/doi.org\/10.1109\/34.273735","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"491_CR31","doi-asserted-by":"publisher","unstructured":"Li J, Zhang J, Bai X et\u00a0al (2024) DNGaussian: Optimizing sparse-view 3D Gaussian radiance fields with global-local depth normalization. In: 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 20775\u201320785. https:\/\/doi.org\/10.1109\/CVPR52733.2024.01963","DOI":"10.1109\/CVPR52733.2024.01963"},{"key":"491_CR32","doi-asserted-by":"publisher","unstructured":"Li L, Shen Z, Wang Z et\u00a0al (2023) Compressing volumetric radiance fields to 1 MB. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2023, Vancouver, BC, Canada, June 17-24, 2023. IEEE, pp 4222\u20134231. https:\/\/doi.org\/10.1109\/CVPR52729.2023.00411","DOI":"10.1109\/CVPR52729.2023.00411"},{"key":"491_CR33","doi-asserted-by":"crossref","unstructured":"Lin CY, Wu CH, Yeh CH et\u00a0al (2025) FrugalNeRF: Fast convergence for extreme few-shot novel view synthesis without learned priors. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2025, Nashville, TN, USA, June 11-15, 2025. Computer Vision Foundation \/ IEEE, pp 11227\u201311238","DOI":"10.1109\/CVPR52734.2025.01049"},{"key":"491_CR34","doi-asserted-by":"publisher","unstructured":"Liu X, Chen J, Kao SH et\u00a0al (2025) Deceptive-NeRF\/3DGS: Diffusion-generated pseudo-observations for\u00a0high-quality sparse-view reconstruction. In: Leonardis A, Ricci E, Roth S et\u00a0al (eds) Computer Vision \u2013 ECCV 2024. Springer Nature Switzerland, Cham, pp 337\u2013355. https:\/\/doi.org\/10.1007\/978-3-031-72640-8_19","DOI":"10.1007\/978-3-031-72640-8_19"},{"key":"491_CR35","doi-asserted-by":"publisher","unstructured":"Lombardi S, Simon T, Saragih J et\u00a0al (2019) Neural volumes: Learning dynamic renderable volumes from images. ACM Trans Graph 38(4):65:1\u201365:14. https:\/\/doi.org\/10.1145\/3306346.3323020","DOI":"10.1145\/3306346.3323020"},{"issue":"2","key":"491_CR36","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1109\/2945.468400","volume":"1","author":"N Max","year":"1995","unstructured":"Max N (1995) Optical models for direct volume rendering. IEEE Trans Vis Comput Graph 1(2):99\u2013108. https:\/\/doi.org\/10.1109\/2945.468400","journal-title":"IEEE Trans Vis Comput Graph"},{"key":"491_CR37","doi-asserted-by":"publisher","unstructured":"Miangoleh SMH, Dille S, Mai L et\u00a0al (2021) Boosting monocular depth estimation models to high-resolution via content-adaptive multi-resolution merging. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 9680\u20139689. https:\/\/doi.org\/10.1109\/CVPR46437.2021.00956","DOI":"10.1109\/CVPR46437.2021.00956"},{"key":"491_CR38","doi-asserted-by":"publisher","unstructured":"Mildenhall B, Srinivasan PP, Ortiz-Cayon R et\u00a0al (2019) Local light field fusion: Practical view synthesis with prescriptive sampling guidelines. ACM Trans Graph 38(4):29:1\u201329:14. https:\/\/doi.org\/10.1145\/3306346.3322980","DOI":"10.1145\/3306346.3322980"},{"key":"491_CR39","doi-asserted-by":"publisher","unstructured":"Mildenhall B, Srinivasan PP, Tancik M et\u00a0al (2020) NeRF: Representing scenes as neural radiance fields for view synthesis. In: Vedaldi A, Bischof H, Brox T et\u00a0al (eds) Computer Vision - ECCV 2020 - 16th European Conference, Glasgow, UK, August 23-28, 2020, Proceedings, Part I, Lecture Notes in Computer Science, vol 12346. Springer, pp 405\u2013421. https:\/\/doi.org\/10.1007\/978-3-030-58452-8_24","DOI":"10.1007\/978-3-030-58452-8_24"},{"key":"491_CR40","doi-asserted-by":"publisher","unstructured":"M\u00fcller T, Evans A, Schied C et\u00a0al (2022) Instant neural graphics primitives with a multiresolution hash encoding. ACM Trans Graph 41(4):102:1\u2013102:15. https:\/\/doi.org\/10.1145\/3528223.3530127","DOI":"10.1145\/3528223.3530127"},{"key":"491_CR41","doi-asserted-by":"publisher","unstructured":"Niemeyer M, Barron JT, Mildenhall B et\u00a0al (2022) RegNeRF: Regularizing neural radiance fields for view synthesis from sparse inputs. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 5470\u20135480. https:\/\/doi.org\/10.1109\/CVPR52688.2022.00540","DOI":"10.1109\/CVPR52688.2022.00540"},{"key":"491_CR42","doi-asserted-by":"publisher","unstructured":"Paliwal A, Ye W, Xiong J et\u00a0al (2025) CoherentGS: Sparse novel view synthesis with\u00a0coherent 3D Gaussians. In: Leonardis A, Ricci E, Roth S et\u00a0al (eds) Computer Vision \u2013 ECCV 2024. Springer Nature Switzerland, Cham, pp 19\u201337. https:\/\/doi.org\/10.1007\/978-3-031-73404-5_2","DOI":"10.1007\/978-3-031-73404-5_2"},{"key":"491_CR43","doi-asserted-by":"publisher","unstructured":"Park H, Ryu G, Kim W (2025) DropGaussian: Structural regularization for sparse-view Gaussian splatting. In: 2025 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 21600\u201321609. https:\/\/doi.org\/10.1109\/CVPR52734.2025.02012","DOI":"10.1109\/CVPR52734.2025.02012"},{"key":"491_CR44","doi-asserted-by":"publisher","unstructured":"Philip J, Deschaintre V (2023) Floaters no more: Radiance field gradient scaling for improved near-camera training. In: Ritschel T, Weidlich A (eds) Eurographics Symposium on Rendering. The Eurographics Association. https:\/\/doi.org\/10.2312\/sr.20231122","DOI":"10.2312\/sr.20231122"},{"key":"491_CR45","doi-asserted-by":"publisher","unstructured":"Ranftl R, Bochkovskiy A, Koltun V (2021) Vision Transformers for dense prediction. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), pp 12159\u201312168. https:\/\/doi.org\/10.1109\/ICCV48922.2021.01196","DOI":"10.1109\/ICCV48922.2021.01196"},{"key":"491_CR46","doi-asserted-by":"publisher","unstructured":"Rau A, Aklilu J, Christopher\u00a0Holsinger F et\u00a0al (2025) Depth-guided NeRF training via\u00a0earth mover\u2019s distance. In: Leonardis A, Ricci E, Roth S et\u00a0al (eds) Computer Vision \u2013 ECCV 2024. Springer Nature Switzerland, Cham, pp 1\u201317. https:\/\/doi.org\/10.1007\/978-3-031-73039-9_1","DOI":"10.1007\/978-3-031-73039-9_1"},{"key":"491_CR47","doi-asserted-by":"publisher","unstructured":"Rho D, Lee B, Nam S et\u00a0al (2023) Masked wavelet representation for compact neural radiance fields. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2023, Vancouver, BC, Canada, June 17-24, 2023. IEEE, pp 20680\u201320690. https:\/\/doi.org\/10.1109\/CVPR52729.2023.01981","DOI":"10.1109\/CVPR52729.2023.01981"},{"key":"491_CR48","doi-asserted-by":"publisher","unstructured":"Roessle B, Barron JT, Mildenhall B et\u00a0al (2022) Dense depth priors for neural radiance fields from sparse input views. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 12882\u201312891. https:\/\/doi.org\/10.1109\/CVPR52688.2022.01255","DOI":"10.1109\/CVPR52688.2022.01255"},{"key":"491_CR49","doi-asserted-by":"publisher","unstructured":"Sch\u00f6nberger JL, Frahm JM (2016) Structure-from-motion revisited. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp 4104\u20134113. https:\/\/doi.org\/10.1109\/CVPR.2016.445","DOI":"10.1109\/CVPR.2016.445"},{"key":"491_CR50","doi-asserted-by":"publisher","unstructured":"Seo S, Chang Y, Kwak N (2023a) FlipNeRF: Flipped reflection rays for few-shot novel view synthesis. In: 2023 IEEE\/CVF International Conference on Computer Vision (ICCV), pp 22826\u201322836. https:\/\/doi.org\/10.1109\/ICCV51070.2023.02092","DOI":"10.1109\/ICCV51070.2023.02092"},{"key":"491_CR51","doi-asserted-by":"publisher","unstructured":"Seo S, Han D, Chang Y et\u00a0al (2023b) MixNeRF: Modeling a ray with mixture density for novel view synthesis from sparse inputs. In: 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 20659\u201320668. https:\/\/doi.org\/10.1109\/CVPR52729.2023.01979","DOI":"10.1109\/CVPR52729.2023.01979"},{"key":"491_CR52","doi-asserted-by":"publisher","unstructured":"Shi R, Wei X, Wang C et\u00a0al (2024) ZeroRF: Fast sparse view 360$$^\\circ $$ reconstruction with zero pretraining. In: 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 21114\u201321124. https:\/\/doi.org\/10.1109\/CVPR52733.2024.01995","DOI":"10.1109\/CVPR52733.2024.01995"},{"key":"491_CR53","doi-asserted-by":"publisher","unstructured":"Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. https:\/\/doi.org\/10.48550\/arXiv.1409.1556. arXiv:1409.1556","DOI":"10.48550\/arXiv.1409.1556"},{"key":"491_CR54","doi-asserted-by":"publisher","unstructured":"Sitzmann V, Thies J, Heide F et\u00a0al (2019a) DeepVoxels: Learning persistent 3D feature embeddings. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 2432\u20132441. https:\/\/doi.org\/10.1109\/CVPR.2019.00254","DOI":"10.1109\/CVPR.2019.00254"},{"key":"491_CR55","volume-title":"Advances in Neural Information Processing Systems","author":"V Sitzmann","year":"2019","unstructured":"Sitzmann V, Zollhoefer M, Wetzstein G (2019) Scene representation networks: Continuous 3D-structure-aware neural scene representations. In: Wallach H, Larochelle H, Beygelzimer A et al (eds) Advances in Neural Information Processing Systems, vol 32. Curran Associates Inc"},{"key":"491_CR56","doi-asserted-by":"publisher","unstructured":"Somraj N, Soundararajan R (2023) ViP-NeRF: Visibility prior for sparse input neural radiance fields. In: ACM SIGGRAPH 2023 Conference Proceedings. Association for Computing Machinery, New York, NY, USA, SIGGRAPH \u201923, pp 1\u201311. https:\/\/doi.org\/10.1145\/3588432.3591539","DOI":"10.1145\/3588432.3591539"},{"key":"491_CR57","doi-asserted-by":"crossref","unstructured":"Song J, Park S, An H et\u00a0al (2023a) D\u00e4RF: Boosting radiance fields from sparse input views with monocular depth adaptation. In: Oh A, Naumann T, Globerson A et\u00a0al (eds) Advances in Neural Information Processing Systems, vol\u00a036. Curran Associates, Inc., pp 68458\u201368470","DOI":"10.52202\/075280-2993"},{"key":"491_CR58","doi-asserted-by":"publisher","unstructured":"Song L, Li Z, Gong X et\u00a0al (2023b) Harnessing low-frequency neural fields for few-shot view synthesis. https:\/\/doi.org\/10.48550\/arXiv.2303.08370. arXiv:2303.08370","DOI":"10.48550\/arXiv.2303.08370"},{"key":"491_CR59","doi-asserted-by":"publisher","unstructured":"Sun C, Sun M, Chen HT (2022) Direct voxel grid optimization: Super-fast convergence for radiance fields reconstruction. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 5449\u20135459. https:\/\/doi.org\/10.1109\/CVPR52688.2022.00538","DOI":"10.1109\/CVPR52688.2022.00538"},{"key":"491_CR60","doi-asserted-by":"publisher","unstructured":"Sun J, Zhang Z, Chen J et\u00a0al (2023) VGOS: Voxel grid optimization for view synthesis from sparse inputs. In: Thirty-Second International Joint Conference on Artificial Intelligence, pp 1414\u20131422. https:\/\/doi.org\/10.24963\/ijcai.2023\/157","DOI":"10.24963\/ijcai.2023\/157"},{"key":"491_CR61","unstructured":"Tieleman T, Hinton G (2012) RMSProp: Divide the gradient by a running average of its recent magnitude. COURSERA: Neural Netw Mach Learn 4(2):26"},{"key":"491_CR62","doi-asserted-by":"publisher","unstructured":"Uy MA, Martin-Brualla R, Guibas L et\u00a0al (2023) SCADE: NeRFs from space carving with ambiguity-aware depth estimates. In: 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 16518\u201316527. https:\/\/doi.org\/10.1109\/CVPR52729.2023.01585","DOI":"10.1109\/CVPR52729.2023.01585"},{"key":"491_CR63","doi-asserted-by":"publisher","unstructured":"Wang G, Chen Z, Loy CC et\u00a0al (2023) SparseNeRF: Distilling depth ranking for few-shot novel view synthesis. In: IEEE\/CVF International Conference on Computer Vision, ICCV 2023, Paris, France, October 1-6, 2023. IEEE, pp 9031\u20139042. https:\/\/doi.org\/10.1109\/ICCV51070.2023.00832","DOI":"10.1109\/ICCV51070.2023.00832"},{"key":"491_CR64","doi-asserted-by":"crossref","unstructured":"Wang J, Chen M, Karaev N et\u00a0al (2025) VGGT: Visual geometry grounded Transformer. In: Proceedings of the Computer Vision and Pattern Recognition Conference, pp 5294\u20135306","DOI":"10.1109\/CVPR52734.2025.00499"},{"key":"491_CR65","doi-asserted-by":"publisher","unstructured":"Wang S, Leroy V, Cabon Y et\u00a0al (2024) DUSt3R: Geometric 3D vision made easy. In: 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 20697\u201320709. https:\/\/doi.org\/10.1109\/CVPR52733.2024.01956","DOI":"10.1109\/CVPR52733.2024.01956"},{"issue":"4","key":"491_CR66","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang Z, Bovik A, Sheikh H et al (2004) Image quality assessment: From error visibility to structural similarity. IEEE Trans Image Process 13(4):600\u2013612. https:\/\/doi.org\/10.1109\/TIP.2003.819861","journal-title":"IEEE Trans Image Process"},{"key":"491_CR67","doi-asserted-by":"publisher","unstructured":"Wu R, Mildenhall B, Henzler P et\u00a0al (2024) ReconFusion: 3D reconstruction with diffusion priors. In: 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) pp 21551\u201321561. https:\/\/doi.org\/10.1109\/CVPR52733.2024.02036","DOI":"10.1109\/CVPR52733.2024.02036"},{"key":"491_CR68","doi-asserted-by":"publisher","unstructured":"Wynn J, Turmukhambetov D (2023) DiffusioNeRF: Regularizing neural radiance fields with denoising diffusion models. In: 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 4180\u20134189. https:\/\/doi.org\/10.1109\/CVPR52729.2023.00407","DOI":"10.1109\/CVPR52729.2023.00407"},{"key":"491_CR69","doi-asserted-by":"publisher","unstructured":"Xiao L, Nouri S, Hegland J et\u00a0al (2022) Neuralpassthrough: Learned real-time view synthesis for VR. In: ACM SIGGRAPH 2022 Conference Proceedings. Association for Computing Machinery, New York, NY, USA, SIGGRAPH \u201922. https:\/\/doi.org\/10.1145\/3528233.3530701","DOI":"10.1145\/3528233.3530701"},{"key":"491_CR70","doi-asserted-by":"publisher","unstructured":"Xiong H, Muttukuru S, Upadhyay R et\u00a0al (2025) SparseGS: Real-time 360$$^\\circ $$ sparse view synthesis using Gaussian splatting. https:\/\/doi.org\/10.48550\/arXiv.2312.00206. arXiv:2312.00206","DOI":"10.48550\/arXiv.2312.00206"},{"key":"491_CR71","doi-asserted-by":"publisher","unstructured":"Xu Y, Liu B, Tang H et\u00a0al (2024) Learning with unreliability: Fast few-shot voxel radiance fields with relative geometric consistency. In: 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 20342\u201320351. https:\/\/doi.org\/10.1109\/CVPR52733.2024.01923","DOI":"10.1109\/CVPR52733.2024.01923"},{"key":"491_CR72","doi-asserted-by":"publisher","unstructured":"Yang J, Pavone M, Wang Y (2023) FreeNeRF: Improving few-shot neural rendering with free frequency regularization. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2023, Vancouver, BC, Canada, June 17-24, 2023. IEEE, pp 8254\u20138263. https:\/\/doi.org\/10.1109\/CVPR52729.2023.00798","DOI":"10.1109\/CVPR52729.2023.00798"},{"key":"491_CR73","doi-asserted-by":"crossref","unstructured":"Yang L, Kang B, Huang Z et\u00a0al (2024a) Depth anything V2. In: Globerson A, Mackey L, Belgrave D et\u00a0al (eds) Advances in Neural Information Processing Systems, vol\u00a037. Curran Associates, Inc., pp 21875\u201321911","DOI":"10.52202\/079017-0688"},{"key":"491_CR74","doi-asserted-by":"publisher","first-page":"112096","DOI":"10.1016\/j.knosys.2024.112096","volume":"299","author":"W Yang","year":"2024","unstructured":"Yang W, Shi J, Bai S et al (2024) Clear-Plenoxels: Floaters free radiance fields without neural networks. Knowl-Based Syst 299:112096. https:\/\/doi.org\/10.1016\/j.knosys.2024.112096","journal-title":"Knowl-Based Syst"},{"key":"491_CR75","doi-asserted-by":"publisher","unstructured":"Yu A, Ye V, Tancik M et\u00a0al (2021) pixelNeRF: Neural radiance fields from one or few Images. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 4576\u20134585. https:\/\/doi.org\/10.1109\/CVPR46437.2021.00455","DOI":"10.1109\/CVPR46437.2021.00455"},{"key":"491_CR76","doi-asserted-by":"publisher","unstructured":"Zhang K, Riegler G, Snavely N et\u00a0al (2020) NeRF++: Analyzing and improving neural radiance fields. https:\/\/doi.org\/10.48550\/arXiv.2010.07492. arXiv:2010.07492","DOI":"10.48550\/arXiv.2010.07492"},{"key":"491_CR77","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TGRS.2024.3453415","volume":"62","author":"L Zhang","year":"2024","unstructured":"Zhang L, Fu S, Wang W et al (2024) Toward cross-domain class-incremental remote sensing scene classification. IEEE Trans Geosci Remote Sens 62:1\u201313. https:\/\/doi.org\/10.1109\/TGRS.2024.3453415","journal-title":"IEEE Trans Geosci Remote Sens"},{"key":"491_CR78","doi-asserted-by":"publisher","unstructured":"Zhang R, Isola P, Efros AA et\u00a0al (2018) The unreasonable effectiveness of deep features as a perceptual metric. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 586\u2013595. https:\/\/doi.org\/10.1109\/CVPR.2018.00068","DOI":"10.1109\/CVPR.2018.00068"},{"key":"491_CR79","doi-asserted-by":"crossref","unstructured":"Zhao T, Chen J, Leng C et al (2023) TinyNeRF: Towards 100 x compression of voxel radiance fields. Proceed AAAI Conf Artif Intell 37(3):3588\u20133596. https:\/\/doi.org\/10.1609\/aaai.v37i3.25469","DOI":"10.1609\/aaai.v37i3.25469"},{"key":"491_CR80","doi-asserted-by":"publisher","unstructured":"Zheng Y, Jiang Z, He S et\u00a0al (2025) NexusGS: Sparse view synthesis with epipolar depth priors in 3D Gaussian splatting. In: 2025 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 26800\u201326809. https:\/\/doi.org\/10.1109\/CVPR52734.2025.02496","DOI":"10.1109\/CVPR52734.2025.02496"},{"key":"491_CR81","doi-asserted-by":"publisher","unstructured":"Zhou Q, Ye Y, Liu Z et\u00a0al (2025) Non-local guided neural fields for 4D CT reconstruction. IEEE Trans Circuits Syst Video Technol pp 1\u20131. https:\/\/doi.org\/10.1109\/TCSVT.2025.3600922","DOI":"10.1109\/TCSVT.2025.3600922"},{"key":"491_CR82","doi-asserted-by":"crossref","unstructured":"Zhu H, He T, Li X et\u00a0al (2024) Is vanilla MLP in neural radiance field enough for few-shot view synthesis? In: 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 20288\u201320298. https:\/\/doi.org\/10.1109\/CVPR52733.2024.01918","DOI":"10.1109\/CVPR52733.2024.01918"},{"key":"491_CR83","doi-asserted-by":"publisher","unstructured":"Zhu Z, Fan Z, Jiang Y et\u00a0al (2025) FSGS: Real-time few-shot view synthesis using Gaussian splatting. In: Leonardis A, Ricci E, Roth S et\u00a0al (eds) Computer Vision \u2013 ECCV 2024. Springer Nature Switzerland, Cham, pp 145\u2013163. https:\/\/doi.org\/10.1007\/978-3-031-72933-1_9","DOI":"10.1007\/978-3-031-72933-1_9"}],"container-title":["Journal of King Saud University Computer and Information Sciences"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s44443-026-00491-2","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44443-026-00491-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44443-026-00491-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T08:41:20Z","timestamp":1775205680000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s44443-026-00491-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1,28]]},"references-count":83,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2026,4]]}},"alternative-id":["491"],"URL":"https:\/\/doi.org\/10.1007\/s44443-026-00491-2","relation":{},"ISSN":["1319-1578","2213-1248"],"issn-type":[{"value":"1319-1578","type":"print"},{"value":"2213-1248","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,1,28]]},"assertion":[{"value":"9 October 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 January 2026","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"28 January 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no relevant financial or non-financial interests to disclose.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing Interests"}}],"article-number":"98"}}