{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T12:16:18Z","timestamp":1773317778190,"version":"3.50.1"},"reference-count":72,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/100031931","name":"Shanghai Artificial Intelligence Laboratory","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100031931","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Displays"],"published-print":{"date-parts":[[2026,4]]},"DOI":"10.1016\/j.displa.2026.103349","type":"journal-article","created":{"date-parts":[[2026,1,10]],"date-time":"2026-01-10T16:00:34Z","timestamp":1768060834000},"page":"103349","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["Direct LiDAR-supervised surface-aligned 3D Gaussian Splatting for high-fidelity digital twin"],"prefix":"10.1016","volume":"92","author":[{"ORCID":"https:\/\/orcid.org\/0009-0003-9980-1392","authenticated-orcid":false,"given":"Xingdong","family":"Sheng","sequence":"first","affiliation":[]},{"given":"Qi","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Xu","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Zhenyang","family":"Qu","sequence":"additional","affiliation":[]},{"given":"Haoyu","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Shijie","family":"Mao","sequence":"additional","affiliation":[]},{"given":"Xiaokang","family":"Yang","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.displa.2026.103349_b1","series-title":"International Conference on Intelligent Computing","first-page":"499","article-title":"State-of-the-art techniques in 3D industrial reconstruction: A detailed survey","author":"Zhu","year":"2025"},{"key":"10.1016\/j.displa.2026.103349_b2","doi-asserted-by":"crossref","DOI":"10.1016\/j.displa.2025.102988","article-title":"High-precision 3D teeth reconstruction based on five-view intra-oral photos","volume":"87","author":"Wang","year":"2025","journal-title":"Displays"},{"key":"10.1016\/j.displa.2026.103349_b3","doi-asserted-by":"crossref","unstructured":"J.L. Schonberger, J.-M. Frahm, Structure-from-motion revisited, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 4104\u20134113.","DOI":"10.1109\/CVPR.2016.445"},{"key":"10.1016\/j.displa.2026.103349_b4","doi-asserted-by":"crossref","unstructured":"N. Snavely, S.M. Seitz, R. Szeliski, Photo tourism: exploring photo collections in 3D, in: ACM Siggraph 2006 Papers, 2006, pp. 835\u2013846.","DOI":"10.1145\/1141911.1141964"},{"key":"10.1016\/j.displa.2026.103349_b5","first-page":"519","article-title":"A comparison and evaluation of multi-view stereo reconstruction algorithms","volume":"vol. 1","author":"Seitz","year":"2006"},{"key":"10.1016\/j.displa.2026.103349_b6","doi-asserted-by":"crossref","unstructured":"J.L. Sch\u00f6nberger, E. Zheng, M. Pollefeys, J.-M. Frahm, Pixelwise View Selection for Unstructured Multi-View Stereo, in: European Conference on Computer Vision, ECCV, 2016.","DOI":"10.1007\/978-3-319-46487-9_31"},{"key":"10.1016\/j.displa.2026.103349_b7","doi-asserted-by":"crossref","DOI":"10.1016\/j.displa.2021.102100","article-title":"RGB-fusion: Monocular 3D reconstruction with learned depth prediction","volume":"70","author":"Duan","year":"2021","journal-title":"Displays"},{"issue":"1","key":"10.1016\/j.displa.2026.103349_b8","doi-asserted-by":"crossref","first-page":"99","DOI":"10.1145\/3503250","article-title":"Nerf: Representing scenes as neural radiance fields for view synthesis","volume":"65","author":"Mildenhall","year":"2021","journal-title":"Commun. ACM"},{"key":"10.1016\/j.displa.2026.103349_b9","doi-asserted-by":"crossref","unstructured":"S. Fridovich-Keil, G. Meanti, F.R. Warburg, B. Recht, A. Kanazawa, K-planes: Explicit radiance fields in space, time, and appearance, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 12479\u201312488.","DOI":"10.1109\/CVPR52729.2023.01201"},{"key":"10.1016\/j.displa.2026.103349_b10","doi-asserted-by":"crossref","unstructured":"S. Fridovich-Keil, A. Yu, M. Tancik, Q. Chen, B. Recht, A. Kanazawa, Plenoxels: Radiance fields without neural networks, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 5501\u20135510.","DOI":"10.1109\/CVPR52688.2022.00542"},{"key":"10.1016\/j.displa.2026.103349_b11","doi-asserted-by":"crossref","unstructured":"S.J. Garbin, M. Kowalski, M. Johnson, J. Shotton, J. Valentin, Fastnerf: High-fidelity neural rendering at 200fps, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2021, pp. 14346\u201314355.","DOI":"10.1109\/ICCV48922.2021.01408"},{"issue":"4","key":"10.1016\/j.displa.2026.103349_b12","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3592433","article-title":"3D gaussian splatting for real-time radiance field rendering","volume":"42","author":"Kerbl","year":"2023","journal-title":"ACM Trans. Graph."},{"key":"10.1016\/j.displa.2026.103349_b13","doi-asserted-by":"crossref","unstructured":"B. Huang, Z. Yu, A. Chen, A. Geiger, S. Gao, 2d gaussian splatting for geometrically accurate radiance fields, in: ACM SIGGRAPH 2024 Conference Papers, 2024, pp. 1\u201311.","DOI":"10.1145\/3641519.3657428"},{"key":"10.1016\/j.displa.2026.103349_b14","doi-asserted-by":"crossref","unstructured":"P. Dai, J. Xu, W. Xie, X. Liu, H. Wang, W. Xu, High-quality surface reconstruction using gaussian surfels, in: ACM SIGGRAPH 2024 Conference Papers, 2024, pp. 1\u201311.","DOI":"10.1145\/3641519.3657441"},{"key":"10.1016\/j.displa.2026.103349_b15","series-title":"2025 IEEE\/CVF Winter Conference on Applications of Computer Vision","first-page":"2421","article-title":"Dn-splatter: Depth and normal priors for gaussian splatting and meshing","author":"Turkulainen","year":"2025"},{"key":"10.1016\/j.displa.2026.103349_b16","series-title":"Autosplat: Constrained gaussian splatting for autonomous driving scene reconstruction","author":"Khan","year":"2024"},{"key":"10.1016\/j.displa.2026.103349_b17","doi-asserted-by":"crossref","unstructured":"H. Zhou, J. Shao, L. Xu, D. Bai, W. Qiu, B. Liu, Y. Wang, A. Geiger, Y. Liao, Hugs: Holistic urban 3d scene understanding via gaussian splatting, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2024, pp. 21336\u201321345.","DOI":"10.1109\/CVPR52733.2024.02016"},{"key":"10.1016\/j.displa.2026.103349_b18","doi-asserted-by":"crossref","DOI":"10.1109\/LRA.2024.3400149","article-title":"Liv-gaussmap: Lidar-inertial-visual fusion for real-time 3d radiance field map rendering","author":"Hong","year":"2024","journal-title":"IEEE Robot. Autom. Lett."},{"key":"10.1016\/j.displa.2026.103349_b19","doi-asserted-by":"crossref","unstructured":"X. Zhou, Z. Lin, X. Shan, Y. Wang, D. Sun, M.-H. Yang, Drivinggaussian: Composite gaussian splatting for surrounding dynamic autonomous driving scenes, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2024, pp. 21634\u201321643.","DOI":"10.1109\/CVPR52733.2024.02044"},{"issue":"6","key":"10.1016\/j.displa.2026.103349_b20","first-page":"1","article-title":"Letsgo: Large-scale garage modeling and rendering via lidar-assisted gaussian primitives","volume":"43","author":"Cui","year":"2024","journal-title":"ACM Trans. Graph."},{"key":"10.1016\/j.displa.2026.103349_b21","series-title":"S3 Gaussian: Self-supervised street Gaussians for autonomous driving","author":"Huang","year":"2024"},{"key":"10.1016\/j.displa.2026.103349_b22","series-title":"Street Gaussians for modeling dynamic urban scenes.(2023)","author":"Yan","year":"2023"},{"key":"10.1016\/j.displa.2026.103349_b23","doi-asserted-by":"crossref","unstructured":"J. Kulhanek, T. Sattler, Tetra-nerf: Representing neural radiance fields using tetrahedra, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2023, pp. 18458\u201318469.","DOI":"10.1109\/ICCV51070.2023.01692"},{"key":"10.1016\/j.displa.2026.103349_b24","doi-asserted-by":"crossref","unstructured":"K. Rematas, A. Liu, P.P. Srinivasan, J.T. Barron, A. Tagliasacchi, T. Funkhouser, V. Ferrari, Urban radiance fields, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 12932\u201312942.","DOI":"10.1109\/CVPR52688.2022.01259"},{"key":"10.1016\/j.displa.2026.103349_b25","doi-asserted-by":"crossref","unstructured":"A. Chen, Z. Xu, F. Zhao, X. Zhang, F. Xiang, J. Yu, H. Su, Mvsnerf: Fast generalizable radiance field reconstruction from multi-view stereo, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2021, pp. 14124\u201314133.","DOI":"10.1109\/ICCV48922.2021.01386"},{"key":"10.1016\/j.displa.2026.103349_b26","series-title":"Neus: Learning neural implicit surfaces by volume rendering for multi-view reconstruction","author":"Wang","year":"2021"},{"key":"10.1016\/j.displa.2026.103349_b27","doi-asserted-by":"crossref","DOI":"10.1016\/j.displa.2025.103000","article-title":"ASR-nesurf: Alleviating structural redundancy in neural surface reconstruction for deformable endoscopic tissues by validity probability","volume":"88","author":"Zhang","year":"2025","journal-title":"Displays"},{"key":"10.1016\/j.displa.2026.103349_b28","doi-asserted-by":"crossref","DOI":"10.1016\/j.displa.2025.102996","article-title":"DARF: Depth-aware generalizable neural radiance field","volume":"88","author":"Shi","year":"2025","journal-title":"Displays"},{"key":"10.1016\/j.displa.2026.103349_b29","doi-asserted-by":"crossref","unstructured":"J.T. Barron, B. Mildenhall, M. Tancik, P. Hedman, R. Martin-Brualla, P.P. Srinivasan, Mip-nerf: A multiscale representation for anti-aliasing neural radiance fields, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2021, pp. 5855\u20135864.","DOI":"10.1109\/ICCV48922.2021.00580"},{"key":"10.1016\/j.displa.2026.103349_b30","doi-asserted-by":"crossref","unstructured":"J.T. Barron, B. Mildenhall, D. Verbin, P.P. Srinivasan, P. Hedman, Mip-nerf 360: Unbounded anti-aliased neural radiance fields, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 5470\u20135479.","DOI":"10.1109\/CVPR52688.2022.00539"},{"issue":"4","key":"10.1016\/j.displa.2026.103349_b31","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3528223.3530127","article-title":"Instant neural graphics primitives with a multiresolution hash encoding","volume":"41","author":"M\u00fcller","year":"2022","journal-title":"ACM Trans. Graph."},{"key":"10.1016\/j.displa.2026.103349_b32","doi-asserted-by":"crossref","unstructured":"C. Sun, M. Sun, H.-T. Chen, Direct voxel grid optimization: Super-fast convergence for radiance fields reconstruction, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 5459\u20135469.","DOI":"10.1109\/CVPR52688.2022.00538"},{"key":"10.1016\/j.displa.2026.103349_b33","doi-asserted-by":"crossref","unstructured":"A. Yu, R. Li, M. Tancik, H. Li, R. Ng, A. Kanazawa, Plenoctrees for real-time rendering of neural radiance fields, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2021, pp. 5752\u20135761.","DOI":"10.1109\/ICCV48922.2021.00570"},{"key":"10.1016\/j.displa.2026.103349_b34","doi-asserted-by":"crossref","unstructured":"Z. Chen, T. Funkhouser, P. Hedman, A. Tagliasacchi, Mobilenerf: Exploiting the polygon rasterization pipeline for efficient neural field rendering on mobile architectures, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 16569\u201316578.","DOI":"10.1109\/CVPR52729.2023.01590"},{"key":"10.1016\/j.displa.2026.103349_b35","series-title":"European Conference on Computer Vision","first-page":"333","article-title":"Tensorf: Tensorial radiance fields","author":"Chen","year":"2022"},{"key":"10.1016\/j.displa.2026.103349_b36","doi-asserted-by":"crossref","unstructured":"Q. Xu, Z. Xu, J. Philip, S. Bi, Z. Shu, K. Sunkavalli, U. Neumann, Point-nerf: Point-based neural radiance fields, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 5438\u20135448.","DOI":"10.1109\/CVPR52688.2022.00536"},{"key":"10.1016\/j.displa.2026.103349_b37","doi-asserted-by":"crossref","unstructured":"M. Tancik, V. Casser, X. Yan, S. Pradhan, B. Mildenhall, P.P. Srinivasan, J.T. Barron, H. Kretzschmar, Block-nerf: Scalable large scene neural view synthesis, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 8248\u20138258.","DOI":"10.1109\/CVPR52688.2022.00807"},{"key":"10.1016\/j.displa.2026.103349_b38","doi-asserted-by":"crossref","unstructured":"H. Turki, D. Ramanan, M. Satyanarayanan, Mega-nerf: Scalable construction of large-scale nerfs for virtual fly-throughs, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 12922\u201312931.","DOI":"10.1109\/CVPR52688.2022.01258"},{"key":"10.1016\/j.displa.2026.103349_b39","doi-asserted-by":"crossref","unstructured":"C. Li, B.Y. Feng, Z. Fan, P. Pan, Z. Wang, Steganerf: Embedding invisible information within neural radiance fields, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2023, pp. 441\u2013453.","DOI":"10.1109\/ICCV51070.2023.00047"},{"key":"10.1016\/j.displa.2026.103349_b40","doi-asserted-by":"crossref","DOI":"10.1109\/TCSVT.2025.3538684","article-title":"3D gaussian splatting: Survey, technologies, challenges, and opportunities","author":"Bao","year":"2025","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.displa.2026.103349_b41","article-title":"Image-based view-dependent appearance for 3D gaussian splatting","author":"Guo","year":"2025","journal-title":"Displays"},{"key":"10.1016\/j.displa.2026.103349_b42","doi-asserted-by":"crossref","unstructured":"Y. Liu, K. Fan, W. Yu, C. Li, H. Lu, Y. Yuan, MonoSplat: Generalizable 3D Gaussian Splatting from Monocular Depth Foundation Models, in: Proceedings of the Computer Vision and Pattern Recognition Conference, 2025, pp. 21570\u201321579.","DOI":"10.1109\/CVPR52734.2025.02009"},{"key":"10.1016\/j.displa.2026.103349_b43","doi-asserted-by":"crossref","unstructured":"H. Liu, Y. Wang, C. Li, R. Cai, K. Wang, W. Li, P. Molchanov, P. Wang, Z. Wang, FlexGS: Train Once, Deploy Everywhere with Many-in-One Flexible 3D Gaussian Splatting, in: Proceedings of the Computer Vision and Pattern Recognition Conference, 2025, pp. 16336\u201316345.","DOI":"10.1109\/CVPR52734.2025.01523"},{"key":"10.1016\/j.displa.2026.103349_b44","unstructured":"C. Li, H. Liu, Z. Fan, W. Li, Y. Liu, P. Pan, Y. Yuan, Instantsplamp: Fast and generalizable stenography framework for generative gaussian splatting, in: The Thirteenth International Conference on Learning Representations, 2025."},{"key":"10.1016\/j.displa.2026.103349_b45","series-title":"ICASSP 2025-2025 IEEE International Conference on Acoustics, Speech and Signal Processing","first-page":"1","article-title":"Concealgs: Concealing invisible copyright information in 3D Gaussian splatting","author":"Yang","year":"2025"},{"key":"10.1016\/j.displa.2026.103349_b46","series-title":"2025 IEEE International Conference on Robotics and Automation","first-page":"2694","article-title":"Hide-in-motion: Embedding steganographic copyright information into 4d Gaussian splatting assets","author":"Liu","year":"2025"},{"key":"10.1016\/j.displa.2026.103349_b47","doi-asserted-by":"crossref","first-page":"140138","DOI":"10.52202\/079017-4447","article-title":"Lightgaussian: Unbounded 3d gaussian compression with 15x reduction and 200+ fps","volume":"37","author":"Fan","year":"2024","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.displa.2026.103349_b48","doi-asserted-by":"crossref","unstructured":"J.C. Lee, D. Rho, X. Sun, J.H. Ko, E. Park, Compact 3d gaussian representation for radiance field, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2024, pp. 21719\u201321728.","DOI":"10.1109\/CVPR52733.2024.02052"},{"key":"10.1016\/j.displa.2026.103349_b49","doi-asserted-by":"crossref","unstructured":"Z. Yu, A. Chen, B. Huang, T. Sattler, A. Geiger, Mip-splatting: Alias-free 3d gaussian splatting, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2024, pp. 19447\u201319456.","DOI":"10.1109\/CVPR52733.2024.01839"},{"key":"10.1016\/j.displa.2026.103349_b50","doi-asserted-by":"crossref","unstructured":"Z. Ye, W. Li, S. Liu, P. Qiao, Y. Dou, Absgs: Recovering fine details in 3d gaussian splatting, in: Proceedings of the 32nd ACM International Conference on Multimedia, 2024, pp. 1053\u20131061.","DOI":"10.1145\/3664647.3681361"},{"key":"10.1016\/j.displa.2026.103349_b51","doi-asserted-by":"crossref","unstructured":"A. Hamdi, L. Melas-Kyriazi, J. Mai, G. Qian, R. Liu, C. Vondrick, B. Ghanem, A. Vedaldi, Ges: Generalized exponential splatting for efficient radiance field rendering, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2024, pp. 19812\u201319822.","DOI":"10.1109\/CVPR52733.2024.01873"},{"key":"10.1016\/j.displa.2026.103349_b52","doi-asserted-by":"crossref","unstructured":"J. Lin, Z. Li, X. Tang, J. Liu, S. Liu, J. Liu, Y. Lu, X. Wu, S. Xu, Y. Yan, et al., Vastgaussian: Vast 3d gaussians for large scene reconstruction, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2024, pp. 5166\u20135175.","DOI":"10.1109\/CVPR52733.2024.00494"},{"key":"10.1016\/j.displa.2026.103349_b53","series-title":"European Conference on Computer Vision","first-page":"265","article-title":"Citygaussian: Real-time high-quality large-scale scene rendering with gaussians","author":"Liu","year":"2024"},{"key":"10.1016\/j.displa.2026.103349_b54","doi-asserted-by":"crossref","unstructured":"G. Wu, T. Yi, J. Fang, L. Xie, X. Zhang, W. Wei, W. Liu, Q. Tian, X. Wang, 4d gaussian splatting for real-time dynamic scene rendering, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2024, pp. 20310\u201320320.","DOI":"10.1109\/CVPR52733.2024.01920"},{"key":"10.1016\/j.displa.2026.103349_b55","series-title":"ACM SIGGRAPH 2024 Conference Papers","article-title":"High-quality surface reconstruction using Gaussian surfels","author":"Dai","year":"2024"},{"key":"10.1016\/j.displa.2026.103349_b56","doi-asserted-by":"crossref","unstructured":"K. Jiang, V. Sivaram, C. Peng, R. Ramamoorthi, Geometry Field Splatting with Gaussian Surfels, in: Proceedings of the Computer Vision and Pattern Recognition Conference, 2025, pp. 5752\u20135762.","DOI":"10.1109\/CVPR52734.2025.00540"},{"key":"10.1016\/j.displa.2026.103349_b57","doi-asserted-by":"crossref","unstructured":"P.-C. Kung, S. Isaacson, R. Vasudevan, K.A. Skinner, SAD-GS: Shape-aligned Depth-supervised Gaussian Splatting, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2024, pp. 2842\u20132851.","DOI":"10.1109\/CVPRW63382.2024.00290"},{"key":"10.1016\/j.displa.2026.103349_b58","doi-asserted-by":"crossref","unstructured":"L. Yang, B. Kang, Z. Huang, X. Xu, J. Feng, H. Zhao, Depth anything: Unleashing the power of large-scale unlabeled data, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2024, pp. 10371\u201310381.","DOI":"10.1109\/CVPR52733.2024.00987"},{"key":"10.1016\/j.displa.2026.103349_b59","doi-asserted-by":"crossref","unstructured":"H. Xu, S. Peng, F. Wang, H. Blum, D. Barath, A. Geiger, M. Pollefeys, Depthsplat: Connecting gaussian splatting and depth, in: Proceedings of the Computer Vision and Pattern Recognition Conference, 2025, pp. 16453\u201316463.","DOI":"10.1109\/CVPR52734.2025.01534"},{"key":"10.1016\/j.displa.2026.103349_b60","doi-asserted-by":"crossref","unstructured":"Z. Peng, T. Shao, Y. Liu, J. Zhou, Y. Yang, J. Wang, K. Zhou, Rtg-slam: Real-time 3d reconstruction at scale using gaussian splatting, in: ACM SIGGRAPH 2024 Conference Papers, 2024, pp. 1\u201311.","DOI":"10.1145\/3641519.3657455"},{"key":"10.1016\/j.displa.2026.103349_b61","article-title":"Liv-GS: Lidar-vision integration for 3D Gaussian splatting SLAM in outdoor environments","author":"Xiao","year":"2024","journal-title":"IEEE Robot. Autom. Lett."},{"key":"10.1016\/j.displa.2026.103349_b62","series-title":"2025 International Conference on Robotics and Automation","article-title":"Gaussian-LIC: Real-time photo-realistic SLAM with Gaussian splatting and lidar-inertial-camera fusion","author":"Lang","year":"2025"},{"key":"10.1016\/j.displa.2026.103349_b63","doi-asserted-by":"crossref","unstructured":"G. Hess, C. Lindstr\u00f6m, M. Fatemi, C. Petersson, L. Svensson, Splatad: Real-time lidar and camera rendering with 3d gaussian splatting for autonomous driving, in: Proceedings of the Computer Vision and Pattern Recognition Conference, 2025, pp. 11982\u201311992.","DOI":"10.1109\/CVPR52734.2025.01119"},{"key":"10.1016\/j.displa.2026.103349_b64","article-title":"Li-gs: Gaussian splatting with lidar incorporated for accurate large-scale reconstruction","author":"Jiang","year":"2024","journal-title":"IEEE Robot. Autom. Lett."},{"key":"10.1016\/j.displa.2026.103349_b65","series-title":"Lihi-gs: Lidar-supervised gaussian splatting for highway driving scene reconstruction","author":"Kung","year":"2024"},{"key":"10.1016\/j.displa.2026.103349_b66","series-title":"2024 IEEE International Conference on Robotics and Automation","first-page":"1723","article-title":"Colmap-pcd: An open-source tool for fine image-to-point cloud registration","author":"Bai","year":"2024"},{"key":"10.1016\/j.displa.2026.103349_b67","doi-asserted-by":"crossref","first-page":"18","DOI":"10.1016\/j.isprsjprs.2020.02.018","article-title":"Robust normal vector estimation in 3D point clouds through iterative principal component analysis","volume":"163","author":"Sanchez","year":"2020","journal-title":"ISPRS J. Photogramm. Remote Sens."},{"key":"10.1016\/j.displa.2026.103349_b68","series-title":"Neusg: Neural implicit surface reconstruction with 3d gaussian splatting guidance","author":"Chen","year":"2023"},{"key":"10.1016\/j.displa.2026.103349_b69","doi-asserted-by":"crossref","unstructured":"F. Warburg, E. Weber, M. Tancik, A. Holynski, A. Kanazawa, Nerfbusters: Removing ghostly artifacts from casually captured nerfs, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2023, pp. 18120\u201318130.","DOI":"10.1109\/ICCV51070.2023.01661"},{"key":"10.1016\/j.displa.2026.103349_b70","first-page":"21875","article-title":"Depth anything v2","volume":"37","author":"Yang","year":"2024","journal-title":"Adv. Neural Inf. Process. Syst."},{"issue":"4","key":"10.1016\/j.displa.2026.103349_b71","doi-asserted-by":"crossref","first-page":"600","DOI":"10.1109\/TIP.2003.819861","article-title":"Image quality assessment: from error visibility to structural similarity","volume":"13","author":"Wang","year":"2004","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.displa.2026.103349_b72","doi-asserted-by":"crossref","unstructured":"R. Zhang, P. Isola, A.A. Efros, E. Shechtman, O. Wang, The unreasonable effectiveness of deep features as a perceptual metric, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 586\u2013595.","DOI":"10.1109\/CVPR.2018.00068"}],"container-title":["Displays"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0141938226000120?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0141938226000120?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,3,11]],"date-time":"2026-03-11T18:24:39Z","timestamp":1773253479000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0141938226000120"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,4]]},"references-count":72,"alternative-id":["S0141938226000120"],"URL":"https:\/\/doi.org\/10.1016\/j.displa.2026.103349","relation":{},"ISSN":["0141-9382"],"issn-type":[{"value":"0141-9382","type":"print"}],"subject":[],"published":{"date-parts":[[2026,4]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Direct LiDAR-supervised surface-aligned 3D Gaussian Splatting for high-fidelity digital twin","name":"articletitle","label":"Article Title"},{"value":"Displays","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.displa.2026.103349","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 Elsevier B.V. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"103349"}}