{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,5]],"date-time":"2025-03-05T05:19:08Z","timestamp":1741151948802,"version":"3.38.0"},"reference-count":74,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/legalcode"}],"funder":[{"name":"Analytical Center under the RF Government","award":["000000D730321P5Q0002","70-2021-00145 02.11.2021"],"award-info":[{"award-number":["000000D730321P5Q0002","70-2021-00145 02.11.2021"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2025]]},"DOI":"10.1109\/access.2025.3536004","type":"journal-article","created":{"date-parts":[[2025,1,29]],"date-time":"2025-01-29T19:02:57Z","timestamp":1738177377000},"page":"36939-36950","source":"Crossref","is-referenced-by-count":0,"title":["NeuSD: Surface Completion With Multi-View Text-to-Image Diffusion"],"prefix":"10.1109","volume":"13","author":[{"given":"Savva","family":"Ignatyev","sequence":"first","affiliation":[{"name":"Skolkovo Institute of Science and Technology, Moscow, Russia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8789-5924","authenticated-orcid":false,"given":"Daniil","family":"Selikhanovych","sequence":"additional","affiliation":[{"name":"Skolkovo Institute of Science and Technology, Moscow, Russia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3666-9166","authenticated-orcid":false,"given":"Oleg","family":"Voynov","sequence":"additional","affiliation":[{"name":"Skolkovo Institute of Science and Technology, Moscow, Russia"}]},{"given":"Yiqun","family":"Wang","sequence":"additional","affiliation":[{"name":"College of Computer Science, Chongqing University, Chongqing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0627-9746","authenticated-orcid":false,"given":"Peter","family":"Wonka","sequence":"additional","affiliation":[{"name":"Division of Computer, Electrical and Mathematical Sciences and Engineering (CEMSE), King Abdullah University of Science and Technology, Thuwal, Saudi Arabia"}]},{"given":"Stamatios","family":"Lefkimmiatis","sequence":"additional","affiliation":[{"name":"MTS AI, Moscow, Russia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8424-0690","authenticated-orcid":false,"given":"Evgeny","family":"Burnaev","sequence":"additional","affiliation":[{"name":"Skolkovo Institute of Science and Technology, Moscow, Russia"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19824-3_13"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00329"},{"key":"ref3","first-page":"62332","article-title":"ReTR: Modeling rendering via transformer for generalizable neural surface reconstruction","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"36","author":"Liang"},{"key":"ref4","first-page":"1","article-title":"DreamFusion: Text-to-3D using 2D diffusion","volume-title":"Proc. 11th Int. Conf. Learn. Represent. (ICLR)","author":"Poole"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1145\/3503250"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.00946"},{"key":"ref7","first-page":"1","article-title":"Instant3D: Fast text-to-3D with sparse-view generation and large reconstruction model","volume-title":"Proc. 12th Int. Conf. Learn. Represent.","author":"Li"},{"key":"ref8","first-page":"1","article-title":"MVDream: Multi-view diffusion for 3D generation","volume-title":"Proc. 12th Int. Conf. Learn. Represent.","author":"Shi"},{"key":"ref9","first-page":"8406","article-title":"ProlificDreamer: High-fidelity and diverse text-to-3D generation with variational score distillation","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"36","author":"Wang"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.02033"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00037"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.00623"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00816"},{"key":"ref14","first-page":"22226","article-title":"One-2\u20133\u201345: Any single image to 3D mesh in 45 seconds without per-shape optimization","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"36","author":"Liu"},{"key":"ref15","first-page":"1","article-title":"Magic123: One image to high-quality 3D object generation using both 2D and 3D diffusion priors","volume-title":"Proc. 12th Int. Conf. Learn. Represent.","author":"Qian"},{"key":"ref16","first-page":"1","article-title":"DreamGaussian: Generative Gaussian splatting for efficient 3D content creation","volume-title":"Proc. 12th Int. Conf. Learn. Represent.","author":"Tang"},{"key":"ref17","first-page":"1","article-title":"Syncdreamer: Generating multiview-consistent images from a single-view image","volume-title":"Proc. The 12th Int. Conf. Learn. Represent.","author":"Liu"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.00951"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01211"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00223"},{"key":"ref21","first-page":"27171","article-title":"NeuS: Learning neural implicit surfaces by volume rendering for multi-view reconstruction","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Wang"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00186"},{"key":"ref24","first-page":"2492","article-title":"Multiview neural surface reconstruction by disentangling geometry and appearance","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Yariv"},{"key":"ref25","first-page":"4805","article-title":"Volume rendering of neural implicit surfaces","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Yariv"},{"key":"ref26","first-page":"3789","article-title":"Implicit geometric regularization for learning shapes","volume-title":"Proc. 37th Int. Conf. Mach. Learn.","author":"Gropp"},{"key":"ref27","first-page":"3403","article-title":"Geo-neus: Geometry-consistent neural implicit surfaces learning for multi-view reconstruction","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Fu"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00616"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19824-3_9"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1145\/3550454.3555514"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00305"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00818"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00817"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01997"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00455"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1145\/3528223.3530127"},{"issue":"4","key":"ref37","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3592433","article-title":"3D Gaussian splatting for real-time radiance field rendering","volume":"42","author":"Kerbl","year":"2023","journal-title":"ACM Trans. Graph."},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1111\/cgf.15063"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.02155"},{"key":"ref40","first-page":"1","article-title":"CAT3D: Create anything in 3D with multi-view diffusion models","volume-title":"Proc. 38th Annu. Conf.","author":"Gao"},{"key":"ref41","first-page":"1","article-title":"LEAP: Liberate sparse-view 3D modeling from camera poses","volume-title":"Proc. 12th Int. Conf. Learn. Represent.","author":"Jiang"},{"key":"ref42","first-page":"1","article-title":"LRM: Large reconstruction model for single image to 3D","volume-title":"Proc. 12th Int. Conf. Learn. Represent.","author":"Hong"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.00983"},{"key":"ref44","article-title":"Generic 3D diffusion adapter using controlled multi-view editing","author":"Chen","year":"2024","journal-title":"arXiv:2403.12032"},{"key":"ref45","first-page":"1","article-title":"An image is worth one word: Personalizing text-to-image generation using textual inversion","volume-title":"Proc. 11th Int. Conf. Learn. Represent.","author":"Gal"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00853"},{"key":"ref47","article-title":"Re-imagine the negative prompt algorithm: Transform 2D diffusion into 3D, alleviate Janus problem and beyond","author":"Armandpour","year":"2023","journal-title":"arXiv:2304.04968"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01263"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.00609"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW63382.2024.00753"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1145\/3641519.3657403"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1109\/3DV62453.2024.00154"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1109\/3DV62453.2024.00150"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1109\/TITS.2022.3195555"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58548-8_17"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00181"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00546"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01227"},{"key":"ref59","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00545"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00614"},{"key":"ref61","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00577"},{"key":"ref62","first-page":"75951","article-title":"DiffComplete: Diffusion-based generative 3D shape completion","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"36","author":"Chu"},{"key":"ref63","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00433"},{"key":"ref64","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-72655-2_1"},{"key":"ref65","article-title":"ShapeNet: An information-rich 3D model repository","author":"Chang","year":"2015","journal-title":"arXiv:1512.03012"},{"key":"ref66","first-page":"12171","article-title":"Point cloud completion with pretrained text-to-image diffusion models","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"36","author":"Kasten"},{"volume-title":"Stable Diffusion V2-1","year":"2022","author":"Rombach","key":"ref67"},{"key":"ref68","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.59"},{"key":"ref69","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-016-0902-9"},{"key":"ref70","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00883"},{"key":"ref71","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"Proc. 38th Int. Conf. Mach. Learn.","volume":"139","author":"Radford"},{"key":"ref72","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00068"},{"key":"ref73","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2003.819861"},{"key":"ref74","doi-asserted-by":"publisher","DOI":"10.1515\/eng-2019-0059"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/6287639\/10820123\/10857276.pdf?arnumber=10857276","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,4]],"date-time":"2025-03-04T05:53:58Z","timestamp":1741067638000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10857276\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"references-count":74,"URL":"https:\/\/doi.org\/10.1109\/access.2025.3536004","relation":{},"ISSN":["2169-3536"],"issn-type":[{"type":"electronic","value":"2169-3536"}],"subject":[],"published":{"date-parts":[[2025]]}}}