{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,11]],"date-time":"2025-09-11T20:49:04Z","timestamp":1757623744225,"version":"3.44.0"},"reference-count":53,"publisher":"Springer Science and Business Media LLC","issue":"9","license":[{"start":{"date-parts":[[2025,6,2]],"date-time":"2025-06-02T00:00:00Z","timestamp":1748822400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,6,2]],"date-time":"2025-06-02T00:00:00Z","timestamp":1748822400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"DOI":"10.13039\/501100002613","name":"Ulsan National Institute of Science and Technology","doi-asserted-by":"crossref","id":[{"id":"10.13039\/501100002613","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Comput Vis"],"published-print":{"date-parts":[[2025,9]]},"DOI":"10.1007\/s11263-025-02485-5","type":"journal-article","created":{"date-parts":[[2025,6,2]],"date-time":"2025-06-02T12:20:50Z","timestamp":1748866850000},"page":"6332-6346","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["DogRecon: Canine Prior-Guided Animatable 3D Gaussian Dog Reconstruction From A Single Image"],"prefix":"10.1007","volume":"133","author":[{"given":"Gyeongsu","family":"Cho","sequence":"first","affiliation":[]},{"given":"Changwoo","family":"Kang","sequence":"additional","affiliation":[]},{"given":"Donghyeon","family":"Soon","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3920-9608","authenticated-orcid":false,"given":"Kyungdon","family":"Joo","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,6,2]]},"reference":[{"key":"2485_CR1","doi-asserted-by":"crossref","unstructured":"AlBahar, B., Saito, S., Tseng, H.-Y., Kim, C., Kopf, J., & Huang, J.-B. (2023) Single-image 3d human digitization with shape-guided diffusion. In SIGGRAPH Asia,","DOI":"10.1145\/3610548.3618153"},{"issue":"4","key":"2485_CR2","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459850","volume":"40","author":"T Bagautdinov","year":"2021","unstructured":"Bagautdinov, T., Wu, C., Simon, T., Prada, F., Shiratori, T., Wei, S.-E., Xu, W., Sheikh, Y., & Saragih, J. (2021). Driving-signal aware full-body avatars. ACM Transactions on Graphics (TOG), 40(4), 1\u201317.","journal-title":"ACM Transactions on Graphics (TOG)"},{"key":"2485_CR3","doi-asserted-by":"crossref","unstructured":"Biggs, B., Boyne, O., Charles, J., Fitzgibbon, A., & Cipolla, R. (2020) Who left the dogs out? 3d animal reconstruction with expectation maximization in the loop. In ECCV,","DOI":"10.1007\/978-3-030-58621-8_12"},{"key":"2485_CR4","unstructured":"Couairon, G., Verbeek, J., Schwenk, H., & Cord, M. (2023) Diffedit: Diffusion-based semantic image editing with mask guidance. In ICLR,"},{"key":"2485_CR5","unstructured":"Dognomics2023. Internet-source dataset. https:\/\/www.dognomics.com\/breed-guide\/, n.d."},{"key":"2485_CR6","unstructured":"He, Z., & Wang, T. (2023) Openlrm: Open-source large reconstruction models. https:\/\/github.com\/3DTopia\/OpenLRM,"},{"key":"2485_CR7","unstructured":"Hong, Y., Zhang, K., Gu, J., Bi, S., Zhou, Y., Liu, D., Liu, F., Sunkavalli, K., Bui, T., & Tan, H. (2023) Lrm: Large reconstruction model for single image to 3d. arXiv preprint arXiv:2311.04400,"},{"key":"2485_CR8","doi-asserted-by":"crossref","unstructured":"Hu, L., Zhang, H., Zhang, Y., Zhou, B., Liu, B., Zhang, S., & Nie, L. (2023a) Gaussianavatar: Towards realistic human avatar modeling from a single video via animatable 3d gaussians. arXiv preprint arXiv:2312.02134,","DOI":"10.1109\/CVPR52733.2024.00067"},{"key":"2485_CR9","doi-asserted-by":"crossref","unstructured":"Hu, S., Hong, F., Pan, L., Mei, H., Yang, L., & Liu, Z. (2023b) Sherf: Generalizable human nerf from a single image. In ICCV,","DOI":"10.1109\/ICCV51070.2023.00858"},{"key":"2485_CR10","doi-asserted-by":"crossref","unstructured":"Huang, Y., Yi, H., Liu,W., Wang, H., Wu, B., Wang, W., Lin, B., Zhang, D., & Cai, D. (2023) One-shot implicit animatable avatars with model-based priors. In ICCV,","DOI":"10.1109\/ICCV51070.2023.00824"},{"key":"2485_CR11","doi-asserted-by":"crossref","unstructured":"Jain, A., Tancik, M., & Abbeel, P. (2021) Putting nerf on a diet: Semantically consistent few-shot view synthesis. In ICCV,","DOI":"10.1109\/ICCV48922.2021.00583"},{"key":"2485_CR12","doi-asserted-by":"crossref","unstructured":"Jain, A., Mildenhall, B., Barron, J.T., Abbeel, P., & Poole, B. (2022) Zero-shot text-guided object generation with dream fields. In CVPR,","DOI":"10.1109\/CVPR52688.2022.00094"},{"key":"2485_CR13","doi-asserted-by":"crossref","unstructured":"Jiang, W., Yi, K.\u00a0M., Samei, G., Tuzel, O., & Ranjan, A. (2022) Neuman: Neural human radiance field from a single video. In ECCV,","DOI":"10.1007\/978-3-031-19824-3_24"},{"key":"2485_CR14","unstructured":"Jun, H., & Nichol. A. (2023) Shap-e: Generating conditional 3d implicit functions. arXiv preprint arXiv:2305.02463,"},{"key":"2485_CR15","doi-asserted-by":"crossref","unstructured":"Kearney, S., Li, W., Parsons, M., Kim, K.I., & Cosker, D. (2020) Rgbd-dog: Predicting canine pose from rgbd sensors. In CVPR,","DOI":"10.1109\/CVPR42600.2020.00836"},{"key":"2485_CR16","doi-asserted-by":"crossref","unstructured":"Kerbl, B., Kopanas, G., Leimk\u00fchler, T., & Drettakis, G. (2023) 3d gaussian splatting for real-time radiance field rendering. ACM TOG, 42(4), July URL https:\/\/repo-sam.inria.fr\/fungraph\/3d-gaussian-splatting\/.","DOI":"10.1145\/3592433"},{"key":"2485_CR17","doi-asserted-by":"crossref","unstructured":"Kirillov, A., Mintun, E., Ravi, N., Mao, H., Rolland, C., Gustafson, L., Xiao, T., Whitehead, S., Berg, A.\u00a0C., Lo, & W.-Y., et\u00a0al. Segment anything. arXiv preprint arXiv:2304.02643, 2023.","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"2485_CR18","doi-asserted-by":"crossref","unstructured":"Kocabas, M., Chang, J.-H.\u00a0R., Gabriel, J., Tuzel, O., & Ranjan, A. (2023) Hugs: Human gaussian splats. arXiv preprint arXiv:2311.17910,","DOI":"10.1109\/CVPR52733.2024.00055"},{"key":"2485_CR19","first-page":"24741","volume":"34","author":"Y Kwon","year":"2021","unstructured":"Kwon, Y., Kim, D., Ceylan, D., & Fuchs, H. (2021). Neural human performer: Learning generalizable radiance fields for human performance rendering. Advances in Neural Information Processing Systems, 34, 24741\u201324752.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"2485_CR20","doi-asserted-by":"crossref","unstructured":"Lei, J., Wang, Y., Pavlakos, G., Liu, L., & Daniilidis, K. (2024) Gart: Gaussian articulated template models. In CVPR,","DOI":"10.1109\/CVPR52733.2024.01879"},{"key":"2485_CR21","doi-asserted-by":"crossref","unstructured":"Li, R., Tanke, J., Vo, M., Zollh\u00f6fer, M., Gall, J., Kanazawa, A., & Lassner, C. (2022) Tava: Template-free animatable volumetric actors. In ECCV,","DOI":"10.1007\/978-3-031-19824-3_25"},{"key":"2485_CR22","doi-asserted-by":"crossref","unstructured":"Li, Z., Litvak, D., Li, R., Zhang, Y., Jakab, T., Rupprecht, C., Wu, S., Vedaldi, A., & Wu, J. (2024) Learning the 3d fauna of the web. In CVPR,","DOI":"10.1109\/CVPR52733.2024.00931"},{"key":"2485_CR23","doi-asserted-by":"crossref","unstructured":"Lin, C.-H., Gao, J., Tang, L., Takikawa, T., Zeng, X., Huang, X., Kreis, K., Fidler, S., Liu, M.-Y., & Lin, T.-Y. (2023) Magic3d: High-resolution text-to-3d content creation. In CVPR,","DOI":"10.1109\/CVPR52729.2023.00037"},{"issue":"6","key":"2485_CR24","first-page":"1","volume":"40","author":"L Liu","year":"2021","unstructured":"Liu, L., Habermann, M., Rudnev, V., Sarkar, K., Gu, J., & Theobalt, C. (2021). Neural actor: Neural free-view synthesis of human actors with pose control. ACM transactions on graphics (TOG), 40(6), 1\u201316.","journal-title":"ACM transactions on graphics (TOG)"},{"key":"2485_CR25","unstructured":"Liu, M., Xu, C., Jin, H., Chen, L., Xu, Z., & Su, H., et\u00a0al. (2023a) One-2-3-45: Any single image to 3d mesh in 45 seconds without per-shape optimization. arXiv preprint arXiv:2306.16928,"},{"key":"2485_CR26","doi-asserted-by":"crossref","unstructured":"Liu, R., Wu, R., Van\u00a0Hoorick, B., Tokmakov, P., Zakharov, S., & Vondrick, C. (2023b) Zero-1-to-3: Zero-shot one image to 3d object. In ICCV,","DOI":"10.1109\/ICCV51070.2023.00853"},{"key":"2485_CR27","unstructured":"Liu, Y., Lin, C., Zeng, Z., Long, X., Liu, L., Komura, T., & Wang, W. (2023c) Syncdreamer: Generating multiview-consistent images from a single-view image. arXiv preprint arXiv:2309.03453,"},{"key":"2485_CR28","doi-asserted-by":"crossref","unstructured":"Luo, H., Xu, T., Jiang, Y., Zhou, C., Qiu, Q., Zhang, Y., Yang, W., Xu, L., & Yu, J. (2022) Artemis: Articulated neural pets with appearance and motion synthesis. ACM Trans. Graph., 41(4),","DOI":"10.1145\/3528223.3530086"},{"key":"2485_CR29","doi-asserted-by":"crossref","unstructured":"Mildenhall, B., Srinivasan, P.\u00a0P., Tancik, M., Barron, J.\u00a0T., Ramamoorthi, R., & Ng, R. (2020) Nerf: Representing scenes as neural radiance fields for view synthesis. In ECCV,","DOI":"10.1007\/978-3-030-58452-8_24"},{"key":"2485_CR30","unstructured":"Nichol, A., Jun, H., Dhariwal, P., Mishkin, P., & Chen, M. (2022) Point-e: A system for generating 3d point clouds from complex prompts. arXiv preprint arXiv:2212.08751,"},{"key":"2485_CR31","doi-asserted-by":"crossref","unstructured":"Peng, S., Dong, J., Wang, Q., Zhang, S., Shuai, Q., Zhou, X., & Bao, H. (2021a) Animatable neural radiance fields for modeling dynamic human bodies. In ICCV,","DOI":"10.1109\/ICCV48922.2021.01405"},{"key":"2485_CR32","doi-asserted-by":"crossref","unstructured":"Peng, S., Zhang, Y., Xu, Y., Wang, Q., Shuai, Q., Bao, H., & Zhou, X. (2021b) Neural body: Implicit neural representations with structured latent codes for novel view synthesis of dynamic humans. In CVPR,","DOI":"10.1109\/CVPR46437.2021.00894"},{"key":"2485_CR33","unstructured":"Poole, B., Jain, A., Barron, J.T., & Mildenhall, B. (2023) Dreamfusion: Text-to-3d using 2d diffusion. In ICLR,"},{"key":"2485_CR34","unstructured":"Radford, A., Kim, J.W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., Krueger, G., & Sutskever, I. (2021) Learning transferable visual models from natural language supervision. In ICML,"},{"key":"2485_CR35","unstructured":"Ramesh, A., Dhariwal, P., Nichol, A., Chu, C., & Chen, M. (2022) Hierarchical text-conditional image generation with clip latents,"},{"key":"2485_CR36","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., & Ommer, B. (2022) High-resolution image synthesis with latent diffusion models. In CVPR,","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"2485_CR37","doi-asserted-by":"crossref","unstructured":"R\u00fcegg, N., Tripathi, S., Schindler, K., Black, M.J., & Zuffi, S. (2023a) Bite: Beyond priors for improved three-d dog pose estimation. In CVPR,","DOI":"10.1109\/CVPR52729.2023.00856"},{"issue":"8","key":"2485_CR38","doi-asserted-by":"publisher","first-page":"1964","DOI":"10.1007\/s11263-023-01780-3","volume":"131","author":"N R\u00fcegg","year":"2023","unstructured":"R\u00fcegg, N., Zuffi, S., Schindler, K., & Black, M. J. (2023). Barc: Breed-augmented regression using classification for 3d dog reconstruction from images. IJCV, 131(8), 1964\u20131979.","journal-title":"IJCV"},{"key":"2485_CR39","doi-asserted-by":"crossref","unstructured":"Saito, S., Yang, J., Ma, Q., & Black, M.J. (2021) Scanimate: Weakly supervised learning of skinned clothed avatar networks. In CVPR,","DOI":"10.1109\/CVPR46437.2021.00291"},{"key":"2485_CR40","doi-asserted-by":"crossref","unstructured":"Sinha, S., Shapovalov, R. Reizenstein, J., Rocco, I., Neverova, N., Vedaldi, A., & Novotny, D. (2023) Common pets in 3d: Dynamic new-view synthesis of real-life deformable categories. In CVPR, 2023.","DOI":"10.1109\/CVPR52729.2023.00473"},{"key":"2485_CR41","unstructured":"Tang, J., Ren, J., Zhou, H., Liu, Z., & Zeng, G. (2023) Dreamgaussian: Generative gaussian splatting for efficient 3d content creation. arXiv preprint arXiv:2309.16653,"},{"key":"2485_CR42","doi-asserted-by":"crossref","unstructured":"Weng, C.-Y., Curless, B., Srinivasan, P.\u00a0P., Barron, J.\u00a0T., & Kemelmacher-Shlizerman, I. (2022) HumanNeRF: Free-viewpoint rendering of moving people from monocular video. In CVPR,","DOI":"10.1109\/CVPR52688.2022.01573"},{"key":"2485_CR43","doi-asserted-by":"crossref","unstructured":"Wu, S., Li, R., Jakab, T., Rupprecht, C., & Vedaldi, A. (2023) MagicPony: Learning articulated 3d animals in the wild. In CVPR,","DOI":"10.1109\/CVPR52729.2023.00849"},{"key":"2485_CR44","doi-asserted-by":"crossref","unstructured":"Xu, J., Wang, X., Cheng, W., Cao, Y.-P., Shan, Y., Qie, X., & Gao, S. (2023) Dream3d: Zero-shot text-to-3d synthesis using 3d shape prior and text-to-image diffusion models. In CVPR,","DOI":"10.1109\/CVPR52729.2023.02003"},{"key":"2485_CR45","doi-asserted-by":"crossref","unstructured":"Yang, G., Sun, D., Jampani, V., Vlasic, D., Cole, F., Chang, H., Ramanan, D., Freeman, W.T., & Liu, C. (2021) Lasr: Learning articulated shape reconstruction from a monocular video. In CVPR,","DOI":"10.1109\/CVPR46437.2021.01572"},{"key":"2485_CR46","doi-asserted-by":"crossref","unstructured":"Yang, G., Vo, M., Neverova, N., Ramanan, D., Vedaldi, A., & Joo, H. (2022) Banmo: Building animatable 3d neural models from many casual videos. In CVPR,","DOI":"10.1109\/CVPR52688.2022.00288"},{"key":"2485_CR47","doi-asserted-by":"crossref","unstructured":"Yang, G., Wang, C., Reddy, N.D., & Ramanan, D. (2023) Reconstructing animatable categories from videos. In CVPR,","DOI":"10.1109\/CVPR52729.2023.01630"},{"key":"2485_CR48","doi-asserted-by":"crossref","unstructured":"Yang, Z., Zhou, M., Shan, M., Wen, B., Xuan, Z., Hill, M., Bai, J., Qi, G.-J., & Wang, Y. (2024) Omnimotiongpt: Animal motion generation with limited data. In CVPR,","DOI":"10.1109\/CVPR52733.2024.00125"},{"key":"2485_CR49","doi-asserted-by":"crossref","unstructured":"Yao, C.-H., Hung, W.-C., Li, Y., Rubinstein, M., Yang, M.-H., & Jampani, V. (2023) Hi-lassie: High-fidelity articulated shape and skeleton discovery from sparse image ensemble. In CVPR,","DOI":"10.1109\/CVPR52729.2023.00470"},{"key":"2485_CR50","doi-asserted-by":"crossref","unstructured":"Ye, J., Wang, P., Li, K., Shi, Y., & Wang, H. (2023) Consistent-1-to-3: Consistent image to 3d view synthesis via geometry-aware diffusion models. 3DV,","DOI":"10.1109\/3DV62453.2024.00027"},{"key":"2485_CR51","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., & Wang, O. (2018) The unreasonable effectiveness of deep features as a perceptual metric. In CVPR,","DOI":"10.1109\/CVPR.2018.00068"},{"key":"2485_CR52","doi-asserted-by":"crossref","unstructured":"Zuffi, S., Kanazawa, A., Jacobs, D., & Black, M.J. (2017) 3D menagerie: Modeling the 3D shape and pose of animals. In CVPR,","DOI":"10.1109\/CVPR.2017.586"},{"key":"2485_CR53","doi-asserted-by":"crossref","unstructured":"Zuffi, S., Kanazawa, A., Berger-Wolf, T., & Black, M.J. (2018) Three-d safari: Learning to estimate zebra pose, shape, and texture from images \"in the wild\". In ICCV,","DOI":"10.1109\/ICCV.2019.00546"}],"container-title":["International Journal of Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-025-02485-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11263-025-02485-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-025-02485-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,9]],"date-time":"2025-09-09T08:06:49Z","timestamp":1757405209000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11263-025-02485-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,2]]},"references-count":53,"journal-issue":{"issue":"9","published-print":{"date-parts":[[2025,9]]}},"alternative-id":["2485"],"URL":"https:\/\/doi.org\/10.1007\/s11263-025-02485-5","relation":{},"ISSN":["0920-5691","1573-1405"],"issn-type":[{"type":"print","value":"0920-5691"},{"type":"electronic","value":"1573-1405"}],"subject":[],"published":{"date-parts":[[2025,6,2]]},"assertion":[{"value":"11 September 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 May 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 June 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}