{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,24]],"date-time":"2025-06-24T07:10:31Z","timestamp":1750749031000,"version":"3.40.3"},"publisher-location":"Cham","reference-count":33,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031479656"},{"type":"electronic","value":"9783031479663"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-47966-3_10","type":"book-chapter","created":{"date-parts":[[2023,12,2]],"date-time":"2023-12-02T09:02:11Z","timestamp":1701507731000},"page":"121-135","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Inpainting of\u00a0Depth Images Using Deep Neural Networks for\u00a0Real-Time Applications"],"prefix":"10.1007","author":[{"given":"Roland","family":"Fischer","sequence":"first","affiliation":[]},{"given":"Janis","family":"Ro\u00dfkamp","sequence":"additional","affiliation":[]},{"given":"Thomas","family":"Hudcovic","sequence":"additional","affiliation":[]},{"given":"Anton","family":"Schlegel","sequence":"additional","affiliation":[]},{"given":"Gabriel","family":"Zachmann","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,3]]},"reference":[{"key":"10_CR1","doi-asserted-by":"crossref","unstructured":"Armanious, K., Mecky, Y., Gatidis, S., Yang, B.: Adversarial inpainting of medical image modalities. In: IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 3267\u20133271 (2019)","DOI":"10.1109\/ICASSP.2019.8682677"},{"key":"10_CR2","doi-asserted-by":"crossref","unstructured":"Chen, H., et al.: Pre-trained image processing transformer. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12299\u201312310 (2021)","DOI":"10.1109\/CVPR46437.2021.01212"},{"key":"10_CR3","doi-asserted-by":"crossref","unstructured":"Deng, Y., Hui, S., Zhou, S., Meng, D., Wang, J.: T-former: an efficient transformer for image inpainting. In: Proceedings of the 30th ACM International Conference on Multimedia, MM 2022, pp. 6559\u20136568. Association for Computing Machinery (2022)","DOI":"10.1145\/3503161.3548446"},{"key":"10_CR4","doi-asserted-by":"crossref","unstructured":"Fujii, R., Hachiuma, R., Saito, H.: RGB-D image inpainting using generative adversarial network with a late fusion approach. ArXiv: abs\/2110.07413 (2020)","DOI":"10.1109\/ISMAR-Adjunct.2019.00-46"},{"key":"10_CR5","doi-asserted-by":"crossref","unstructured":"Isola, P., Zhu, J.Y., Zhou, T., Efros, A.: Image-to-image translation with conditional adversarial networks. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5967\u20135976 (2017)","DOI":"10.1109\/CVPR.2017.632"},{"issue":"3","key":"10_CR6","doi-asserted-by":"publisher","first-page":"6391","DOI":"10.1109\/LRA.2022.3171727","volume":"7","author":"J Jeon","year":"2022","unstructured":"Jeon, J., Lim, H., Seo, D.U., Myung, H.: Struct-mdc: mesh-refined unsupervised depth completion leveraging structural regularities from visual slam. IEEE Robot. Autom. Lett. 7(3), 6391\u20136398 (2022)","journal-title":"IEEE Robot. Autom. Lett."},{"issue":"6","key":"10_CR7","doi-asserted-by":"publisher","first-page":"1797","DOI":"10.3390\/s20061797","volume":"20","author":"W Jin","year":"2020","unstructured":"Jin, W., Zun, L., Yong, L.: Double-constraint inpainting model of a single-depth image. Sensors 20(6), 1797 (2020)","journal-title":"Sensors"},{"key":"10_CR8","doi-asserted-by":"crossref","unstructured":"Lee, S., Yi, E., Lee, J., Kim, J.: Multi-scaled and densely connected locally convolutional layers for depth completion. In: 2022 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 8360\u20138367 (2022)","DOI":"10.1109\/IROS47612.2022.9982179"},{"key":"10_CR9","doi-asserted-by":"crossref","unstructured":"Li, W., Lin, Z., Kun, Z., Qi, L., Wang, Y., Jia, J.: Mat: mask-aware transformer for large hole image inpainting. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10748\u201310758 (2022)","DOI":"10.1109\/CVPR52688.2022.01049"},{"issue":"18","key":"10_CR10","doi-asserted-by":"publisher","first-page":"3941","DOI":"10.3390\/s19183941","volume":"19","author":"Z Li","year":"2019","unstructured":"Li, Z., et al.: Promising generative adversarial network based sinogram inpainting method for ultra-limited-angle computed tomography imaging. Sensors 19(18), 3941 (2019)","journal-title":"Sensors"},{"issue":"6","key":"10_CR11","doi-asserted-by":"publisher","first-page":"1103","DOI":"10.3390\/app9061103","volume":"9","author":"Z Li","year":"2019","unstructured":"Li, Z., Wu, J.: Learning deep CNN denoiser priors for depth image inpainting. Appl. Sci. 9(6), 1103 (2019)","journal-title":"Appl. Sci."},{"key":"10_CR12","unstructured":"Liu, G.: Pytorch implementation of the partial convolution layer for padding and image inpainting (2018). https:\/\/github.com\/NVIDIA\/partialconv"},{"key":"10_CR13","doi-asserted-by":"crossref","unstructured":"Liu, G., Reda, F.A., Shih, K.J., Wang, T.C., Tao, A., Catanzaro, B.: Image inpainting for irregular holes using partial convolutions. In: European Conference on Computer Vision (2018)","DOI":"10.1007\/978-3-030-01252-6_6"},{"key":"10_CR14","doi-asserted-by":"crossref","unstructured":"Makarov, I., Borisenko, G.: Depth inpainting via vision transformer. In: 2021 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct), pp. 286\u2013291 (10 2021)","DOI":"10.1109\/ISMAR-Adjunct54149.2021.00065"},{"key":"10_CR15","doi-asserted-by":"crossref","unstructured":"Mal, F., Karaman, S.: Sparse-to-dense: depth prediction from sparse depth samples and a single image. In: 2018 IEEE International Conference on Robotics and Automation (ICRA), pp. 1\u20138 (2018)","DOI":"10.1109\/ICRA.2018.8460184"},{"key":"10_CR16","doi-asserted-by":"crossref","unstructured":"Ning, W., Li, J., Zhang, L., Du, B.: Musical: multi-scale image contextual attention learning for inpainting. In: Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligent, pp. 3748\u20133754 (2019)","DOI":"10.24963\/ijcai.2019\/520"},{"key":"10_CR17","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10684\u201310695 (2022)","DOI":"10.1109\/CVPR52688.2022.01042"},{"issue":"1","key":"10_CR18","doi-asserted-by":"publisher","first-page":"444","DOI":"10.1109\/TIP.2014.2372479","volume":"24","author":"T Ruzic","year":"2015","unstructured":"Ruzic, T., Pizurica, A.: Context-aware patch-based image inpainting using Markov random field modeling. IEEE Trans. Image Process. 24(1), 444\u2013456 (2015)","journal-title":"IEEE Trans. Image Process."},{"key":"10_CR19","volume":"98","author":"S Satapathy","year":"2021","unstructured":"Satapathy, S., Sahay, R.R.: Robust depth map inpainting using superpixels and non-local gauss-Markov random field prior. Signal Process.: Image Commun. 98, 116378 (2021)","journal-title":"Signal Process.: Image Commun."},{"key":"10_CR20","doi-asserted-by":"publisher","first-page":"105778","DOI":"10.1016\/j.knosys.2020.105778","volume":"196","author":"M Shao","year":"2020","unstructured":"Shao, M., Zhang, W., Zuo, W., Meng, D.: Multi-scale generative adversarial inpainting network based on cross-layer attention transfer mechanism. Knowl.-Based Syst. 196, 105778 (2020)","journal-title":"Knowl.-Based Syst."},{"key":"10_CR21","doi-asserted-by":"crossref","unstructured":"Shen, L., Hong, R., Zhang, H., Zhang, H., Wang, M.: Single-shot semantic image inpainting with densely connected generative networks. In: Proceedings of the 27th ACM International Conference on Multimedia, MM 2019, pp. 1861\u20131869 (2019)","DOI":"10.1145\/3343031.3350903"},{"issue":"10","key":"10_CR22","doi-asserted-by":"publisher","first-page":"1570","DOI":"10.1109\/TIP.2005.852206","volume":"14","author":"JL Starck","year":"2005","unstructured":"Starck, J.L., Elad, M., Donoho, D.: Image decomposition via the combination of sparse representations and a variational approach. IEEE Trans. Image Process. 14(10), 1570\u20131582 (2005)","journal-title":"IEEE Trans. Image Process."},{"key":"10_CR23","doi-asserted-by":"crossref","unstructured":"Suvorov, R., et al.: Resolution-robust large mask inpainting with Fourier convolutions. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision (WACV), pp. 2149\u20132159 (2022)","DOI":"10.1109\/WACV51458.2022.00323"},{"key":"10_CR24","doi-asserted-by":"crossref","unstructured":"Tao, Y., Popovic, M., Wang, Y., Digumarti, S., Chebrolu, N., Fallon, M.: 3d lidar reconstruction with probabilistic depth completion for robotic navigation. In: 2022 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 5339\u20135346 (2022)","DOI":"10.1109\/IROS47612.2022.9981531"},{"issue":"4","key":"10_CR25","doi-asserted-by":"publisher","first-page":"506","DOI":"10.1109\/TPAMI.2005.87","volume":"27","author":"D Tschumperle","year":"2005","unstructured":"Tschumperle, D., Deriche, R.: Vector-valued image regularization with PDEs: a common framework for different applications. IEEE Trans. Pattern Anal. Machine Intell. 27(4), 506\u2013517 (2005)","journal-title":"IEEE Trans. Pattern Anal. Machine Intell."},{"key":"10_CR26","unstructured":"Wongsa, R.: Pytorch implementation of the paper: image inpainting for irregular holes using partial convolutions. https:\/\/github.com\/ryanwongsa\/Image-Inpainting (2020)"},{"key":"10_CR27","doi-asserted-by":"crossref","unstructured":"Xie, C., et al.: Image inpainting with learnable bidirectional attention maps. In: 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 8857\u20138866 (2019)","DOI":"10.1109\/ICCV.2019.00895"},{"key":"10_CR28","doi-asserted-by":"crossref","unstructured":"Yan, Z., Li, X., Li, M., Zuo, W., Shan, S.: Shift-net: image inpainting via deep feature rearrangement. In: European Conference on Computer Vision (2018)","DOI":"10.1007\/978-3-030-01264-9_1"},{"key":"10_CR29","doi-asserted-by":"crossref","unstructured":"Yeh, R., Chen, C., Lim, T.Y., Schwing, A., Hasegawa-Johnson, M., Do, M.: Semantic image inpainting with deep generative models. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6882\u20136890 (2017)","DOI":"10.1109\/CVPR.2017.728"},{"key":"10_CR30","doi-asserted-by":"crossref","unstructured":"Yu, J., Lin, Z., Yang, J., Shen, X., Lu, X., Huang, T.: Free-form image inpainting with gated convolution. In: 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 4470\u20134479 (2019)","DOI":"10.1109\/ICCV.2019.00457"},{"key":"10_CR31","doi-asserted-by":"crossref","unstructured":"Yu, Y., et al.: Diverse image inpainting with bidirectional and autoregressive transformers. In Proceedings of the 29th ACM International Conference on Multimedia (2021)","DOI":"10.1145\/3474085.3475436"},{"key":"10_CR32","doi-asserted-by":"publisher","first-page":"74","DOI":"10.1016\/j.inffus.2022.08.033","volume":"90","author":"X Zhang","year":"2022","unstructured":"Zhang, X., Zhai, D., Li, T., Zhou, Y., Lin, Y.: Image inpainting based on deep learning: a review. Inf. Fusion 90, 74\u201394 (2022)","journal-title":"Inf. Fusion"},{"key":"10_CR33","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Funkhouser, T.: Deep depth completion of a single RGB-D image. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2018)","DOI":"10.1109\/CVPR.2018.00026"}],"container-title":["Lecture Notes in Computer Science","Advances in Visual Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-47966-3_10","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,4]],"date-time":"2024-11-04T19:28:12Z","timestamp":1730748492000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-47966-3_10"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031479656","9783031479663"],"references-count":33,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-47966-3_10","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"3 December 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ISVC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Symposium on Visual Computing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Lake Tahoe, NV","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"USA","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16 October 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"isvc2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.isvc.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"25","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"58","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"232% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"43 (oral), 15 (poster), 25 (special tracks) out of 34 submissions","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}