{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T03:04:50Z","timestamp":1740107090315,"version":"3.37.3"},"reference-count":67,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2023,8,12]],"date-time":"2023-08-12T00:00:00Z","timestamp":1691798400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,8,12]],"date-time":"2023-08-12T00:00:00Z","timestamp":1691798400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100003787","name":"Natural Science Foundation of Hebei Province","doi-asserted-by":"publisher","award":["No.F2022201009"],"award-info":[{"award-number":["No.F2022201009"]}],"id":[{"id":"10.13039\/501100003787","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Hebei University High-level Scientific Research Foundation for the Introduction of Talent","award":["No.521100221029"],"award-info":[{"award-number":["No.521100221029"]}]},{"name":"Science and Technology Project of Hebei Education Department","award":["No.QN2023186"],"award-info":[{"award-number":["No.QN2023186"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2024,5]]},"DOI":"10.1007\/s00371-023-03045-z","type":"journal-article","created":{"date-parts":[[2023,8,12]],"date-time":"2023-08-12T11:02:07Z","timestamp":1691838127000},"page":"3457-3472","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Dynamic context-driven progressive image inpainting with auxiliary generative units"],"prefix":"10.1007","volume":"40","author":[{"given":"Zhiwen","family":"Wang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6429-8465","authenticated-orcid":false,"given":"Kai","family":"Li","sequence":"additional","affiliation":[]},{"given":"Jinjia","family":"Peng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,8,12]]},"reference":[{"key":"3045_CR1","doi-asserted-by":"crossref","unstructured":"Bertalmio, M., Sapiro, G., Caselles, V., Ballester, C.: Image inpainting. In: Proceedings of the 27th Annual Conference on Computer Graphics and Interactive Techniques, pp. 417\u2013424 (2000)","DOI":"10.1145\/344779.344972"},{"issue":"4","key":"3045_CR2","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2185520.2185578","volume":"31","author":"S Darabi","year":"2012","unstructured":"Darabi, S., Shechtman, E., Barnes, C., Goldman, D.B., Sen, P.: Image melding: combining inconsistent images using patch-based synthesis. ACM Trans. Graph. (TOG) 31(4), 1\u201310 (2012)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"3045_CR3","doi-asserted-by":"crossref","unstructured":"Pathak, D., Krahenbuhl, P., Donahue, J., Darrell, T., Efros, A.A.: Context encoders: feature learning by inpainting. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2536\u20132544 (2016)","DOI":"10.1109\/CVPR.2016.278"},{"key":"3045_CR4","doi-asserted-by":"crossref","unstructured":"Ren, Y., Yu, X., Zhang, R., Li, T.H., Liu, S., Li, G.: Structureflow: image inpainting via structure-aware appearance flow. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 181\u2013190 (2019)","DOI":"10.1109\/ICCV.2019.00027"},{"key":"3045_CR5","doi-asserted-by":"crossref","unstructured":"Xiong, W., Yu, J., Lin, Z., Yang, J., Lu, X., Barnes, C., Luo, J.: Foreground-aware image inpainting. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5840\u20135848 (2019)","DOI":"10.1109\/CVPR.2019.00599"},{"key":"3045_CR6","doi-asserted-by":"publisher","first-page":"1691","DOI":"10.1007\/s00371-020-01932-3","volume":"37","author":"Y Chen","year":"2021","unstructured":"Chen, Y., Liu, L., Tao, J., Xia, R., Zhang, Q., Yang, K., Xiong, J., Chen, X.: The improved image inpainting algorithm via encoder and similarity constraint. Vis. Comput. 37, 1691\u20131705 (2021)","journal-title":"Vis. Comput."},{"key":"3045_CR7","doi-asserted-by":"crossref","unstructured":"Yan, Z., Li, X., Li, M., Zuo, W., Shan, S.: Shift-net: image inpainting via deep feature rearrangement. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 1\u201317 (2018)","DOI":"10.1007\/978-3-030-01264-9_1"},{"key":"3045_CR8","doi-asserted-by":"crossref","unstructured":"Xie, C., Liu, S., Li, C., Cheng, M.-M., Zuo, W., Liu, X., Wen, S., Ding, E.: Image inpainting with learnable bidirectional attention maps. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 8858\u20138867 (2019)","DOI":"10.1109\/ICCV.2019.00895"},{"key":"3045_CR9","unstructured":"Nazeri, K., Ng, E., Joseph, T., Qureshi, F.Z., Ebrahimi, M.: Edgeconnect: generative image inpainting with adversarial edge learning. arXiv:1901.00212 (2019)"},{"key":"3045_CR10","doi-asserted-by":"crossref","unstructured":"Guo, X., Yang, H., Huang, D.: Image inpainting via conditional texture and structure dual generation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 14134\u201314143 (2021)","DOI":"10.1109\/ICCV48922.2021.01387"},{"key":"3045_CR11","doi-asserted-by":"crossref","unstructured":"Liu, H., Jiang, B., Song, Y., Huang, W., Yang, C.: Rethinking image inpainting via a mutual encoder\u2013decoder with feature equalizations. In: European Conference on Computer Vision, pp. 725\u2013741. Springer (2020)","DOI":"10.1007\/978-3-030-58536-5_43"},{"key":"3045_CR12","doi-asserted-by":"crossref","unstructured":"Lin, J., Wang, Y.-G., Tang, W., Li, A.: Multi-feature co-learning for image inpainting. arXiv:2205.10578 (2022)","DOI":"10.1109\/ICPR56361.2022.9956475"},{"key":"3045_CR13","doi-asserted-by":"crossref","unstructured":"Yu, J., Lin, Z., Yang, J., Shen, X., Lu, X., Huang, T.S.: Generative image inpainting with contextual attention. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5505\u20135514 (2018)","DOI":"10.1109\/CVPR.2018.00577"},{"key":"3045_CR14","doi-asserted-by":"crossref","unstructured":"Zhang, W., Zhu, J., Tai, Y., Wang, Y., Chu, W., Ni, B., Wang, C., Yang, X.: Context-aware image inpainting with learned semantic priors. arXiv:2106.07220 (2021)","DOI":"10.24963\/ijcai.2021\/183"},{"issue":"4","key":"3045_CR15","doi-asserted-by":"publisher","DOI":"10.1117\/1.JEI.29.4.043026","volume":"29","author":"L Dou","year":"2020","unstructured":"Dou, L., Qian, Z., Qin, C., Feng, G., Zhang, X.: Anti-forensics of diffusion-based image inpainting. J. Electron. Imaging 29(4), 043026 (2020)","journal-title":"J. Electron. Imaging"},{"issue":"3","key":"3045_CR16","doi-asserted-by":"publisher","first-page":"885","DOI":"10.1007\/s00500-014-1547-7","volume":"20","author":"K Li","year":"2016","unstructured":"Li, K., Wei, Y., Yang, Z., Wei, W.: Image inpainting algorithm based on tv model and evolutionary algorithm. Soft Comput. 20(3), 885\u2013893 (2016)","journal-title":"Soft Comput."},{"issue":"8","key":"3045_CR17","doi-asserted-by":"publisher","first-page":"3802","DOI":"10.1007\/s00034-019-01029-w","volume":"38","author":"G Sridevi","year":"2019","unstructured":"Sridevi, G., Srinivas Kumar, S.: Image inpainting based on fractional-order nonlinear diffusion for image reconstruction. Circuits Syst. Signal Process. 38(8), 3802\u20133817 (2019)","journal-title":"Circuits Syst. Signal Process."},{"issue":"4","key":"3045_CR18","doi-asserted-by":"publisher","first-page":"1705","DOI":"10.1109\/TIP.2018.2880681","volume":"28","author":"D Ding","year":"2018","unstructured":"Ding, D., Ram, S., Rodr\u00edguez, J.J.: Image inpainting using nonlocal texture matching and nonlinear filtering. IEEE Trans. Image Process. 28(4), 1705\u20131719 (2018)","journal-title":"IEEE Trans. Image Process."},{"key":"3045_CR19","doi-asserted-by":"publisher","first-page":"69728","DOI":"10.1109\/ACCESS.2018.2877401","volume":"6","author":"M Isogawa","year":"2018","unstructured":"Isogawa, M., Mikami, D., Iwai, D., Kimata, H., Sato, K.: Mask optimization for image inpainting. IEEE Access 6, 69728\u201369741 (2018)","journal-title":"IEEE Access"},{"issue":"4","key":"3045_CR20","doi-asserted-by":"publisher","first-page":"3549","DOI":"10.1007\/s13369-018-3592-5","volume":"44","author":"J Zeng","year":"2019","unstructured":"Zeng, J., Fu, X., Leng, L., Wang, C.: Image inpainting algorithm based on saliency map and gray entropy. Arab. J. Sci. Eng. 44(4), 3549\u20133558 (2019)","journal-title":"Arab. J. Sci. Eng."},{"issue":"6","key":"3045_CR21","doi-asserted-by":"publisher","first-page":"2023","DOI":"10.1109\/TVCG.2017.2702738","volume":"24","author":"Q Guo","year":"2017","unstructured":"Guo, Q., Gao, S., Zhang, X., Yin, Y., Zhang, C.: Patch-based image inpainting via two-stage low rank approximation. IEEE Trans. Vis. Comput. Graph. 24(6), 2023\u20132036 (2017)","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"3045_CR22","doi-asserted-by":"publisher","first-page":"8561","DOI":"10.1109\/TIP.2020.3015545","volume":"29","author":"Z Zha","year":"2020","unstructured":"Zha, Z., Yuan, X., Zhou, J., Zhu, C., Wen, B.: Image restoration via simultaneous nonlocal self-similarity priors. IEEE Trans. Image Process. 29, 8561\u20138576 (2020)","journal-title":"IEEE Trans. Image Process."},{"issue":"6","key":"3045_CR23","doi-asserted-by":"publisher","first-page":"4172","DOI":"10.1016\/j.jfranklin.2023.01.041","volume":"360","author":"Y Li","year":"2023","unstructured":"Li, Y., Jiang, Y., Zhang, H., Liu, J., Ding, X., Gui, G.: Nonconvex l1\/2-regularized nonlocal self-similarity denoiser for compressive sensing based CT reconstruction. J. Frankl. Inst. 360(6), 4172\u20134195 (2023)","journal-title":"J. Frankl. Inst."},{"key":"3045_CR24","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2023.120456","volume":"228","author":"Y Li","year":"2023","unstructured":"Li, Y., Gao, L., Hu, S., Gui, G., Chen, C.-Y.: Nonlocal low-rank plus deep denoising prior for robust image compressed sensing reconstruction. Expert Syst. Appl. 228, 120456 (2023)","journal-title":"Expert Syst. Appl."},{"issue":"9\u201310","key":"3045_CR25","doi-asserted-by":"publisher","first-page":"3205","DOI":"10.1007\/s00371-022-02532-z","volume":"38","author":"X Zhang","year":"2022","unstructured":"Zhang, X., Ma, W., Varinlioglu, G., Rauh, N., He, L., Aliaga, D.: Guided pluralistic building contour completion. Vis. Comput. 38(9\u201310), 3205\u20133216 (2022)","journal-title":"Vis. Comput."},{"key":"3045_CR26","doi-asserted-by":"crossref","unstructured":"Li, J., Wang, N., Zhang, L., Du, B., Tao, D.: Recurrent feature reasoning for image inpainting. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7760\u20137768 (2020)","DOI":"10.1109\/CVPR42600.2020.00778"},{"key":"3045_CR27","doi-asserted-by":"crossref","unstructured":"Lugmayr, A., Danelljan, M., Romero, A., Yu, F., Timofte, R., Van Gool, L.: Repaint: inpainting using denoising diffusion probabilistic models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11461\u201311471 (2022)","DOI":"10.1109\/CVPR52688.2022.01117"},{"key":"3045_CR28","doi-asserted-by":"crossref","unstructured":"Peng, J., Liu, D., Xu, S., Li, H.: Generating diverse structure for image inpainting with hierarchical VQ-VAE. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10775\u201310784 (2021)","DOI":"10.1109\/CVPR46437.2021.01063"},{"key":"3045_CR29","doi-asserted-by":"publisher","first-page":"376","DOI":"10.1016\/j.neucom.2022.11.074","volume":"520","author":"A Li","year":"2023","unstructured":"Li, A., Zhao, L., Zuo, Z., Wang, Z., Xing, W., Lu, D.: Migt: multi-modal image inpainting guided with text. Neurocomputing 520, 376\u2013385 (2023)","journal-title":"Neurocomputing"},{"key":"3045_CR30","doi-asserted-by":"crossref","unstructured":"Yang, Y., Cheng, Z., Yu, H., Zhang, Y., Cheng, X., Zhang, Z., Xie, G.: MSE-NET: generative image inpainting with multi-scale encoder. Vis. Comput. 1\u201313 (2021)","DOI":"10.1007\/s00371-021-02143-0"},{"issue":"9\u201310","key":"3045_CR31","doi-asserted-by":"publisher","first-page":"3149","DOI":"10.1007\/s00371-022-02523-0","volume":"38","author":"Y Xie","year":"2022","unstructured":"Xie, Y., Lin, Z., Yang, Z., Deng, H., Wu, X., Mao, X., Li, Q., Liu, W.: Learning semantic alignment from image for text-guided image inpainting. Vis. Comput. 38(9\u201310), 3149\u20133161 (2022)","journal-title":"Vis. Comput."},{"key":"3045_CR32","doi-asserted-by":"crossref","unstructured":"Lahiri, A., Jain, A.K., Agrawal, S., Mitra, P., Biswas, P.K.: Prior guided GAN based semantic inpainting. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13696\u201313705 (2020)","DOI":"10.1109\/CVPR42600.2020.01371"},{"key":"3045_CR33","doi-asserted-by":"crossref","unstructured":"Yi, Z., Tang, Q., Azizi, S., Jang, D., Xu, Z.: Contextual residual aggregation for ultra high-resolution image inpainting. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7508\u20137517 (2020)","DOI":"10.1109\/CVPR42600.2020.00753"},{"key":"3045_CR34","doi-asserted-by":"crossref","unstructured":"Zhao, L., Mo, Q., Lin, S., Wang, Z., Zuo, Z., Chen, H., Xing, W., Lu, D.: Uctgan: diverse image inpainting based on unsupervised cross-space translation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5741\u20135750 (2020)","DOI":"10.1109\/CVPR42600.2020.00578"},{"key":"3045_CR35","doi-asserted-by":"crossref","unstructured":"Liao, L., Xiao, J., Wang, Z., Lin, C.-W., Satoh, S.: Image inpainting guided by coherence priors of semantics and textures. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6539\u20136548 (2021)","DOI":"10.1109\/CVPR46437.2021.00647"},{"key":"3045_CR36","doi-asserted-by":"crossref","unstructured":"Zeng, Y., Fu, J., Chao, H., Guo, B.: Aggregated contextual transformations for high-resolution image inpainting. IEEE Trans. Vis. Comput. Graph. (2022)","DOI":"10.1109\/TVCG.2022.3156949"},{"key":"3045_CR37","doi-asserted-by":"crossref","unstructured":"Liu, H., Wan, Z., Huang, W., Song, Y., Han, X., Liao, J.: Pd-gan: probabilistic diverse gan for image inpainting. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 9371\u20139381 (2021)","DOI":"10.1109\/CVPR46437.2021.00925"},{"key":"3045_CR38","doi-asserted-by":"publisher","first-page":"2895","DOI":"10.1007\/s00371-021-02218-y","volume":"37","author":"H Li","year":"2021","unstructured":"Li, H., Zhong, Z., Guan, W., Du, C., Yang, Y., Wei, Y., Ye, C.: Generative character inpainting guided by structural information. Vis. Comput. 37, 2895\u20132906 (2021)","journal-title":"Vis. Comput."},{"issue":"4","key":"3045_CR39","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3072959.3073659","volume":"36","author":"S Iizuka","year":"2017","unstructured":"Iizuka, S., Simo-Serra, E., Ishikawa, H.: Globally and locally consistent image completion. ACM Trans. Graph. (ToG) 36(4), 1\u201314 (2017)","journal-title":"ACM Trans. Graph. (ToG)"},{"key":"3045_CR40","doi-asserted-by":"crossref","unstructured":"Liu, G., Reda, F.A., Shih, K.J., Wang, T.-C., Tao, A., Catanzaro, B.: Image inpainting for irregular holes using partial convolutions. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 85\u2013100 (2018)","DOI":"10.1007\/978-3-030-01252-6_6"},{"key":"3045_CR41","doi-asserted-by":"crossref","unstructured":"Yu, J., Lin, Z., Yang, J., Shen, X., Lu, X., Huang, T.S.: Free-form image inpainting with gated convolution. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4471\u20134480 (2019)","DOI":"10.1109\/ICCV.2019.00457"},{"key":"3045_CR42","doi-asserted-by":"crossref","unstructured":"Suvorov, R., Logacheva, E., Mashikhin, A., Remizova, A., Ashukha, A., Silvestrov, A., Kong, N., Goka, H., Park, K., Lempitsky, V.: Resolution-robust large mask inpainting with Fourier convolutions. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 2149\u20132159 (2022)","DOI":"10.1109\/WACV51458.2022.00323"},{"key":"3045_CR43","first-page":"4479","volume":"33","author":"L Chi","year":"2020","unstructured":"Chi, L., Jiang, B., Mu, Y.: Fast Fourier convolution. Adv. Neural Inf. Process. Syst. 33, 4479\u20134488 (2020)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"3045_CR44","unstructured":"Wang, D., Xie, C., Liu, S., Niu, Z., Zuo, W.: Image inpainting with edge-guided learnable bidirectional attention maps. arXiv:2104.12087 (2021)"},{"key":"3045_CR45","doi-asserted-by":"crossref","unstructured":"Wang, C., Zhu, Y., Yuan, C.: Diverse image inpainting with normalizing flow. In: Computer Vision\u2013ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23\u201327, 2022, Proceedings, Part XXIII, pp. 53\u201369. Springer (2022)","DOI":"10.1007\/978-3-031-20050-2_4"},{"issue":"12","key":"3045_CR46","doi-asserted-by":"publisher","first-page":"9717","DOI":"10.1007\/s00521-022-06961-8","volume":"34","author":"J Yu","year":"2022","unstructured":"Yu, J., Li, K., Peng, J.: Reference-guided face inpainting with reference attention network. Neural Comput. Appl. 34(12), 9717\u20139731 (2022)","journal-title":"Neural Comput. Appl."},{"key":"3045_CR47","doi-asserted-by":"crossref","unstructured":"Li, W., Lin, Z., Zhou, K., Qi, L., Wang, Y., Jia, J.: Mat: mask-aware transformer for large hole image inpainting. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10758\u201310768 (2022)","DOI":"10.1109\/CVPR52688.2022.01049"},{"key":"3045_CR48","unstructured":"Zhao, S., Cui, J., Sheng, Y., Dong, Y., Liang, X., Chang, E.I., Xu, Y.: Large scale image completion via co-modulated generative adversarial networks. arXiv:2103.10428 (2021)"},{"key":"3045_CR49","doi-asserted-by":"crossref","unstructured":"Yu, Y., Zhang, L., Fan, H., Luo, T.: High-fidelity image inpainting with gan inversion. In: Computer Vision\u2013ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23\u201327, 2022, Proceedings, Part XVI, Springer, pp. 242\u2013258 (2022)","DOI":"10.1007\/978-3-031-19787-1_14"},{"key":"3045_CR50","unstructured":"Dumoulin, V., Shlens, J., Kudlur, M.: A learned representation for artistic style. arXiv:1610.07629 (2016)"},{"key":"3045_CR51","doi-asserted-by":"crossref","unstructured":"Huang, X., Belongie, S.: Arbitrary style transfer in real-time with adaptive instance normalization. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1501\u20131510 (2017)","DOI":"10.1109\/ICCV.2017.167"},{"key":"3045_CR52","unstructured":"Razavi, A., Van den Oord, A., Vinyals, O.: Generating diverse high-fidelity images with vq-vae-2. Adv. Neural Inf. Process. Syst. 32 (2019)"},{"key":"3045_CR53","doi-asserted-by":"crossref","unstructured":"Esser, P., Rombach, R., Ommer, B.: Taming transformers for high-resolution image synthesis. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12873\u201312883 (2021)","DOI":"10.1109\/CVPR46437.2021.01268"},{"key":"3045_CR54","first-page":"6840","volume":"33","author":"J Ho","year":"2020","unstructured":"Ho, J., Jain, A., Abbeel, P.: Denoising diffusion probabilistic models. Adv. Neural Inf. Process. Syst. 33, 6840\u20136851 (2020)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"3045_CR55","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7132\u20137141 (2018)","DOI":"10.1109\/CVPR.2018.00745"},{"key":"3045_CR56","doi-asserted-by":"crossref","unstructured":"Huang, Z., Wang, X., Huang, L., Huang, C., Wei, Y., Liu, W.: Ccnet: criss-cross attention for semantic segmentation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 603\u2013612 (2019)","DOI":"10.1109\/ICCV.2019.00069"},{"key":"3045_CR57","doi-asserted-by":"crossref","unstructured":"Johnson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time style transfer and super-resolution. In: European Conference on Computer Vision, pp. 694\u2013711. Springer (2016)","DOI":"10.1007\/978-3-319-46475-6_43"},{"key":"3045_CR58","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556 (2014)"},{"issue":"3","key":"3045_CR59","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., Huang, Z., Karpathy, A., Khosla, A., Bernstein, M., et al.: Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 115(3), 211\u2013252 (2015)","journal-title":"Int. J. Comput. Vis."},{"key":"3045_CR60","unstructured":"Karras, T., Aila, T., Laine, S., Lehtinen, J.: Progressive growing of gans for improved quality, stability, and variation. arXiv:1710.10196 (2017)"},{"issue":"4","key":"3045_CR61","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2185520.2185597","volume":"31","author":"C Doersch","year":"2012","unstructured":"Doersch, C., Singh, S., Gupta, A., Sivic, J., Efros, A.: What makes Paris look like Paris? ACM Trans. Graph. 31(4), 1\u20139 (2012)","journal-title":"ACM Trans. Graph."},{"key":"3045_CR62","first-page":"8024","volume":"32","author":"A Paszke","year":"2019","unstructured":"Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., et al.: Pytorch: an imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 32, 8024\u20138035 (2019)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"3045_CR63","unstructured":"Wang, Y., Tao, X., Qi, X., Shen, X., Jia, J.: Image inpainting via generative multi-column convolutional neural networks. Adv. Neural Inf. Process. Syst. 31 (2018)"},{"key":"3045_CR64","doi-asserted-by":"crossref","unstructured":"Zheng, C., Cham, T.-J., Cai, J.: Pluralistic image completion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1438\u20131447 (2019)","DOI":"10.1109\/CVPR.2019.00153"},{"key":"3045_CR65","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2020.107448","volume":"106","author":"N Wang","year":"2020","unstructured":"Wang, N., Ma, S., Li, J., Zhang, Y., Zhang, L.: Multistage attention network for image inpainting. Pattern Recognit. 106, 107448 (2020)","journal-title":"Pattern Recognit."},{"key":"3045_CR66","doi-asserted-by":"crossref","unstructured":"Wadhwa, G., Dhall, A., Murala, S., Tariq, U.: Hyperrealistic image inpainting with hypergraphs. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 3912\u20133921 (2021)","DOI":"10.1109\/WACV48630.2021.00396"},{"key":"3045_CR67","doi-asserted-by":"crossref","unstructured":"Liu, W., Cun, X., Pun, C.-M., Xia, M., Zhang, Y., Wang, J.: Coordfill: efficient high-resolution image inpainting via parameterized coordinate querying. arXiv:2303.08524 (2023)","DOI":"10.1609\/aaai.v37i2.25263"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-03045-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-023-03045-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-03045-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,4,19]],"date-time":"2024-04-19T09:09:21Z","timestamp":1713517761000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-023-03045-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,8,12]]},"references-count":67,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2024,5]]}},"alternative-id":["3045"],"URL":"https:\/\/doi.org\/10.1007\/s00371-023-03045-z","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"type":"print","value":"0178-2789"},{"type":"electronic","value":"1432-2315"}],"subject":[],"published":{"date-parts":[[2023,8,12]]},"assertion":[{"value":"23 July 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 August 2023","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"All the authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"All data used in this study were obtained from publicly available datasets.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical and informed consent for data used"}}]}}