{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,4,26]],"date-time":"2025-04-26T09:06:59Z","timestamp":1745658419382,"version":"3.37.3"},"reference-count":41,"publisher":"Springer Science and Business Media LLC","issue":"32","license":[{"start":{"date-parts":[[2024,2,23]],"date-time":"2024-02-23T00:00:00Z","timestamp":1708646400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,2,23]],"date-time":"2024-02-23T00:00:00Z","timestamp":1708646400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-024-18590-5","type":"journal-article","created":{"date-parts":[[2024,2,23]],"date-time":"2024-02-23T08:02:37Z","timestamp":1708675357000},"page":"78387-78406","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Bridging partial-gated convolution with transformer for smooth-variation image inpainting"],"prefix":"10.1007","volume":"83","author":[{"given":"Zeyu","family":"Wang","sequence":"first","affiliation":[]},{"given":"Haibin","family":"Shen","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3722-9979","authenticated-orcid":false,"given":"Kejie","family":"Huang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,2,23]]},"reference":[{"key":"18590_CR1","unstructured":"Jia J, Tang C-K (2003) Image repairing: robust image synthesis by adaptive nd tensor voting. In: 2003 IEEE Computer society conference on computer vision and pattern recognition, 2003. Proceedings, vol 1, p. IEEE"},{"issue":"9","key":"18590_CR2","doi-asserted-by":"publisher","first-page":"1200","DOI":"10.1109\/TIP.2004.833105","volume":"13","author":"A Criminisi","year":"2004","unstructured":"Criminisi A, P\u00e9rez P, Toyama K (2004) Region filling and object removal by exemplar-based image inpainting. IEEE Trans Image Process 13(9):1200\u20131212","journal-title":"IEEE Trans Image Process"},{"key":"18590_CR3","doi-asserted-by":"crossref","unstructured":"Ono N, Aizawa K, Matsui Y (2021) Comic image inpainting via distance transform. In: SIGGRAPH Asia 2021 technical communications, pp 1\u20134","DOI":"10.1145\/3478512.3488607"},{"key":"18590_CR4","doi-asserted-by":"crossref","unstructured":"Bertalmio M, Sapiro G, Caselles V, Ballester C (2000) Image inpainting. In: Proceedings of the 27th annual conference on computer graphics and interactive techniques, pp 417\u2013424","DOI":"10.1145\/344779.344972"},{"key":"18590_CR5","doi-asserted-by":"crossref","unstructured":"Daisy M, Tschumperl\u00e9 D, L\u00e9zoray O (2013) A fast spatial patch blending algorithm for artefact reduction in pattern-based image inpainting. In: SIGGRAPH Asia 2013 Technical Briefs, pp 1\u20134","DOI":"10.1145\/2542355.2542365"},{"issue":"1","key":"18590_CR6","doi-asserted-by":"publisher","first-page":"444","DOI":"10.1109\/TIP.2014.2372479","volume":"24","author":"T Ru\u017ei\u0107","year":"2014","unstructured":"Ru\u017ei\u0107 T, Pi\u017eurica A (2014) Context-aware patch-based image inpainting using Markov random field modeling. IEEE Trans Image Process 24(1):444\u2013456","journal-title":"IEEE Trans Image Process"},{"key":"18590_CR7","doi-asserted-by":"crossref","unstructured":"Pathak D, Krahenbuhl P, Donahue J, Darrell T, Efros AA (2016) Context encoders: feature learning by inpainting. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp","DOI":"10.1109\/CVPR.2016.278"},{"issue":"4","key":"18590_CR8","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3072959.3073659","volume":"36","author":"S Iizuka","year":"2017","unstructured":"Iizuka S, Simo-Serra E, Ishikawa H (2017) Globally and locally consistent image completion. ACM Trans Graph (ToG) 36(4):1\u201314","journal-title":"ACM Trans Graph (ToG)"},{"key":"18590_CR9","unstructured":"Demir U, Unal G (2018) Patch-based image inpainting with generative adversarial networks. arXiv:1803.07422"},{"key":"18590_CR10","doi-asserted-by":"crossref","unstructured":"Liu G, Reda FA, Shih KJ, Wang T-C, Tao A, Catanzaro B (2018) Image inpainting for irregular holes using partial convolutions. In: Proceedings of the european conference on computer vision (ECCV), pp 85\u2013100","DOI":"10.1007\/978-3-030-01252-6_6"},{"key":"18590_CR11","doi-asserted-by":"crossref","unstructured":"Yu J, Lin Z, Yang J, Shen X, Lu X, Huang TS (2019) Free-form image inpainting with gated convolution. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 4471\u20134480","DOI":"10.1109\/ICCV.2019.00457"},{"key":"18590_CR12","unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser \u0141, Polosukhin I (2017) Attention is all you need. Adv Neural Inf Process 30"},{"issue":"12","key":"18590_CR13","doi-asserted-by":"publisher","first-page":"3050","DOI":"10.1109\/TIFS.2017.2730822","volume":"12","author":"H Li","year":"2017","unstructured":"Li H, Luo W, Huang J (2017) Localization of diffusion-based inpainting in digital images. IEEE Trans Inf Forensics Secur 12(12):3050\u20133064","journal-title":"IEEE Trans Inf Forensics Secur"},{"issue":"8","key":"18590_CR14","doi-asserted-by":"publisher","first-page":"1203","DOI":"10.1109\/LGRS.2017.2702106","volume":"14","author":"N Amrani","year":"2017","unstructured":"Amrani N, Serra-Sagrist\u00e0 J, Peter P, Weickert J (2017) Diffusion-based inpainting for coding remote-sensing data. IEEE Geosci Remote Sens 14(8):1203\u20131207","journal-title":"IEEE Geosci Remote Sens"},{"issue":"6","key":"18590_CR15","doi-asserted-by":"publisher","first-page":"2023","DOI":"10.1109\/TVCG.2017.2702738","volume":"24","author":"Q Guo","year":"2017","unstructured":"Guo Q, Gao S, Zhang X, Yin Y, Zhang C (2017) Patch-based image inpainting via two-stage low rank approximation. IEEE Trans Vis Comput Graph 24(6):2023\u20132036","journal-title":"IEEE Trans Vis Comput Graph"},{"issue":"7553","key":"18590_CR16","first-page":"436","volume":"521","author":"Y LeCun","year":"2015","unstructured":"LeCun Y, Bengio Y, Hinton G (2015) Deep Learn Nat 521(7553):436\u2013444","journal-title":"Deep Learn Nat"},{"key":"18590_CR17","doi-asserted-by":"crossref","unstructured":"Yi Z, Tang Q, Azizi S, Jang D, Xu Z (2020) Contextual residual aggregation for ultra high-resolution image inpainting. In: Proceedings of the IEEE\/cvf conference on computer vision and pattern recognition, pp 7508\u20137517","DOI":"10.1109\/CVPR42600.2020.00753"},{"key":"18590_CR18","doi-asserted-by":"crossref","unstructured":"Zeng Y, Lin Z, Lu H, Patel VM (2021) Cr-fill: generative image inpainting with auxiliary contextual reconstruction. In: Proceedings of the IEEE\/CVF international conference on computer vision (ICCV), pp 14164\u201314173","DOI":"10.1109\/ICCV48922.2021.01390"},{"key":"18590_CR19","doi-asserted-by":"crossref","unstructured":"Liu H, Wan Z, Huang W, Song Y, Han X, Liao J (2021) Pd-gan: probabilistic diverse gan for image inpainting. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 9371\u20139381","DOI":"10.1109\/CVPR46437.2021.00925"},{"key":"18590_CR20","doi-asserted-by":"crossref","unstructured":"Zheng C, Cham T-J, Cai J, Phung D (2022) Bridging global context interactions for high-fidelity image completion. In: Proceedings of the IEEE conference on computer vision and pattern recognition","DOI":"10.1109\/CVPR52688.2022.01122"},{"key":"18590_CR21","unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S, Uszkoreit J, Houlsby N (2021) An image is worth 16x16 words: transformers for image recognition at scale. In: 9th International conference on learning representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021"},{"key":"18590_CR22","doi-asserted-by":"crossref","unstructured":"Carion N, Massa F, Synnaeve G, Usunier N, Kirillov A, Zagoruyko S (2020) End-to-end object detection with transformers. In: European conference on computer vision, Springer, pp 213\u2013229","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"18590_CR23","doi-asserted-by":"crossref","unstructured":"Zheng S, Lu J, Zhao H, Zhu X, Luo Z, Wang Y, Fu Y, Feng J, Xiang T, Torr PH et al (2021) Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 6881\u20136890","DOI":"10.1109\/CVPR46437.2021.00681"},{"key":"18590_CR24","doi-asserted-by":"crossref","unstructured":"Esser P, Rombach R, Ommer B (2021) Taming transformers for high-resolution image synthesis. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 12873\u201312883","DOI":"10.1109\/CVPR46437.2021.01268"},{"key":"18590_CR25","doi-asserted-by":"crossref","unstructured":"Wan Z, Zhang J, Chen D, Liao J (2021) High-fidelity pluralistic image completion with transformers. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 4692\u20134701","DOI":"10.1109\/ICCV48922.2021.00465"},{"key":"18590_CR26","doi-asserted-by":"crossref","unstructured":"Li W, Lin Z, Zhou K, Qi L, Wang Y, Jia J (2022) Mat: Mask-aware transformer for large hole image inpainting. arXiv:2203.15270","DOI":"10.1109\/CVPR52688.2022.01049"},{"key":"18590_CR27","doi-asserted-by":"crossref","unstructured":"Yu J, Lin Z, Yang J, Shen X, Lu X, Huang TS (2018) Generative image inpainting with contextual attention. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5505\u20135514","DOI":"10.1109\/CVPR.2018.00577"},{"key":"18590_CR28","doi-asserted-by":"crossref","unstructured":"Nazeri K, Ng E, Joseph T, Qureshi F, Ebrahimi M (2019) Edgeconnect: structure guided image inpainting using edge prediction. In: Proceedings of the IEEE\/CVF international conference on computer vision workshops, pp 0\u20130","DOI":"10.1109\/ICCVW.2019.00408"},{"key":"18590_CR29","doi-asserted-by":"crossref","unstructured":"Wu H, Zhou J, Li Y (2021) Deep generative model for image inpainting with local binary pattern learning and spatial attention. IEEE Trans Multimedia","DOI":"10.1109\/TMM.2021.3111491"},{"key":"18590_CR30","doi-asserted-by":"crossref","unstructured":"Zeng Y, Fu J, Chao H, Guo B (2019) Learning pyramid-context encoder network for high-quality image inpainting. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 1486\u20131494","DOI":"10.1109\/CVPR.2019.00158"},{"issue":"2","key":"18590_CR31","doi-asserted-by":"publisher","first-page":"71","DOI":"10.3390\/info13020071","volume":"13","author":"H Luo","year":"2022","unstructured":"Luo H, Zheng Y (2022) Semantic residual pyramid network for image inpainting. Information 13(2):71","journal-title":"Information"},{"key":"18590_CR32","unstructured":"Devlin J, Chang M-W, Lee K, Toutanova K (2019) Bert: pre-training of deep bidirectional transformers for language understanding. In: NAACL"},{"key":"18590_CR33","unstructured":"Karras T, Aila T, Laine S, Lehtinen J (2018) Progressive growing of gans for improved quality, stability, and variation. In: 6th International conference on learning representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings"},{"key":"18590_CR34","doi-asserted-by":"crossref","unstructured":"Liu Z, Luo P, Wang X, Tang X (2015) Deep learning face attributes in the wild. In: Proceedings of the IEEE international conference on computer vision, pp 3730\u20133738","DOI":"10.1109\/ICCV.2015.425"},{"key":"18590_CR35","doi-asserted-by":"crossref","unstructured":"Karras T, Laine S, Aila T (2019) A style-based generator architecture for generative adversarial networks. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 4401\u20134410","DOI":"10.1109\/CVPR.2019.00453"},{"key":"18590_CR36","doi-asserted-by":"crossref","unstructured":"Doersch C, Singh S, Gupta A, Sivic J, Efros A (2012) What makes paris look like Paris? ACM Trans Graph 31(4)","DOI":"10.1145\/2185520.2335452"},{"key":"18590_CR37","doi-asserted-by":"crossref","unstructured":"Zhou B, Lapedriza A, Khosla A, Oliva A, Torralba A (2017) Places: a 10 million image database for scene recognition. IEEE Trans Pattern Anal Mach Intell 40(6):1452\u20131464","DOI":"10.1109\/TPAMI.2017.2723009"},{"key":"18590_CR38","doi-asserted-by":"publisher","first-page":"4855","DOI":"10.1109\/TIP.2021.3076310","volume":"30","author":"M Zhu","year":"2021","unstructured":"Zhu M, He D, Li X, Li C, Li F, Liu X, Ding E, Zhang Z (2021) Image inpainting by end-to-end cascaded refinement with mask awareness. IEEE Trans Image Process 30:4855\u20134866","journal-title":"IEEE Trans Image Process"},{"issue":"4","key":"18590_CR39","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang Z, Bovik AC, Sheikh HR, Simoncelli EP (2004) Image quality assessment: from error visibility to structural similarity. IEEE Trans Image Process 13(4):600\u2013612","journal-title":"IEEE Trans Image Process"},{"key":"18590_CR40","doi-asserted-by":"crossref","unstructured":"Zhang R, Isola P, Efros AA, Shechtman E, Wang O (2018) The unreasonable effectiveness of deep features as a perceptual metric. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 586\u2013595","DOI":"10.1109\/CVPR.2018.00068"},{"key":"18590_CR41","unstructured":"Heusel M, Ramsauer H, Unterthiner T, Nessler B, Hochreiter S (2017) Gans trained by a two time-scale update rule converge to a local nash equilibrium. Adv Neural Inf Process Systs 30"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-18590-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-024-18590-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-18590-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,4]],"date-time":"2024-09-04T04:14:32Z","timestamp":1725423272000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-024-18590-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,2,23]]},"references-count":41,"journal-issue":{"issue":"32","published-online":{"date-parts":[[2024,9]]}},"alternative-id":["18590"],"URL":"https:\/\/doi.org\/10.1007\/s11042-024-18590-5","relation":{},"ISSN":["1573-7721"],"issn-type":[{"type":"electronic","value":"1573-7721"}],"subject":[],"published":{"date-parts":[[2024,2,23]]},"assertion":[{"value":"14 November 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 January 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 February 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 February 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of interest"}}]}}