{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T18:10:17Z","timestamp":1772907017353,"version":"3.50.1"},"publisher-location":"Cham","reference-count":29,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030585440","type":"print"},{"value":"9783030585457","type":"electronic"}],"license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-3-030-58545-7_27","type":"book-chapter","created":{"date-parts":[[2020,11,4]],"date-time":"2020-11-04T10:04:51Z","timestamp":1604484291000},"page":"467-482","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Conditional Image Repainting via Semantic Bridge and Piecewise Value Function"],"prefix":"10.1007","author":[{"given":"Shuchen","family":"Weng","sequence":"first","affiliation":[]},{"given":"Wenbo","family":"Li","sequence":"additional","affiliation":[]},{"given":"Dawei","family":"Li","sequence":"additional","affiliation":[]},{"given":"Hongxia","family":"Jin","sequence":"additional","affiliation":[]},{"given":"Boxin","family":"Shi","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,11,5]]},"reference":[{"key":"27_CR1","unstructured":"FaceApp. https:\/\/www.faceapp.com\/"},{"key":"27_CR2","doi-asserted-by":"crossref","unstructured":"Caesar, H., Uijlings, J.R.R., Ferrari, V.: Coco-stuff: thing and stuff classes in context. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00132"},{"key":"27_CR3","doi-asserted-by":"crossref","unstructured":"Chen, B., Kae, A.: Toward realistic image compositing with adversarial learning. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00861"},{"key":"27_CR4","unstructured":"Chen, X., Qing, L., He, X., Luo, X., Xu, Y.: FTGAN: a fully-trained generative adversarial networks for text to face generation. CoRR abs\/1904.05729 (2019)"},{"key":"27_CR5","unstructured":"Cong, W., et al.: Deep image harmonization via domain verification. CoRR abs\/1911.13239 (2019)"},{"key":"27_CR6","unstructured":"Cun, X., Pun, C.: Improving the harmony of the composite image by spatial-separated attention module. CoRR abs\/1907.06406 (2019)"},{"key":"27_CR7","unstructured":"Goodfellow, I.J., et al.: Generative adversarial nets. In: NIPS (2014)"},{"key":"27_CR8","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Klambauer, G., Hochreiter, S.: GANs trained by a two time-scale update rule converge to a nash equilibrium. In: NIPS (2017)"},{"key":"27_CR9","first-page":"214","volume":"29","author":"H Huang","year":"2020","unstructured":"Huang, H., Xu, S., Cai, J., Liu, W., Hu, S.: Temporally coherent video harmonization using adversarial networks. TIP 29, 214\u2013224 (2020)","journal-title":"TIP"},{"key":"27_CR10","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00453"},{"key":"27_CR11","doi-asserted-by":"crossref","unstructured":"Li, W., et al.: Object-driven text-to-image synthesis via adversarial training. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.01245"},{"key":"27_CR12","doi-asserted-by":"crossref","unstructured":"Li, Y., Singh, K.K., Ojha, U., Lee, Y.J.: Mixnmatch: multifactor disentanglement and encoding for conditional image generation. CoRR abs\/1911.11758 (2019)","DOI":"10.1109\/CVPR42600.2020.00806"},{"key":"27_CR13","doi-asserted-by":"crossref","unstructured":"Park, T., Liu, M., Wang, T., Zhu, J.: Semantic image synthesis with spatially-adaptive normalization. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00244"},{"key":"27_CR14","doi-asserted-by":"crossref","unstructured":"Pennington, J., Socher, R., Manning, C.D.: Glove: global vectors for word representation. In: EMNLP (2014)","DOI":"10.3115\/v1\/D14-1162"},{"key":"27_CR15","doi-asserted-by":"crossref","unstructured":"Qiao, T., Zhang, J., Xu, D., Tao, D.: Mirrorgan: learning text-to-image generation by redescription. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00160"},{"key":"27_CR16","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"234","DOI":"10.1007\/978-3-319-24574-4_28","volume-title":"Medical Image Computing and Computer-Assisted Intervention \u2013 MICCAI 2015","author":"O Ronneberger","year":"2015","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234\u2013241. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28"},{"key":"27_CR17","doi-asserted-by":"crossref","unstructured":"Tan, H., Liu, X., Li, X., Zhang, Y., Yin, B.: Semantics-enhanced adversarial nets for text-to-image synthesis. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.01060"},{"key":"27_CR18","doi-asserted-by":"crossref","unstructured":"Tripathi, S., Chandra, S., Agrawal, A., Tyagi, A., Rehg, J.M., Chari, V.: Learning to generate synthetic data via compositing. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00055"},{"key":"27_CR19","doi-asserted-by":"crossref","unstructured":"Tsai, Y., Shen, X., Lin, Z., Sunkavalli, K., Lu, X., Yang, M.: Deep image harmonization. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.299"},{"key":"27_CR20","doi-asserted-by":"publisher","first-page":"149","DOI":"10.1145\/2897824.2925942","volume":"35","author":"Y Tsai","year":"2016","unstructured":"Tsai, Y., Shen, X., Lin, Z., Sunkavalli, K., Yang, M.: Sky is not the limit: semantic-aware sky replacement. TOG 35, 149 (2016)","journal-title":"TOG"},{"key":"27_CR21","unstructured":"Wah, C., Branson, S., Welinder, P., Perona, P., Belongie, S.: The Caltech-UCSD Birds-200-2011 Dataset. Technical report CNS-TR-2011-001, California Institute of Technology (2011)"},{"key":"27_CR22","doi-asserted-by":"crossref","unstructured":"Wang, T., Liu, M., Zhu, J., Tao, A., Kautz, J., Catanzaro, B.: High-resolution image synthesis and semantic manipulation with conditional gans. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00917"},{"key":"27_CR23","doi-asserted-by":"crossref","unstructured":"Weng, S., Li, W., Li, D., Jin, H., Shi, B.: Misc: multi-condition injection and spatially-adaptive compositing for conditional person image synthesis. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00776"},{"key":"27_CR24","doi-asserted-by":"crossref","unstructured":"Wu, H., Zheng, S., Zhang, J., Huang, K.: GP-GAN: towards realistic high-resolution image blending. In: ACM MM (2019)","DOI":"10.1145\/3343031.3350944"},{"key":"27_CR25","doi-asserted-by":"crossref","unstructured":"Xu, T., et al.: Attngan: fine-grained text to image generation with attentional generative adversarial networks. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00143"},{"key":"27_CR26","doi-asserted-by":"crossref","unstructured":"Yin, G., Liu, B., Sheng, L., Yu, N., Wang, X., Shao, J.: Semantics disentangling for text-to-image generation. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00243"},{"key":"27_CR27","doi-asserted-by":"crossref","unstructured":"Yu, J., Lin, Z., Yang, J., Shen, X., Lu, X., Huang, T.S.: Free-form image inpainting with gated convolution. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00457"},{"key":"27_CR28","doi-asserted-by":"crossref","unstructured":"Zhou, P., Han, X., Morariu, V.I., Davis, L.S.: Learning rich features for image manipulation detection. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00116"},{"key":"27_CR29","doi-asserted-by":"crossref","unstructured":"Zhu, M., Pan, P., Chen, W., Yang, Y.: DM-GAN: dynamic memory generative adversarial networks for text-to-image synthesis. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00595"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2020"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-58545-7_27","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,4]],"date-time":"2024-11-04T01:11:30Z","timestamp":1730682690000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-58545-7_27"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"ISBN":["9783030585440","9783030585457"],"references-count":29,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-58545-7_27","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020]]},"assertion":[{"value":"5 November 2020","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Glasgow","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 August 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 August 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2020.eu\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"OpenReview","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5025","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1360","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"27% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"7","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"The conference was held virtually due to the COVID-19 pandemic. From the ECCV Workshops 249 full papers, 18 short papers, and 21 further contributions were published out of a total of 467 submissions.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}