{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,5]],"date-time":"2026-03-05T16:41:18Z","timestamp":1772728878573,"version":"3.50.1"},"publisher-location":"Cham","reference-count":45,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031197895","type":"print"},{"value":"9783031197901","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19790-1_1","type":"book-chapter","created":{"date-parts":[[2022,10,23]],"date-time":"2022-10-23T07:02:44Z","timestamp":1666508564000},"page":"1-17","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":16,"title":["Editing Out-of-Domain GAN Inversion via\u00a0Differential Activations"],"prefix":"10.1007","author":[{"given":"Haorui","family":"Song","sequence":"first","affiliation":[]},{"given":"Yong","family":"Du","sequence":"additional","affiliation":[]},{"given":"Tianyi","family":"Xiang","sequence":"additional","affiliation":[]},{"given":"Junyu","family":"Dong","sequence":"additional","affiliation":[]},{"given":"Jing","family":"Qin","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3802-4644","authenticated-orcid":false,"given":"Shengfeng","family":"He","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,24]]},"reference":[{"key":"1_CR1","doi-asserted-by":"crossref","unstructured":"Abdal, R., Qin, Y., Wonka, P.: Image2stylegan: how to embed images into the stylegan latent space? In: ICCV, pp. 4432\u20134441 (2019)","DOI":"10.1109\/ICCV.2019.00453"},{"key":"1_CR2","doi-asserted-by":"crossref","unstructured":"Abdal, R., Qin, Y., Wonka, P.: Image2stylegan++: how to edit the embedded images? In: CVPR, pp. 8296\u20138305 (2020)","DOI":"10.1109\/CVPR42600.2020.00832"},{"issue":"3","key":"1_CR3","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3447648","volume":"40","author":"R Abdal","year":"2021","unstructured":"Abdal, R., Zhu, P., Mitra, N.J., Wonka, P.: Styleflow: attribute-conditioned exploration of stylegan-generated images using conditional continuous normalizing flows. ACM TOG 40(3), 1\u201321 (2021)","journal-title":"ACM TOG"},{"key":"1_CR4","doi-asserted-by":"crossref","unstructured":"Alaluf, Y., Patashnik, O., Cohen-Or, D.: Restyle: a residual-based StyleGAN encoder via iterative refinement. In: ICCV, pp. 6711\u20136720 (2021)","DOI":"10.1109\/ICCV48922.2021.00664"},{"key":"1_CR5","unstructured":"Arjovsky, M., Chintala, S., Bottou, L.: Wasserstein generative adversarial networks. In: ICML, pp. 214\u2013223. PMLR (2017)"},{"key":"1_CR6","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"618","DOI":"10.1007\/978-3-030-58555-6_37","volume-title":"Computer Vision \u2013 ECCV 2020","author":"W Bae","year":"2020","unstructured":"Bae, W., Noh, J., Kim, G.: Rethinking class activation mapping for weakly supervised object localization. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12360, pp. 618\u2013634. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58555-6_37"},{"key":"1_CR7","unstructured":"Bau, D., et al.: Inverting layers of a large generator. In: ICLR, vol. 2, p. 4 (2019)"},{"key":"1_CR8","unstructured":"Chai, L., Wulff, J., Isola, P.: Using latent space regression to analyze and leverage compositionality in GANS. arXiv preprint arXiv:2103.10426 (2021)"},{"key":"1_CR9","doi-asserted-by":"crossref","unstructured":"Choi, Y., Choi, M., Kim, M., Ha, J.W., Kim, S., Choo, J.: StarGAN: unified generative adversarial networks for multi-domain image-to-image translation. In: CVPR, pp. 8789\u20138797 (2018)","DOI":"10.1109\/CVPR.2018.00916"},{"key":"1_CR10","doi-asserted-by":"crossref","unstructured":"Collins, E., Bala, R., Price, B., Susstrunk, S.: Editing in style: uncovering the local semantics of GANS. In: CVPR, pp. 5771\u20135780 (2020)","DOI":"10.1109\/CVPR42600.2020.00581"},{"key":"1_CR11","doi-asserted-by":"crossref","unstructured":"Fukui, H., Hirakawa, T., Yamashita, T., Fujiyoshi, H.: Attention branch network: learning of attention mechanism for visual explanation. In: CVPR, pp. 10705\u201310714 (2019)","DOI":"10.1109\/CVPR.2019.01096"},{"key":"1_CR12","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. Adv. Neural Inf. Process. Syst. 27 (2014)"},{"key":"1_CR13","doi-asserted-by":"crossref","unstructured":"Gu, J., Shen, Y., Zhou, B.: Image processing using multi-code GAN prior. In: CVPR, pp. 3012\u20133021 (2020)","DOI":"10.1109\/CVPR42600.2020.00308"},{"key":"1_CR14","unstructured":"Guan, S., Tai, Y., Ni, B., Zhu, F., Huang, F., Yang, X.: Collaborative learning for faster stylegan embedding. arXiv preprint arXiv:2007.01758 (2020)"},{"issue":"11","key":"1_CR15","first-page":"5464","volume":"28","author":"Z He","year":"2019","unstructured":"He, Z., Zuo, W., Kan, M., Shan, S., Chen, X.: Attgan: facial attribute editing by only changing what you want. IEEE TIP 28(11), 5464\u20135478 (2019)","journal-title":"IEEE TIP"},{"key":"1_CR16","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANS trained by a two time-scale update rule converge to a local nash equilibrium. In: NeurIPS (2017)"},{"key":"1_CR17","doi-asserted-by":"crossref","unstructured":"Isola, P., Zhu, J.Y., Zhou, T., Efros, A.A.: Image-to-image translation with conditional adversarial networks. In: CVPR, pp. 1125\u20131134 (2017)","DOI":"10.1109\/CVPR.2017.632"},{"key":"1_CR18","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: CVPR, pp. 4401\u20134410 (2019)","DOI":"10.1109\/CVPR.2019.00453"},{"key":"1_CR19","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of StyleGAN. In: CVPR, pp. 8110\u20138119 (2020)","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"1_CR20","doi-asserted-by":"crossref","unstructured":"Kim, D., Khan, M.A., Choo, J.: Not just compete, but collaborate: local image-to-image translation via cooperative mask prediction. In: CVPR, pp. 6509\u20136518 (2021)","DOI":"10.1109\/CVPR46437.2021.00644"},{"key":"1_CR21","unstructured":"Kim, T., Cha, M., Kim, H., Lee, J.K., Kim, J.: Learning to discover cross-domain relations with generative adversarial networks. In: ICML, pp. 1857\u20131865. PMLR (2017)"},{"key":"1_CR22","doi-asserted-by":"crossref","unstructured":"Lee, K.H., Park, C., Oh, J., Kwak, N.: LFI-CAM: learning feature importance for better visual explanation. In: ICCV, pp. 1355\u20131363 (2021)","DOI":"10.1109\/ICCV48922.2021.00139"},{"key":"1_CR23","doi-asserted-by":"crossref","unstructured":"Liu, M., Ding, Y., Xia, M., Liu, X., Ding, E., Zuo, W., Wen, S.: STGAN: a unified selective transfer network for arbitrary image attribute editing. In: CVPR, pp. 3673\u20133682 (2019)","DOI":"10.1109\/CVPR.2019.00379"},{"key":"1_CR24","doi-asserted-by":"crossref","unstructured":"Liu, Z., Luo, P., Wang, X., Tang, X.: Deep learning face attributes in the wild. In: ICCV, pp. 3730\u20133738 (2015)","DOI":"10.1109\/ICCV.2015.425"},{"key":"1_CR25","doi-asserted-by":"crossref","unstructured":"Long, J., Shelhamer, E., Darrell, T.: Fully convolutional networks for semantic segmentation. In: CVPR, pp. 3431\u20133440 (2015)","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"1_CR26","doi-asserted-by":"crossref","unstructured":"Patro, B.N., Lunayach, M., Patel, S., Namboodiri, V.P.: U-cam: visual explanation using uncertainty based class activation maps. In: ICCV, pp. 7444\u20137453 (2019)","DOI":"10.1109\/ICCV.2019.00754"},{"key":"1_CR27","doi-asserted-by":"crossref","unstructured":"Raj, A., Li, Y., Bresler, Y.: Gan-based projector for faster recovery with convergence guarantees in linear inverse problems. In: ICCV, pp. 5602\u20135611 (2019)","DOI":"10.1109\/ICCV.2019.00570"},{"key":"1_CR28","doi-asserted-by":"crossref","unstructured":"Richardson, E., et al.: Encoding in style: a StyleGAN encoder for image-to-image translation. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00232"},{"key":"1_CR29","doi-asserted-by":"crossref","unstructured":"Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., Batra, D.: Grad-cam: visual explanations from deep networks via gradient-based localization. In: ICCV, pp. 618\u2013626 (2017)","DOI":"10.1109\/ICCV.2017.74"},{"key":"1_CR30","doi-asserted-by":"crossref","unstructured":"Shen, Y., Gu, J., Tang, X., Zhou, B.: Interpreting the latent space of GANS for semantic face editing. In: CVPR, pp. 9243\u20139252 (2020)","DOI":"10.1109\/CVPR42600.2020.00926"},{"key":"1_CR31","unstructured":"Shen, Y., Yang, C., Tang, X., Zhou, B.: InterFaceGAN: interpreting the disentangled face representation learned by GANS. IEEE TPAMI (2020)"},{"issue":"4","key":"1_CR32","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459838","volume":"40","author":"O Tov","year":"2021","unstructured":"Tov, O., Alaluf, Y., Nitzan, Y., Patashnik, O., Cohen-Or, D.: Designing an encoder for StyleGAN image manipulation. ACM TOG 40(4), 1\u201314 (2021)","journal-title":"ACM TOG"},{"key":"1_CR33","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"170","DOI":"10.1007\/978-3-030-58542-6_11","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Y Viazovetskyi","year":"2020","unstructured":"Viazovetskyi, Y., Ivashkin, V., Kashin, E.: StyleGAN2 distillation for feed-forward image manipulation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12367, pp. 170\u2013186. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58542-6_11"},{"key":"1_CR34","doi-asserted-by":"crossref","unstructured":"Xu, Y., Deng, B., Wang, J., Jing, Y., Pan, J., He, S.: High-resolution face swapping via latent semantics disentanglement. In: CVPR, pp. 7642\u20137651 (2022)","DOI":"10.1109\/CVPR52688.2022.00749"},{"key":"1_CR35","doi-asserted-by":"crossref","unstructured":"Xu, Y., Du, Y., Xiao, W., Xu, X., He, S.: From continuity to editability: inverting GANS with consecutive images. In: ICCV, pp. 13910\u201313918 (2021)","DOI":"10.1109\/ICCV48922.2021.01365"},{"key":"1_CR36","doi-asserted-by":"crossref","unstructured":"Xu, Y., Shen, Y., Zhu, J., Yang, C., Zhou, B.: Generative hierarchical features from synthesizing images. In: CVPR, pp. 4432\u20134442 (2021)","DOI":"10.1109\/CVPR46437.2021.00441"},{"key":"1_CR37","doi-asserted-by":"crossref","unstructured":"Yang, H., Chai, L., Wen, Q., Zhao, S., Sun, Z., He, S.: Discovering interpretable latent space directions of GANS beyond binary attributes. In: CVPR, pp. 12177\u201312185 (2021)","DOI":"10.1109\/CVPR46437.2021.01200"},{"key":"1_CR38","doi-asserted-by":"crossref","unstructured":"Yang, T., Ren, P., Xie, X., Zhang, L.: Gan prior embedded network for blind face restoration in the wild. In: CVPR, pp. 672\u2013681 (2021)","DOI":"10.1109\/CVPR46437.2021.00073"},{"key":"1_CR39","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: CVPR, pp. 586\u2013595 (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"1_CR40","doi-asserted-by":"crossref","unstructured":"Zhong, Z., Chai, L., Zhou, Y., Deng, B., Pan, J., He, S.: Faithful extreme rescaling via generative prior reciprocated invertible representations. In: CVPR, pp. 5708\u20135717 (2022)","DOI":"10.1109\/CVPR52688.2022.00562"},{"key":"1_CR41","doi-asserted-by":"crossref","unstructured":"Zhou, B., Khosla, A., Lapedriza, A., Oliva, A., Torralba, A.: Learning deep features for discriminative localization. In: CVPR, pp. 2921\u20132929 (2016)","DOI":"10.1109\/CVPR.2016.319"},{"key":"1_CR42","first-page":"1230","volume":"31","author":"Y Zhou","year":"2022","unstructured":"Zhou, Y., Xu, Y., Du, Y., Wen, Q., He, S.: Pro-pulse: Learning progressive encoders of latent semantics in GANS for photo upsampling. IEEE TIP 31, 1230\u20131242 (2022)","journal-title":"IEEE TIP"},{"key":"1_CR43","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"592","DOI":"10.1007\/978-3-030-58520-4_35","volume-title":"Computer Vision \u2013 ECCV 2020","author":"J Zhu","year":"2020","unstructured":"Zhu, J., Shen, Y., Zhao, D., Zhou, B.: In-domain GAN inversion for real image editing. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12362, pp. 592\u2013608. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58520-4_35"},{"key":"1_CR44","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"597","DOI":"10.1007\/978-3-319-46454-1_36","volume-title":"Computer Vision \u2013 ECCV 2016","author":"J-Y Zhu","year":"2016","unstructured":"Zhu, J.-Y., Kr\u00e4henb\u00fchl, P., Shechtman, E., Efros, A.A.: Generative visual manipulation on the natural image manifold. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9909, pp. 597\u2013613. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46454-1_36"},{"key":"1_CR45","doi-asserted-by":"crossref","unstructured":"Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: ICCV, pp. 2223\u20132232 (2017)","DOI":"10.1109\/ICCV.2017.244"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19790-1_1","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,12]],"date-time":"2024-03-12T10:57:28Z","timestamp":1710241048000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19790-1_1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031197895","9783031197901"],"references-count":45,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19790-1_1","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"24 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}