{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,6]],"date-time":"2026-02-06T04:47:55Z","timestamp":1770353275593,"version":"3.49.0"},"publisher-location":"Cham","reference-count":30,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031726422","type":"print"},{"value":"9783031726439","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,11,22]],"date-time":"2024-11-22T00:00:00Z","timestamp":1732233600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,22]],"date-time":"2024-11-22T00:00:00Z","timestamp":1732233600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-72643-9_14","type":"book-chapter","created":{"date-parts":[[2024,11,21]],"date-time":"2024-11-21T20:48:23Z","timestamp":1732222103000},"page":"231-246","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["RegionDrag: Fast Region-Based Image Editing with\u00a0Diffusion Models"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0005-1373-231X","authenticated-orcid":false,"given":"Jingyi","family":"Lu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3797-5082","authenticated-orcid":false,"given":"Xinghui","family":"Li","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7995-9999","authenticated-orcid":false,"given":"Kai","family":"Han","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,22]]},"reference":[{"key":"14_CR1","doi-asserted-by":"crossref","unstructured":"Brooks, T., Holynski, A., Efros, A.A.: InstructPix2Pix: learning to follow image editing instructions. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.01764"},{"key":"14_CR2","doi-asserted-by":"crossref","unstructured":"Cao, M., Wang, X., Qi, Z., Shan, Y., Qie, X., Zheng, Y.: MasaCtrl: tuning-free mutual self-attention control for consistent image synthesis and editing. arXiv preprint arXiv:2304.08465 (2023)","DOI":"10.1109\/ICCV51070.2023.02062"},{"issue":"1","key":"14_CR3","doi-asserted-by":"publisher","first-page":"53","DOI":"10.1109\/MSP.2017.2765202","volume":"35","author":"A Creswell","year":"2018","unstructured":"Creswell, A., White, T., Dumoulin, V., Arulkumaran, K., Sengupta, B., Bharath, A.A.: Generative adversarial networks: an overview. IEEE Signal Process. Mag. 35(1), 53\u201365 (2018). https:\/\/doi.org\/10.1109\/MSP.2017.2765202","journal-title":"IEEE Signal Process. Mag."},{"key":"14_CR4","doi-asserted-by":"crossref","unstructured":"Endo, Y.: User-controllable latent transformer for styleGan image layout editing. In: CGF (2022)","DOI":"10.1111\/cgf.14686"},{"key":"14_CR5","unstructured":"Epstein, D., Jabri, A., Poole, B., Efros, A., Holynski, A.: Diffusion self-guidance for controllable image generation. In: NeurIPS (2024)"},{"key":"14_CR6","unstructured":"Goodfellow, I., et al: Generative adversarial nets. In: NeurIPS (2014)"},{"key":"14_CR7","unstructured":"Ho, J., Jain, A., Abbeel, P.: Denoising diffusion probabilistic models. In: NeurIPS (2020)"},{"key":"14_CR8","unstructured":"Hu, E.J., et al.: LoRA: low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685 (2021)"},{"key":"14_CR9","doi-asserted-by":"crossref","unstructured":"Kang, M., et al.: Scaling up GANs for text-to-image synthesis. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.00976"},{"key":"14_CR10","unstructured":"Karras, T., et al.: Alias-free generative adversarial networks. In: NeurIPS (2021)"},{"key":"14_CR11","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of StyleGan. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"14_CR12","doi-asserted-by":"crossref","unstructured":"Kawar, B., et al: Imagic: text-based real image editing with diffusion models. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.00582"},{"key":"14_CR13","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. In: NeurIPS (2012)"},{"key":"14_CR14","unstructured":"Ling, P., Chen, L., Zhang, P., Chen, H., Jin, Y.: FreeDrag: point tracking is not you need for interactive point-based image editing. arXiv preprint arXiv:2307.04684 (2023)"},{"key":"14_CR15","unstructured":"Luo, M., Cheng, W., Yang, J.: RotationDrag: point-based image editing with rotated diffusion features. arXiv preprint arXiv:2401.06442 (2024)"},{"key":"14_CR16","unstructured":"Mou, C., Wang, X., Song, J., Shan, Y., Zhang, J.: DragonDiffusion: enabling drag-style manipulation on diffusion models. arXiv preprint arXiv:2307.02421 (2023)"},{"key":"14_CR17","doi-asserted-by":"crossref","unstructured":"Mou, C., Wang, X., Song, J., Shan, Y., Zhang, J.: DiffEditor: boosting accuracy and flexibility on diffusion-based image editing. arXiv preprint arXiv:2402.02583 (2024)","DOI":"10.1109\/CVPR52733.2024.00811"},{"key":"14_CR18","unstructured":"Nie, S., Guo, H.A., Lu, C., Zhou, Y., Zheng, C., Li, C.: The blessing of randomness: SDE beats ode in general diffusion-based image editing. arXiv preprint arXiv:2311.01410 (2023)"},{"key":"14_CR19","doi-asserted-by":"crossref","unstructured":"Pan, X., Tewari, A., Leimk\u00fchler, T., Liu, L., Meka, A., Theobalt, C.: Drag your GAN: interactive point-based manipulation on the generative image manifold. In: ACM SIGGRAPH (2023)","DOI":"10.1145\/3588432.3591500"},{"key":"14_CR20","unstructured":"Paszke, A., et\u00a0al.: PyTorch: an imperative style, high-performance deep learning library. In: NeurIPS (2019)"},{"key":"14_CR21","doi-asserted-by":"crossref","unstructured":"Patashnik, O., Wu, Z., Shechtman, E., Cohen-Or, D., Lischinski, D.: StyleCLIP: text-driven manipulation of StyleGan imagery. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00209"},{"key":"14_CR22","unstructured":"von Platen, P., et al.: Diffusers: state-of-the-art diffusion models. https:\/\/github.com\/huggingface\/diffusers (2022)"},{"key":"14_CR23","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"14_CR24","doi-asserted-by":"crossref","unstructured":"Shi, Y., Xue, C., Pan, J., Zhang, W., Tan, V.Y., Bai, S.: DragDiffusion: harnessing diffusion models for interactive point-based image editing. arXiv preprint arXiv:2306.14435 (2023)","DOI":"10.1109\/CVPR52733.2024.00844"},{"key":"14_CR25","unstructured":"Song, J., Meng, C., Ermon, S.: Denoising diffusion implicit models. arXiv preprint arXiv:2010.02502 (2020)"},{"key":"14_CR26","unstructured":"Tang, L., Jia, M., Wang, Q., Phoo, C.P., Hariharan, B.: Emergent correspondence from image diffusion. In: NeurIPS (2024)"},{"key":"14_CR27","doi-asserted-by":"crossref","unstructured":"Wang, T., Zhang, Y., Fan, Y., Wang, J., Chen, Q.: High-fidelity GAN inversion for image attribute editing. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01109"},{"key":"14_CR28","unstructured":"Wu, C.H., De\u00a0la Torre, F.: Unifying diffusion models\u2019 latent space, with applications to CycleDiffusion and guidance. arXiv preprint arXiv:2210.05559 (2022)"},{"key":"14_CR29","doi-asserted-by":"crossref","unstructured":"Xia, W., Zhang, Y., Yang, Y., Xue, J.H., Zhou, B., Yang, M.H.: GAN inversion: a survey. IEEE TPAMI (2022)","DOI":"10.1109\/TPAMI.2022.3181070"},{"key":"14_CR30","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00068"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-72643-9_14","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,21]],"date-time":"2024-11-21T21:26:23Z","timestamp":1732224383000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-72643-9_14"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,22]]},"ISBN":["9783031726422","9783031726439"],"references-count":30,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-72643-9_14","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,22]]},"assertion":[{"value":"22 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}