{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,23]],"date-time":"2026-01-23T15:35:26Z","timestamp":1769182526840,"version":"3.49.0"},"publisher-location":"Singapore","reference-count":43,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819556816","type":"print"},{"value":"9789819556823","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-5682-3_2","type":"book-chapter","created":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T21:13:46Z","timestamp":1769116426000},"page":"17-32","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["PDAttack: Enhancing Transferability of\u00a0Unrestricted Adversarial Examples via\u00a0Prompt-Driven Diffusion"],"prefix":"10.1007","author":[{"given":"Shijie","family":"Zhao","sequence":"first","affiliation":[]},{"given":"Siyu","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Anjie","family":"Peng","sequence":"additional","affiliation":[]},{"given":"Hui","family":"Zeng","sequence":"additional","affiliation":[]},{"given":"Zhenyu","family":"Liang","sequence":"additional","affiliation":[]},{"given":"Xing","family":"Yang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,23]]},"reference":[{"key":"2_CR1","doi-asserted-by":"crossref","unstructured":"Kurakin, A., Goodfellow, I.J., Bengio, S.: Adversarial examples in the physical world. In: Artificial Intelligence Safety and Security, pp. 99\u2013112 (2018)","DOI":"10.1201\/9781351251389-8"},{"key":"2_CR2","doi-asserted-by":"crossref","unstructured":"Dong, Y., et al.: Boosting adversarial attacks with momentum. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 9185\u20139193 (2018)","DOI":"10.1109\/CVPR.2018.00957"},{"issue":"9","key":"2_CR3","doi-asserted-by":"publisher","first-page":"2805","DOI":"10.1109\/TNNLS.2018.2886017","volume":"30","author":"X Yuan","year":"2019","unstructured":"Yuan, X., He, P., Zhu, Q., Li, X.: Adversarial examples: attacks and defenses for deep learning. IEEE Trans. Neural Netw. Learn. Syst. 30(9), 2805\u20132824 (2019)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"2_CR4","unstructured":"Bhattad, A., Chong, M.J., Liang, K., Li, B., Forsyth, D.A.: Unrestricted adversarial examples via semantic manipulation. In: International Conference on Learning Representations (2019)"},{"key":"2_CR5","doi-asserted-by":"crossref","unstructured":"Dai, X., Liang, K., Xiao, B.: AdvDiff: generating unrestricted adversarial examples using diffusion models. In: European Conference on Computer Vision, pp. 93\u2013109 (2024)","DOI":"10.1007\/978-3-031-72952-2_6"},{"key":"2_CR6","doi-asserted-by":"crossref","unstructured":"Chen, J., Chen, H., Chen, K., Zhang, Y., Zou, Z., Shi, Z.: Diffusion models for imperceptible and transferable adversarial attack. IEEE Trans. Pattern Anal. Mach. Intell. (2024)","DOI":"10.1109\/TPAMI.2024.3480519"},{"key":"2_CR7","unstructured":"Yuan, S., Zhang, Q., Gao, L., Cheng, Y., Song, J.: Natural color fool: towards boosting black-box unrestricted attacks. In: Advances in Neural Information Processing Systems, vol. 35, pp. 7546\u20137560 (2022)"},{"key":"2_CR8","doi-asserted-by":"crossref","unstructured":"Mou, N., Guo, B., Zhao, L., Wang, C., Zhao, Y., Wang, Q.: No-box universal adversarial perturbations against image classifiers via artificial textures. IEEE Trans. Inf. Forensics Secur. (2024)","DOI":"10.1109\/TIFS.2024.3478828"},{"key":"2_CR9","doi-asserted-by":"crossref","unstructured":"Zhao, Z., Liu, Z., Larson, M.: Adversarial color enhancement: generating unrestricted adversarial images by optimizing a color filter. arXiv preprint arXiv:2002.01008 (2020)","DOI":"10.5244\/C.34.26"},{"key":"2_CR10","unstructured":"Zhang, Q., et al.: Beyond ImageNet attack: towards crafting adversarial examples for black-box domains. In: International Conference on Learning Representations (2022)"},{"key":"2_CR11","unstructured":"Song, J., Meng, C., Ermon, S.: Denoising diffusion implicit models. In: International Conference on Learning Representations, pp. 1\u201320 (2020)"},{"key":"2_CR12","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10684\u201310695 (2022)","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"2_CR13","doi-asserted-by":"crossref","unstructured":"Wang, W., et al.: SinDiffusion: learning a diffusion model from a single natural image. IEEE Trans. Pattern Anal. Mach. Intell. (2025)","DOI":"10.1109\/TPAMI.2025.3532956"},{"key":"2_CR14","doi-asserted-by":"crossref","unstructured":"Liu, M., et al.: LLM4GEN: leveraging semantic representation of LLMs for text-to-image generation. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 39, no. 5, pp. 5523\u20135531 (2025)","DOI":"10.1609\/aaai.v39i5.32588"},{"key":"2_CR15","unstructured":"Chen, Z., Li, B., Wu, S., Jiang, K., Ding, S., Zhang, W.: Content-based unrestricted adversarial attack. In: Advances in Neural Information Processing Systems, vol. 36, pp. 51719\u201351733 (2023)"},{"key":"2_CR16","unstructured":"Li, J., Li, D., Savarese, S., Hoi, S.: BLIP-2: bootstrapping language-image pre-training with frozen image encoders and large language models. In: International Conference on Machine Learning, pp. 19730\u201319742 (2023)"},{"key":"2_CR17","doi-asserted-by":"crossref","unstructured":"Liu, B., Wang, C., Cao, T., Jia, K., Huang, J.: Towards understanding cross and self-attention in stable diffusion for text-guided image editing. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7817\u20137826 (2024)","DOI":"10.1109\/CVPR52733.2024.00747"},{"key":"2_CR18","doi-asserted-by":"crossref","unstructured":"Zhu, R., Zhang, Z., Liang, S., Liu, Z., Xu, C.: Learning to transform dynamically for better adversarial transferability. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 24273\u201324283 (2024)","DOI":"10.1109\/CVPR52733.2024.02291"},{"key":"2_CR19","doi-asserted-by":"crossref","unstructured":"Zhang, J., et al.: Improving adversarial transferability via neuron attribution-based attacks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14993\u201315002 (2022)","DOI":"10.1109\/CVPR52688.2022.01457"},{"key":"2_CR20","doi-asserted-by":"crossref","unstructured":"Chen, B., Yin, J., Chen, S., Chen, B., Liu, X.: An adaptive model ensemble adversarial attack for boosting adversarial transferability. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4489\u20134498 (2023)","DOI":"10.1109\/ICCV51070.2023.00414"},{"key":"2_CR21","unstructured":"Kuurila-Zhang, H., Chen, H., Zhao, G.: VENOM: text-driven unrestricted adversarial example generation with diffusion models. arXiv preprint arXiv:2501.07922 (2025)"},{"key":"2_CR22","doi-asserted-by":"crossref","unstructured":"Mokady, R., Hertz, A., Aberman, K., Pritch, Y., Cohen-Or, D.: Null-text inversion for editing real images using guided diffusion models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6038\u20136047 (2023)","DOI":"10.1109\/CVPR52729.2023.00585"},{"key":"2_CR23","unstructured":"Shuai, Z., Wu, C., Tang, Z., Song, B., Shen, L.: Latent space disentanglement in diffusion transformers enables precise zero-shot semantic editing. arXiv preprint arXiv:2411.08196 (2024)"},{"key":"2_CR24","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky, O., et al.: ImageNet large scale visual recognition challenge. Int. J. Comput. Vision 115, 211\u2013252 (2015)","journal-title":"Int. J. Comput. Vision"},{"key":"2_CR25","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"2_CR26","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"2_CR27","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.-C.: MobileNetV2: inverted residuals and linear bottlenecks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4510\u20134520 (2018)","DOI":"10.1109\/CVPR.2018.00474"},{"key":"2_CR28","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Van Der Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4700\u20134708 (2017)","DOI":"10.1109\/CVPR.2017.243"},{"key":"2_CR29","unstructured":"Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. In: International Conference on Learning Representations (2021)"},{"key":"2_CR30","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"2_CR31","unstructured":"Touvron, H., Cord, M., Douze, M., Massa, F., Sablayrolles, A., J\u00e9gou, H.: Training data-efficient image transformers & distillation through attention. In: International Conference on Machine Learning, pp. 10347\u201310357 (2021)"},{"key":"2_CR32","doi-asserted-by":"crossref","unstructured":"Touvron, H., Cord, M., Sablayrolles, A., Synnaeve, G., J\u00e9gou, H.: Going deeper with image transformers. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 32\u201342 (2021)","DOI":"10.1109\/ICCV48922.2021.00010"},{"key":"2_CR33","doi-asserted-by":"crossref","unstructured":"Liao, F., Liang, M., Dong, Y., Hu, X., Zhu, J.: Defense against adversarial attacks using high-level representation guided denoiser. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1778\u20131787 (2018)","DOI":"10.1109\/CVPR.2018.00191"},{"key":"2_CR34","unstructured":"Xie, C., Wang, J., Ren, Z., Yuille, A.: Mitigating adversarial effects through randomization. In: International Conference on Learning Representations (2018)"},{"key":"2_CR35","unstructured":"Guo, C., Rana, M., Cisse, M., Van Der Maaten, L.: Countering adversarial images using input transformations. In: International Conference on Learning Representations (2018)"},{"key":"2_CR36","doi-asserted-by":"crossref","unstructured":"Xu, W., Evans, D., Qi, Y.: Feature squeezing: detecting adversarial examples in deep neural networks. In: Network and Distributed System Security Symposium (2018)","DOI":"10.14722\/ndss.2018.23198"},{"key":"2_CR37","unstructured":"Nie, W., Guo, B., Huang, Y., Xiao, C., Vahdat, A., Anandkumar, A.: Diffusion models for adversarial purification. In: International Conference on Machine Learning (2022)"},{"key":"2_CR38","unstructured":"Zhu, D., Chen, J., Shen, X., Li, X., Elhoseiny, M.: MiniGPT-4: enhancing vision-language understanding with advanced large language models. arXiv preprint arXiv:2304.10592 (2023)"},{"key":"2_CR39","unstructured":"Liu, H., Li, C., Wu, Q., Lee, Y.J.: Visual instruction tuning. In: Advances in Neural Information Processing Systems, vol. 36, pp. 34892\u201334916 (2023)"},{"key":"2_CR40","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANs trained by a two time-scale update rule converge to a local Nash equilibrium. In: Advances in Neural Information Processing Systems, vol. 30, pp. 1\u201320 (2017)"},{"issue":"4","key":"2_CR41","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004)","journal-title":"IEEE Trans. Image Process."},{"issue":"13","key":"2_CR42","doi-asserted-by":"publisher","first-page":"800","DOI":"10.1049\/el:20080522","volume":"44","author":"Q Huynh-Thu","year":"2008","unstructured":"Huynh-Thu, Q., Ghanbari, M.: Scope of validity of PSNR in image\/video quality assessment. Electron. Lett. 44(13), 800\u2013801 (2008)","journal-title":"Electron. Lett."},{"key":"2_CR43","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 586\u2013595 (2018)","DOI":"10.1109\/CVPR.2018.00068"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-5682-3_2","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T21:13:55Z","timestamp":1769116435000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-5682-3_2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9789819556816","9789819556823"],"references-count":43,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-5682-3_2","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"23 January 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shanghai","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2025.prcv.cn\/index.asp","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}