{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,29]],"date-time":"2026-03-29T20:23:15Z","timestamp":1774815795035,"version":"3.50.1"},"reference-count":70,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"10","license":[{"start":{"date-parts":[[2025,10,1]],"date-time":"2025-10-01T00:00:00Z","timestamp":1759276800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2025,10,1]],"date-time":"2025-10-01T00:00:00Z","timestamp":1759276800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,10,1]],"date-time":"2025-10-01T00:00:00Z","timestamp":1759276800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001381","name":"National Research Foundation Singapore","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001381","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001381","name":"National Research Foundation Singapore","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001381","id-type":"DOI","asserted-by":"publisher"}]},{"name":"National Cybersecurity R\\&#x0026;D Programme"},{"name":"National Research Foundation, Prime Minister&#x2019;s Office, Singapore"},{"name":"Research Excellence and Technological Enterprise (CREATE) programme"},{"DOI":"10.13039\/501100001381","name":"National Research Foundation Singapore","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001381","id-type":"DOI","asserted-by":"publisher"}]},{"name":"National Large Language Models Funding Initiative"},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Ningbo Science and Technology Innovation 2025 Major Project"},{"name":"Shenzhen Science and Technology Program"},{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"publisher"}]},{"name":"National Research Foundation, Singapore and Cyber Security Agency of Singapore"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Pattern Anal. Mach. Intell."],"published-print":{"date-parts":[[2025,10]]},"DOI":"10.1109\/tpami.2025.3581476","type":"journal-article","created":{"date-parts":[[2025,6,20]],"date-time":"2025-06-20T13:27:53Z","timestamp":1750426073000},"page":"8489-8505","source":"Crossref","is-referenced-by-count":2,"title":["Semantic-Aligned Adversarial Evolution Triangle for High-Transferability Vision-Language Attack"],"prefix":"10.1109","volume":"47","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2018-9344","authenticated-orcid":false,"given":"Xiaojun","family":"Jia","sequence":"first","affiliation":[{"name":"Nanyang Technological University, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-2282-491X","authenticated-orcid":false,"given":"Sensen","family":"Gao","sequence":"additional","affiliation":[{"name":"Department of Computer Vision, Mohamed Bin Zayed University of Artificial Intelligence, Masdar City, United Arab Emirates"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0974-9299","authenticated-orcid":false,"given":"Qing","family":"Guo","sequence":"additional","affiliation":[{"name":"VCIP, School of Computer Science, Nankai University, Tianjin, China"}]},{"given":"Simeng","family":"Qin","sequence":"additional","affiliation":[{"name":"Northeastern University, Shenyang, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4178-0907","authenticated-orcid":false,"given":"Ke","family":"Ma","sequence":"additional","affiliation":[{"name":"School of Electronic, Electrical and Communication Engineering, University of Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5784-770X","authenticated-orcid":false,"given":"Yihao","family":"Huang","sequence":"additional","affiliation":[{"name":"Nanyang Technological University, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7300-9215","authenticated-orcid":false,"given":"Yang","family":"Liu","sequence":"additional","affiliation":[{"name":"Nanyang Technological University, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8095-4637","authenticated-orcid":false,"given":"Ivor W.","family":"Tsang","sequence":"additional","affiliation":[{"name":"Centre for Frontier AI Research (CFAR) and Institute of High Performance Computing(IHPC), Agency for Science, Technology and Research(A<sup>&#x002A;<\/sup>STAR), Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7141-708X","authenticated-orcid":false,"given":"Xiaochun","family":"Cao","sequence":"additional","affiliation":[{"name":"School of Cyber Science and Technology, Shenzhen Campus of Sun Yat-sen University, Shenzhen, China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00130"},{"key":"ref2","article-title":"Gpt-4 technical report","author":"Achiam","year":"2023"},{"key":"ref3","first-page":"23716","article-title":"Flamingo: A visual language model for few-shot learning","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"35","author":"Alayrac","year":"2022"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46454-1_24"},{"key":"ref5","article-title":"Claude 3 Haiku: Our fastest model yet","year":"2024","journal-title":""},{"issue":"2","key":"ref6","first-page":"3","article-title":"Qwen-vl: A frontier large vision-language model with versatile abilities","volume":"1","author":"Bai","year":"2023"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58595-2_7"},{"key":"ref8","first-page":"65","article-title":"METEOR: An automatic metric for mt evaluation with improved correlation with human judgments","volume-title":"Proc. ACL Workshop Intrinsic Extrinsic Eval. Measures Mach. Transl. Summarization","author":"Banerjee","year":"2005"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.01267"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58577-8_7"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00512"},{"key":"ref12","article-title":"Towards transferable attacks against vision-llms in autonomous driving with typography","author":"Chung","year":"2024"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00444"},{"key":"ref14","article-title":"An image is worth 16x16 words: Transformers for image recognition at scale","author":"Dosovitskiy","year":"2010"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01763"},{"key":"ref16","article-title":"Improving adversarial transferability of visual-language pre-training models through collaborative multimodal interaction","author":"Fu","year":"2024"},{"key":"ref17","article-title":"Adversarial robustness for visual grounding of multimodal large language models","author":"Gao","year":"2024"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-72998-0_25"},{"key":"ref19","article-title":"A survey on transferability of adversarial examples across deep neural networks","author":"Gu","year":"2023"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01066"},{"key":"ref21","article-title":"Sa-attack: Improving adversarial transferability of vision-language pre-training models via self-augmentation","author":"He","year":"2023"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i1.25154"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/cvpr52688.2022.01745"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/WACV57701.2024.00297"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1145\/3581783.3611914"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/tpami.2024.3381180\/mm1"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1145\/3474085.3475692"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.1980.1102314"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1145\/3474085.3475431"},{"issue":"PMLR","key":"ref31","first-page":"19730","article-title":"Blip-2: Bootstrapping language-image pre-training with frozen image encoders and large language models","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Li","year":"2023"},{"issue":"PMLR","key":"ref32","first-page":"12888","article-title":"BLIP: Bootstrapping language-image pre-training for unified vision-language understanding and generation","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Li","year":"2022"},{"key":"ref33","first-page":"9694","article-title":"Align before fuse: Vision and language representation learning with momentum distillation","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"34","author":"Li","year":"2021"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.02304"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.500"},{"key":"ref36","first-page":"32900","article-title":"Improving adversarial transferability via intermediate-level perturbation decay","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"36","author":"Li","year":"2024"},{"issue":"PMLR","key":"ref37","first-page":"6577","article-title":"Uncovering the connections between adversarial transferability and knowledge transferability","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Liang","year":"2021"},{"key":"ref38","first-page":"17612","article-title":"Mind the gap: Understanding the modality gap in multi-modal contrastive representation learning","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"35","author":"Liang","year":"2022"},{"key":"ref39","first-page":"74","article-title":"Rouge: A package for automatic evaluation of summaries","volume-title":"Proc. Text Summarization Branches Out","author":"Lin","year":"2004"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"ref41","first-page":"34892","article-title":"Visual instruction tuning","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"36","author":"Liu","year":"2023"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00016"},{"key":"ref43","article-title":"Towards deep learning models resistant to adversarial attacks","author":"Madry","year":"2017"},{"key":"ref44","article-title":"Understanding zero-shot adversarial robustness for large-scale models","author":"Mao","year":"2022"},{"key":"ref45","first-page":"23296","article-title":"Intriguing properties of vision transformers","volume":"34","author":"Naseer","year":"2021","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/WACV57701.2024.00394"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/iccv.2015.303"},{"key":"ref48","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Radford","year":"2021"},{"key":"ref49","article-title":"Robust clip: Unsupervised adversarial fine-tuning of vision embeddings for robust large vision-language models","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Schlarmann","year":"2024"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.74"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1515\/9781400881970-018"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1145\/3474085.3475637"},{"key":"ref53","first-page":"10347","article-title":"Training data-efficient image transformers & distillation through attention","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Touvron","year":"2021"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7299087"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.02218"},{"key":"ref56","first-page":"22680","article-title":"Vlmixer: Unpaired vision-language pre-training via cross-modal cutmix","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Wang","year":"2022"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01585"},{"key":"ref58","article-title":"A unified approach to interpreting and boosting adversarial transferability","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Wang","year":"2021"},{"key":"ref59","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2023.3331582"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01522"},{"key":"ref61","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46475-6_5"},{"key":"ref62","doi-asserted-by":"publisher","DOI":"10.1109\/satml64287.2025.00049"},{"key":"ref63","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00060"},{"key":"ref64","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2024.3367773"},{"key":"ref65","doi-asserted-by":"publisher","DOI":"10.1145\/3503161.3547801"},{"key":"ref66","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00790"},{"key":"ref67","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00553"},{"key":"ref68","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00359"},{"key":"ref69","doi-asserted-by":"publisher","DOI":"10.1145\/3588432.3591532"},{"key":"ref70","article-title":"Minigpt-4: Enhancing vision-language understanding with advanced large language models","author":"Zhu","year":"2023"}],"container-title":["IEEE Transactions on Pattern Analysis and Machine Intelligence"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/34\/11163533\/11045302.pdf?arnumber=11045302","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,18]],"date-time":"2025-09-18T20:04:50Z","timestamp":1758225890000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11045302\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10]]},"references-count":70,"journal-issue":{"issue":"10"},"URL":"https:\/\/doi.org\/10.1109\/tpami.2025.3581476","relation":{},"ISSN":["0162-8828","2160-9292","1939-3539"],"issn-type":[{"value":"0162-8828","type":"print"},{"value":"2160-9292","type":"electronic"},{"value":"1939-3539","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,10]]}}}