{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,6]],"date-time":"2026-01-06T05:42:51Z","timestamp":1767678171236,"version":"3.48.0"},"reference-count":29,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"funder":[{"name":"Selected Projects by the Third Research Institute of the Ministry of Public Security","award":["C25853"],"award-info":[{"award-number":["C25853"]}]},{"name":"National Key Research and Development Program of China","award":["2021YFF0602104-1"],"award-info":[{"award-number":["2021YFF0602104-1"]}]},{"name":"Key Project of the Special Fund for Basic Scientific Research Business Expenses by the Third Research Institute of the Ministry of Public Security, in 2025","award":["C25352"],"award-info":[{"award-number":["C25352"]}]},{"name":"Special Youth Fund for Basic Research Business Expenses by the Third Research Institute of the Ministry of Public Security, in 2025","award":["C25354"],"award-info":[{"award-number":["C25354"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2026]]},"DOI":"10.1109\/access.2025.3648351","type":"journal-article","created":{"date-parts":[[2025,12,26]],"date-time":"2025-12-26T18:24:39Z","timestamp":1766773479000},"page":"875-885","source":"Crossref","is-referenced-by-count":0,"title":["Evolutionary-Algorithm-Based Automatic Prompt Generation for Vision-Language Model Evaluation"],"prefix":"10.1109","volume":"14","author":[{"given":"Wang","family":"Li","sequence":"first","affiliation":[{"name":"Third Research Institute of the Ministry of Public Security, Shanghai, China"}]},{"given":"Mei","family":"Hui","sequence":"additional","affiliation":[{"name":"802 Institute of Shanghai Academy of Space Flight Technology, Shanghai, China"}]},{"given":"Wang","family":"Xiaolu","sequence":"additional","affiliation":[{"name":"Third Research Institute of the Ministry of Public Security, Shanghai, China"}]},{"given":"Li","family":"Shujuan","sequence":"additional","affiliation":[{"name":"Third Research Institute of the Ministry of Public Security, Shanghai, China"}]}],"member":"263","reference":[{"key":"ref1","first-page":"23716","article-title":"Flamingo: A visual language model for few-shot learning","volume-title":"Proc. NeurIPS","author":"Alayrac"},{"key":"ref2","first-page":"4904","article-title":"Scaling up visual and vision-language representation learning with noisy text supervision","volume-title":"Proc. ICML","author":"Jia"},{"key":"ref3","first-page":"12888","article-title":"BLIP: Bootstrapping language-image pre-training for unified vision-language understanding and generation","volume-title":"Proc. ICML","author":"Li"},{"key":"ref4","first-page":"34892","article-title":"Visual instruction tuning","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Liu"},{"key":"ref5","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"Proc. ICML","author":"Radford"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19809-0_30"},{"key":"ref7","article-title":"EVA-CLIP: Improved training techniques for CLIP at scale","author":"Sun","year":"2023","journal-title":"arXiv:2303.15389"},{"key":"ref8","article-title":"MiniGPT-4: Enhancing vision-language understanding with advanced large language models","author":"Zhu","year":"2023","journal-title":"arXiv:2304.10592"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.01206"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.01438"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-022-01653-1"},{"key":"ref12","article-title":"Democratizing contrastive language-image pretraining: A CLIP benchmark of data, model, and supervision","author":"Cui","year":"2022","journal-title":"arXiv:2203.05796"},{"article-title":"FILIP: Fine-grained interactive language-image pre-training","volume-title":"Proc. ICLR","author":"Yao","key":"ref13"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW60793.2023.00034"},{"article-title":"Visual classification via description from large language models","volume-title":"Proc. ICLR","author":"Menon","key":"ref15"},{"article-title":"ChatGPT-powered hierarchical comparisons for image classification","volume-title":"Proc. NeurIPS","author":"Ren","key":"ref16"},{"key":"ref17","article-title":"GPT4 Vis: What can GPT-4 do for zero-shot visual recognition?","author":"Wu","year":"2023","journal-title":"arXiv:2311.15732"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.52202\/079017-0804"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.01443"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1145\/1216504.1216510"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.7551\/mitpress\/1090.001.0001"},{"volume-title":"An Introduction to Genetic Algorithms","year":"1998","author":"Mitchell","key":"ref22"},{"key":"ref23","article-title":"GPT-4 technical report","volume-title":"arXiv:2303.08774","author":"Achiam","year":"2023"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"ref25","article-title":"Hierarchical text-conditional image generation with CLIP latents","author":"Ramesh","year":"2022","journal-title":"arXiv:2204.06125"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.3115\/1073083.1073135"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7299087"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.91"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00371"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/6287639\/11323511\/11316124.pdf?arnumber=11316124","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,6]],"date-time":"2026-01-06T05:39:41Z","timestamp":1767677981000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11316124\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"references-count":29,"URL":"https:\/\/doi.org\/10.1109\/access.2025.3648351","relation":{},"ISSN":["2169-3536"],"issn-type":[{"type":"electronic","value":"2169-3536"}],"subject":[],"published":{"date-parts":[[2026]]}}}