{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,1,21]],"date-time":"2025-01-21T07:10:19Z","timestamp":1737443419290,"version":"3.33.0"},"reference-count":32,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,10,6]],"date-time":"2024-10-06T00:00:00Z","timestamp":1728172800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,10,6]],"date-time":"2024-10-06T00:00:00Z","timestamp":1728172800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62176048"],"award-info":[{"award-number":["62176048"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,10,6]]},"DOI":"10.1109\/smc54092.2024.10832083","type":"proceedings-article","created":{"date-parts":[[2025,1,20]],"date-time":"2025-01-20T18:39:20Z","timestamp":1737398360000},"page":"4074-4079","source":"Crossref","is-referenced-by-count":0,"title":["SSP: A Simple and Safe Prompt Engineering for Visual Generation"],"prefix":"10.1109","author":[{"given":"Weijin","family":"Cheng","sequence":"first","affiliation":[{"name":"College of Computer Science and Engineering, University of Electronic Science and Technology of China,Chengdu"}]},{"given":"Jianzhi","family":"Liu","sequence":"additional","affiliation":[{"name":"College of Computer Science and Engineering, University of Electronic Science and Technology of China,Chengdu"}]},{"given":"Ziyun","family":"Jiao","sequence":"additional","affiliation":[{"name":"College of Computer Science and Engineering, University of Electronic Science and Technology of China,Chengdu"}]},{"given":"Jiawen","family":"Deng","sequence":"additional","affiliation":[{"name":"College of Computer Science and Engineering, University of Electronic Science and Technology of China,Chengdu"}]},{"given":"Fuji","family":"Ren","sequence":"additional","affiliation":[{"name":"College of Computer Science and Engineering, University of Electronic Science and Technology of China,Chengdu"}]}],"member":"263","reference":[{"key":"ref1","article-title":"Glide: Towards photorealistic image generation and editing with text-guided diffusion models","author":"Nichol","year":"2021","journal-title":"arXiv preprint"},{"key":"ref2","first-page":"19822","article-title":"Cogview: Mastering text-to-image generation via transformers","volume":"34","author":"Ding","year":"2021","journal-title":"Advances in Neural Information Processing Systems"},{"doi-asserted-by":"publisher","key":"ref3","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"ref4","first-page":"16890","article-title":"Cogview2: Faster and better text-to-image generation via hierarchical transformers","volume":"35","author":"Ding","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref5","first-page":"36479","article-title":"Photorealistic text-to-image diffusion models with deep language understanding","volume":"35","author":"Saharia","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref6","article-title":"Cogvideo: Large-scale pretraining for text-to-video generation via transformers","author":"Hong","year":"2022","journal-title":"arXiv preprint"},{"key":"ref7","article-title":"Latent video dif-fusion models for high-fidelity long video generation","author":"He","year":"2022","journal-title":"arXiv preprint"},{"key":"ref8","article-title":"DreamArtist: Towards Controllable One-Shot Text-to-Image Generation via Positive-Negative Prompt-Tuning","author":"Dong","year":"2022","journal-title":"arXiv preprint"},{"doi-asserted-by":"publisher","key":"ref9","DOI":"10.1162\/tacl_a_00324"},{"key":"ref10","article-title":"Making pre-trained language models better few-shot learners","author":"Gao","year":"2020","journal-title":"arXiv preprint"},{"doi-asserted-by":"publisher","key":"ref11","DOI":"10.1080\/0144929x.2023.2286532"},{"key":"ref12","article-title":"Beauti-fulprompt: Towards automatic prompt engineering for text-to-image synthesis","author":"Cao","year":"2023","journal-title":"arXiv preprint"},{"doi-asserted-by":"publisher","key":"ref13","DOI":"10.1145\/3539618.3592000"},{"doi-asserted-by":"publisher","key":"ref14","DOI":"10.1109\/TVCG.2023.3327168"},{"key":"ref15","article-title":"SurrogatePrompt: Bypassing the Safety Filter of Text- To-Image Models via Substitution","author":"Ba","year":"2023","journal-title":"arXiv preprint"},{"volume-title":"DALL.E 3","year":"2023","key":"ref16"},{"key":"ref17","article-title":"Gpt-4 technical report","author":"Achiam","year":"2023","journal-title":"arXiv preprint"},{"key":"ref18","article-title":"Bert: Pre-training of deep bidirectional transformers for language understanding","author":"Devlin","year":"2018","journal-title":"arXiv preprint"},{"key":"ref19","first-page":"1060","article-title":"Generative adversarial text to image synthesis","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Reed"},{"issue":"2","key":"ref20","first-page":"3","article-title":"Hier-archical text-conditional image generation with CLIP latents","volume":"1","author":"Ramesh","year":"2022","journal-title":"arXiv preprint"},{"key":"ref21","first-page":"36479","article-title":"Photorealistic text-to-image diffusion models with deep language understanding","volume":"35","author":"Saharia","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"doi-asserted-by":"publisher","key":"ref22","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"ref23","first-page":"8821","article-title":"Zero-shot text-to-image generation","volume-title":"International Conference on Machine Learning","author":"Ramesh"},{"doi-asserted-by":"publisher","key":"ref24","DOI":"10.1007\/978-3-319-10602-1_48"},{"doi-asserted-by":"publisher","key":"ref25","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref26","article-title":"Diffusiondb: A large-scale prompt gallery dataset for text-to-image generative models","author":"Wang","year":"2022","journal-title":"arXiv preprint"},{"key":"ref27","article-title":"Gans trained by a two time-scale update rule converge to a local nash equilibrium","volume":"30","author":"Heusel","year":"2017","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref28","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"International Conference on Machine Learning","author":"Radford"},{"volume-title":"Detoxify","year":"2020","author":"Hanu","key":"ref29"},{"volume-title":"Improved Aesthetic Predictor","author":"Schuhmann","key":"ref30"},{"volume-title":"ChatGPT","year":"2023","key":"ref31"},{"volume-title":"MagicPrompt-Stable Diffusion","key":"ref32"}],"event":{"name":"2024 IEEE International Conference on Systems, Man, and Cybernetics (SMC)","start":{"date-parts":[[2024,10,6]]},"location":"Kuching, Malaysia","end":{"date-parts":[[2024,10,10]]}},"container-title":["2024 IEEE International Conference on Systems, Man, and Cybernetics (SMC)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10830919\/10830920\/10832083.pdf?arnumber=10832083","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,21]],"date-time":"2025-01-21T06:33:36Z","timestamp":1737441216000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10832083\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,6]]},"references-count":32,"URL":"https:\/\/doi.org\/10.1109\/smc54092.2024.10832083","relation":{},"subject":[],"published":{"date-parts":[[2024,10,6]]}}}