{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,26]],"date-time":"2025-06-26T04:02:37Z","timestamp":1750910557484,"version":"3.41.0"},"reference-count":40,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,6,8]],"date-time":"2025-06-08T00:00:00Z","timestamp":1749340800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,6,8]],"date-time":"2025-06-08T00:00:00Z","timestamp":1749340800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,6,8]]},"DOI":"10.1109\/cec65147.2025.11043024","type":"proceedings-article","created":{"date-parts":[[2025,6,24]],"date-time":"2025-06-24T17:32:32Z","timestamp":1750786352000},"page":"1-8","source":"Crossref","is-referenced-by-count":0,"title":["Diverse Prompts: Illuminating the Prompt Space of Large Language Models with MAP-Elites"],"prefix":"10.1109","author":[{"given":"Gabriel Machado","family":"Santos","sequence":"first","affiliation":[{"name":"Federal University of Uberl&#x00E2;ndia,Computer Science Faculty,S&#x00E3;o Paulo,Brazil"}]},{"given":"Rita","family":"Maria Da Silva Julia","sequence":"additional","affiliation":[{"name":"Federal University of Uberl&#x00E2;ndia,Computer Science Faculty,S&#x00E3;o Paulo,Brazil"}]},{"given":"Marcelo Zanchetta","family":"Do Nascimento","sequence":"additional","affiliation":[{"name":"Federal University of Uberl&#x00E2;ndia,Computer Science Faculty,S&#x00E3;o Paulo,Brazil"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2024.3365742"},{"article-title":"The prompt report: A systematic survey of prompting techniques","year":"2024","author":"Schulhoff","key":"ref2"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/TSE.2023.3313881"},{"key":"ref4","article-title":"Hard prompts made easy: Gradient-based discrete optimization for prompt tuning and discovery","volume":"36","author":"Wen","year":"2024","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref5","article-title":"Prompt engineering for chatgpt: a quick guide to techniques, tips, and best practices","author":"Ekin","year":"2023","journal-title":"Authorea Preprints"},{"article-title":"Illuminating search spaces by mapping elites","year":"2015","author":"Mouret","key":"ref6"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1145\/3638529.3654049"},{"year":"2023","key":"ref8","article-title":"Big-bench lite"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1310.4546"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/D14-1162"},{"article-title":"Bert: Pre-training of deep bidirectional transformers for language understanding","year":"2018","author":"Devlin","key":"ref11"},{"article-title":"Improving language understanding by generative pre-training","year":"2018","author":"Radford","key":"ref12"},{"key":"ref13","article-title":"Attention is all you need","author":"Vaswani","year":"2017","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref14","first-page":"1877","article-title":"Language models are few-shot learners","volume":"33","author":"Brown","year":"2020","journal-title":"Advances in neural information processing systems"},{"article-title":"Llama: Open and efficient foundation language models","year":"2023","author":"Touvron","key":"ref15"},{"article-title":"Megatron-lm: Training multi-billion parameter language models using model parallelism","year":"2019","author":"Shoeybi","key":"ref16"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1145\/3560815"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.759"},{"key":"ref19","first-page":"24 824","article-title":"Chain-of-thought prompting elicits reasoning in large language models","volume":"35","author":"Wei","year":"2022","journal-title":"Advances in neural information processing systems"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D18-1298"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.eacl-main.20"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1145\/3641289"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.3389\/frobt.2016.00040"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1038\/nature14422"},{"article-title":"Deep curiosity search: Intra-life exploration can improve performance on challenging deep reinforcement learning problems","year":"2018","author":"Stanton","key":"ref25"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1162\/EVCO_a_00189"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acl-long.353"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.346"},{"issue":"140","key":"ref29","first-page":"1","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume":"21","author":"Raffel","year":"2020","journal-title":"Journal of machine learning research"},{"key":"ref30","article-title":"Grammar prompting for domain-specific language generation with large language models","volume":"36","author":"Wang","year":"2024","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.21236\/ad0616323"},{"article-title":"Beyond the imitation game: Quantifying and extrapolating the capabilities of language models","year":"2022","author":"Srivastava","key":"ref32"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-acl.824"},{"article-title":"Hugging face inference endpoints","year":"2024","author":"Face","key":"ref34"},{"article-title":"Starling-lm-7b-alpha","year":"2024","author":"Face","key":"ref35"},{"article-title":"Llama-3.1-8b-instruct","year":"2024","author":"Face","key":"ref36"},{"article-title":"Phi-3.5-mini-instruct","year":"2024","author":"Face","key":"ref37"},{"article-title":"Qwen2.5-7b-instruct","year":"2024","author":"Face","key":"ref38"},{"key":"ref39","article-title":"Schema-learning and rebinding as mechanisms of in-context learning and emergence","volume":"36","author":"Swaminathan","year":"2024","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref40","first-page":"22 199","article-title":"Large language models are zero-shot reasoners","volume":"35","author":"Kojima","year":"2022","journal-title":"Advances in neural information processing systems"}],"event":{"name":"2025 IEEE Congress on Evolutionary Computation (CEC)","start":{"date-parts":[[2025,6,8]]},"location":"Hangzhou, China","end":{"date-parts":[[2025,6,12]]}},"container-title":["2025 IEEE Congress on Evolutionary Computation (CEC)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11042929\/11042912\/11043024.pdf?arnumber=11043024","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,25]],"date-time":"2025-06-25T05:27:01Z","timestamp":1750829221000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11043024\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,8]]},"references-count":40,"URL":"https:\/\/doi.org\/10.1109\/cec65147.2025.11043024","relation":{},"subject":[],"published":{"date-parts":[[2025,6,8]]}}}