{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,11]],"date-time":"2026-02-11T21:45:32Z","timestamp":1770846332728,"version":"3.50.1"},"reference-count":31,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,10,5]],"date-time":"2025-10-05T00:00:00Z","timestamp":1759622400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,10,5]],"date-time":"2025-10-05T00:00:00Z","timestamp":1759622400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,10,5]]},"DOI":"10.1109\/smc58881.2025.11342966","type":"proceedings-article","created":{"date-parts":[[2026,1,28]],"date-time":"2026-01-28T20:54:44Z","timestamp":1769633684000},"page":"193-199","source":"Crossref","is-referenced-by-count":0,"title":["KGSS:Knowledge-guided Sample Selection for Prompt Generation in Large Language Model"],"prefix":"10.1109","author":[{"given":"Yan","family":"Chen","sequence":"first","affiliation":[{"name":"Beijing University of Technology,College of Information Science and Technology"}]},{"given":"Guang","family":"Yang","sequence":"additional","affiliation":[{"name":"Beijing University of Technology,College of Information Science and Technology"}]},{"given":"Yutao","family":"Zhu","sequence":"additional","affiliation":[{"name":"Renmin University of China,Gaoling School of Artificial Intelligence"}]},{"given":"Zhicheng","family":"Dou","sequence":"additional","affiliation":[{"name":"Renmin University of China,Gaoling School of Artificial Intelligence"}]},{"given":"Lifang","family":"Wu","sequence":"additional","affiliation":[{"name":"Beijing University of Technology,College of Information Science and Technology"}]}],"member":"263","reference":[{"key":"ref1","first-page":"1877","article-title":"Language models are few-shot learners","volume":"33","author":"Brown","year":"2020","journal-title":"Advances in neural information processing systems"},{"issue":"240","key":"ref2","first-page":"1","article-title":"Palm: Scaling language modeling with pathways","volume-title":"Journal of Machine Learning Research","volume":"24","author":"Chowdhery","year":"2023"},{"key":"ref3","article-title":"Gpt-4 technical report","year":"2024"},{"key":"ref4","article-title":"Llama: Open and efficient foundation language models","author":"Touvron","year":"2023"},{"key":"ref5","first-page":"100","article-title":"What makes good in-context examples for GPT-3?","volume-title":"Proceedings of Deep Learning Inside Out (DeeLIO 2022): The 3rd Workshop on Knowledge Extraction and Integration for Deep Learning Architectures","author":"Liu"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.831"},{"key":"ref7","doi-asserted-by":"crossref","first-page":"4469","DOI":"10.18653\/v1\/2023.findings-acl.273","article-title":"Complementary explanations for effective in-context learning","volume-title":"Findings of the Association for Computational Linguistics: ACL 2023","author":"Ye","year":"2023"},{"key":"ref8","first-page":"12 697","article-title":"Calibrate before use: Improving few-shot performance of language models","volume-title":"Proceedings of the 38th International Conference on Machine Learning","volume":"139","author":"Zhao"},{"key":"ref9","first-page":"3558","article-title":"Improving in-context few-shot learning via self-supervised training","volume-title":"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies","author":"Chen"},{"key":"ref10","first-page":"2791","article-title":"MetaICL: Learning to learn in context","volume-title":"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies","author":"Min"},{"key":"ref11","first-page":"968","article-title":"Symbol tuning improves in-context learning in language models","volume-title":"Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing","author":"Wei"},{"key":"ref12","doi-asserted-by":"crossref","DOI":"10.18653\/v1\/2023.acl-long.78","article-title":"Diverse demonstrations improve in-context compositional generalization","author":"Levy","year":"2023"},{"key":"ref13","doi-asserted-by":"crossref","DOI":"10.18653\/v1\/2023.acl-long.79","article-title":"Self-adaptive in-context learning: An information compression perspective for in-context example selection and ordering","author":"Wu","year":"2023"},{"key":"ref14","article-title":"In-context example selection with influences","author":"Nguyen","year":"2023"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-emnlp.411"},{"key":"ref16","article-title":"Fairness-guided few-shot prompting for large language models","author":"Ma","year":"2023"},{"key":"ref17","article-title":"Uncertainty decomposition and quantification for in-context learning of large language models","author":"Ling","year":"2024","journal-title":"CoRR"},{"key":"ref18","article-title":"Misconfidence-based demonstration selection for llm in-context learning","author":"Xu","year":"2024"},{"key":"ref19","first-page":"8086","article-title":"Fantastically ordered prompts and where to find them: Overcoming few-shot prompt order sensitivity","volume-title":"Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"Lu"},{"key":"ref20","article-title":"Let\u2019s learn step by step: Enhancing in-context learning ability with curriculum learning","author":"Liu","year":"2024"},{"key":"ref21","article-title":"Chain-of-thought prompting elicits reasoning in large language models","author":"Wei","year":"2023"},{"key":"ref22","article-title":"Large language models are zero-shot reasoners","volume-title":"Advances in Neural Information Processing Systems","author":"Kojima","year":"2022"},{"key":"ref23","article-title":"Large language models can self-improve","author":"Huang","year":"2022"},{"key":"ref24","article-title":"Automatic chain of thought prompting in large language models","volume-title":"The Eleventh International Conference on Learning Representations","author":"Zhang"},{"key":"ref25","article-title":"Complexity-based prompting for multi-step reasoning","volume-title":"The Eleventh International Conference on Learning Representations","author":"Fu"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i16.29720"},{"key":"ref27","article-title":"Self-consistency improves chain of thought reasoning in language models","volume-title":"The Eleventh International Conference on Learning Representations","author":"Wang"},{"key":"ref28","article-title":"Least-to-most prompting enables complex reasoning in large language models","volume-title":"The Eleventh International Conference on Learning Representations","author":"Zhou"},{"key":"ref29","article-title":"Compositional semantic parsing with large language models","volume-title":"The Eleventh International Conference on Learning Representations","author":"Drozdov"},{"key":"ref30","first-page":"4171","article-title":"BERT: Pre-training of deep bidirectional transformers for language understanding","volume-title":"Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)","author":"Devlin"},{"key":"ref31","article-title":"Evaluating the performance of large language models on gaokao benchmark","author":"Zhang","year":"2023"}],"event":{"name":"2025 IEEE International Conference on Systems, Man, and Cybernetics (SMC)","location":"Vienna, Austria","start":{"date-parts":[[2025,10,5]]},"end":{"date-parts":[[2025,10,8]]}},"container-title":["2025 IEEE International Conference on Systems, Man, and Cybernetics (SMC)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11342430\/11342431\/11342966.pdf?arnumber=11342966","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,11]],"date-time":"2026-02-11T20:53:12Z","timestamp":1770843192000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11342966\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,5]]},"references-count":31,"URL":"https:\/\/doi.org\/10.1109\/smc58881.2025.11342966","relation":{},"subject":[],"published":{"date-parts":[[2025,10,5]]}}}