{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,6]],"date-time":"2026-05-06T02:44:23Z","timestamp":1778035463222,"version":"3.51.4"},"publisher-location":"Singapore","reference-count":28,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819533428","type":"print"},{"value":"9789819533435","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,11,23]],"date-time":"2025-11-23T00:00:00Z","timestamp":1763856000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,11,23]],"date-time":"2025-11-23T00:00:00Z","timestamp":1763856000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-3343-5_33","type":"book-chapter","created":{"date-parts":[[2025,11,22]],"date-time":"2025-11-22T06:30:09Z","timestamp":1763793009000},"page":"427-438","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Self-Prompt Tuning: Enable Autonomous Role-Playing in\u00a0LLMs"],"prefix":"10.1007","author":[{"given":"Aobo","family":"Kong","sequence":"first","affiliation":[]},{"given":"Shiwan","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Hao","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Qicheng","family":"Li","sequence":"additional","affiliation":[]},{"given":"Yong","family":"Qin","sequence":"additional","affiliation":[]},{"given":"Jiaming","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Haoqin","family":"Sun","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,11,23]]},"reference":[{"key":"33_CR1","unstructured":"Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan: Language models are few-shot learners. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol.\u00a033, pp. 1877\u20131901. Curran Associates, Inc. (2020). https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2020\/file\/1457c0d6bfcb4967418bfb8ac142f64a-Paper.pdf"},{"key":"33_CR2","unstructured":"Chen, M., et al.: Evaluating large language models trained on code (2021)"},{"key":"33_CR3","unstructured":"Cobbe, K., et al.: Training verifiers to solve math word problems (2021)"},{"key":"33_CR4","unstructured":"Dubois, Y., et al.: Alpacafarm: a simulation framework for methods that learn from human feedback (2023)"},{"key":"33_CR5","doi-asserted-by":"publisher","first-page":"346","DOI":"10.1162\/tacl_a_00370","volume":"9","author":"M Geva","year":"2021","unstructured":"Geva, M., Khashabi, D., Segal, E., Khot, T., Roth, D., Berant, J.: Did aristotle use a laptop? A question answering benchmark with implicit reasoning strategies. Trans. Assoc. Comput. Linguist. 9, 346\u2013361 (2021). https:\/\/doi.org\/10.1162\/tacl_a_00370","journal-title":"Trans. Assoc. Comput. Linguist."},{"key":"33_CR6","unstructured":"Gudibande, A., et al.: The false promise of imitating proprietary language models. In: The Twelfth International Conference on Learning Representations (2024). https:\/\/openreview.net\/forum?id=Kz3yckpCN5"},{"key":"33_CR7","unstructured":"Hendrycks, D., et al.: Measuring massive multitask language understanding. In: International Conference on Learning Representations (2021). https:\/\/openreview.net\/forum?id=d7KBjmI3GmQ"},{"key":"33_CR8","unstructured":"Jiang, A.Q., et\u00a0al.: Mistral 7b. arXiv preprint arXiv:2310.06825 (2023)"},{"key":"33_CR9","unstructured":"Kojima, T., Gu, S.S., Reid, M., Matsuo, Y., Iwasawa, Y.: Large language models are zero-shot reasoners. In: Koyejo, S., Mohamed, S., Agarwal, A., Belgrave, D., Cho, K., Oh, A. (eds.) Advances in Neural Information Processing Systems, vol.\u00a035, pp. 22199\u201322213. Curran Associates, Inc. (2022). https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2022\/file\/8bb0d291acd4acf06ef112099c16f326-Paper-Conference.pdf"},{"key":"33_CR10","doi-asserted-by":"publisher","unstructured":"Kong, A., et al.: Better zero-shot reasoning with role-play prompting. In: Duh, K., Gomez, H., Bethard, S. (eds.) NAACL: HLT (Volume 1: Long Papers), pp. 4099\u20134113. Association for Computational Linguistics, Mexico City (2024). https:\/\/doi.org\/10.18653\/v1\/2024.naacl-long.228","DOI":"10.18653\/v1\/2024.naacl-long.228"},{"key":"33_CR11","doi-asserted-by":"publisher","unstructured":"Lin, S., Hilton, J., Evans, O.: TruthfulQA: measuring how models mimic human falsehoods. In: Muresan, S., Nakov, P., Villavicencio, A. (eds.) Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 3214\u20133252. Association for Computational Linguistics, Dublin (2022). https:\/\/doi.org\/10.18653\/v1\/2022.acl-long.229","DOI":"10.18653\/v1\/2022.acl-long.229"},{"key":"33_CR12","doi-asserted-by":"publisher","unstructured":"Mihaylov, T., Clark, P., Khot, T., Sabharwal, A.: Can a suit of armor conduct electricity? A new dataset for open book question answering. In: Riloff, E., Chiang, D., Hockenmaier, J., Tsujii, J. (eds.) Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 2381\u20132391. Association for Computational Linguistics, Brussels (2018). https:\/\/doi.org\/10.18653\/v1\/D18-1260, https:\/\/aclanthology.org\/D18-1260","DOI":"10.18653\/v1\/D18-1260"},{"key":"33_CR13","unstructured":"Ouyang, L., et al.: Training language models to follow instructions with human feedback. In: Koyejo, S., Mohamed, S., Agarwal, A., Belgrave, D., Cho, K., Oh, A. (eds.) Advances in Neural Information Processing Systems, vol.\u00a035, pp. 27730\u201327744. Curran Associates, Inc. (2022)"},{"key":"33_CR14","unstructured":"Salewski, L., Alaniz, S., Rio-Torto, I., Schulz, E., Akata, Z.: In-context impersonation reveals large language models\u2019 strengths and biases. In: Thirty-seventh Conference on Neural Information Processing Systems (2023). https:\/\/openreview.net\/forum?id=CbsJ53LdKc"},{"issue":"7987","key":"33_CR15","doi-asserted-by":"publisher","first-page":"493","DOI":"10.1038\/s41586-023-06647-8","volume":"623","author":"M Shanahan","year":"2023","unstructured":"Shanahan, M., McDonell, K., Reynolds, L.: Role play with large language models. Nature 623(7987), 493\u2013498 (2023)","journal-title":"Nature"},{"key":"33_CR16","unstructured":"Srivastava, A., et al.: Beyond the imitation game: quantifying and extrapolating the capabilities of language models. Trans. Mach. Learn. Res. (2023). https:\/\/openreview.net\/forum?id=uyTL5Bvosj"},{"key":"33_CR17","doi-asserted-by":"publisher","unstructured":"Talmor, A., Herzig, J., Lourie, N., Berant, J.: CommonsenseQA: a question answering challenge targeting commonsense knowledge. In: Burstein, J., Doran, C., Solorio, T. (eds.) ACL, Volume 1 (Long and Short Papers), pp. 4149\u20134158. Association for Computational Linguistics, Minneapolis (2019). https:\/\/doi.org\/10.18653\/v1\/N19-1421, https:\/\/aclanthology.org\/N19-1421","DOI":"10.18653\/v1\/N19-1421"},{"key":"33_CR18","unstructured":"Touvron, H., et\u00a0al.: Llama 2: open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288 (2023)"},{"key":"33_CR19","doi-asserted-by":"publisher","unstructured":"Wang, Y., et al.: Self-instruct: aligning language models with self-generated instructions. In: Rogers, A., Boyd-Graber, J., Okazaki, N. (eds.) ACL (Volume 1: Long Papers), pp. 13484\u201313508. Association for Computational Linguistics, Toronto (2023). https:\/\/doi.org\/10.18653\/v1\/2023.acl-long.754, https:\/\/aclanthology.org\/2023.acl-long.754","DOI":"10.18653\/v1\/2023.acl-long.754"},{"key":"33_CR20","doi-asserted-by":"crossref","unstructured":"Wang, Z.M., et al.: Rolellm: benchmarking, eliciting, and enhancing role-playing abilities of large language models (2023)","DOI":"10.18653\/v1\/2024.findings-acl.878"},{"key":"33_CR21","unstructured":"Wei, J., et al.: Finetuned language models are zero-shot learners. In: International Conference on Learning Representations (2022). https:\/\/openreview.net\/forum?id=gEZrGCozdqR"},{"key":"33_CR22","unstructured":"Wei, J., et al.: Chain-of-thought prompting elicits reasoning in large language models. In: Koyejo, S., Mohamed, S., Agarwal, A., Belgrave, D., Cho, K., Oh, A. (eds.) Advances in Neural Information Processing Systems, vol.\u00a035, pp. 24824\u201324837. Curran Associates, Inc. (2022)"},{"key":"33_CR23","doi-asserted-by":"crossref","unstructured":"Wu, N., Gong, M., Shou, L., Liang, S., Jiang, D.: Large language models are diverse role-players for summarization evaluation. arXiv preprint arXiv:2303.15078 (2023)","DOI":"10.1007\/978-3-031-44693-1_54"},{"key":"33_CR24","unstructured":"Xu, C., et al.: WizardLM: empowering large pre-trained language models to follow complex instructions. In: The Twelfth International Conference on Learning Representations (2024). https:\/\/openreview.net\/forum?id=CfXh93NDgH"},{"key":"33_CR25","unstructured":"Yao, S., et al.: Tree of thoughts: deliberate problem solving with large language models. In: Thirty-seventh Conference on Neural Information Processing Systems (2023). https:\/\/openreview.net\/forum?id=5Xc1ecxO1h"},{"key":"33_CR26","unstructured":"Zheng, H.S., et al.: Take a step back: evoking reasoning via abstraction in large language models (2024)"},{"key":"33_CR27","unstructured":"Zhou, C., et al.: LIMA: less is more for alignment. In: Thirty-seventh Conference on Neural Information Processing Systems (2023). https:\/\/openreview.net\/forum?id=KBMOKmX2he"},{"key":"33_CR28","unstructured":"Zhou, D., et al.: Least-to-most prompting enables complex reasoning in large language models. In: The Eleventh International Conference on Learning Representations (2023). https:\/\/openreview.net\/forum?id=WZH7099tgfM"}],"container-title":["Lecture Notes in Computer Science","Natural Language Processing and Chinese Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-3343-5_33","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,22]],"date-time":"2025-11-22T06:30:24Z","timestamp":1763793024000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-3343-5_33"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,23]]},"ISBN":["9789819533428","9789819533435"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-3343-5_33","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,11,23]]},"assertion":[{"value":"23 November 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"NLPCC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"CCF International Conference on Natural Language Processing and Chinese Computing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Urumqi","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"7 August 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"9 August 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"14","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"nlpcc2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/tcci.ccf.org.cn\/conference\/2025\/index.php","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}