{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,15]],"date-time":"2026-04-15T20:16:18Z","timestamp":1776284178964,"version":"3.50.1"},"reference-count":23,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,7,3]],"date-time":"2025-07-03T00:00:00Z","timestamp":1751500800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,7,3]],"date-time":"2025-07-03T00:00:00Z","timestamp":1751500800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,7,3]]},"DOI":"10.1109\/icecet63943.2025.11472287","type":"proceedings-article","created":{"date-parts":[[2026,4,9]],"date-time":"2026-04-09T19:42:35Z","timestamp":1775763755000},"page":"1-7","source":"Crossref","is-referenced-by-count":0,"title":["Prompt Injection Vulnerabilities and Data Leakage in ChatGPT and Claude: Toward Safer Conversational AI"],"prefix":"10.1109","author":[{"given":"Hyun Jung","family":"Kim","sequence":"first","affiliation":[{"name":"Sang-Huh College and the Graduate School of Information &amp; Communication Konkuk University,Dept. of Convergence Information Technology (Artificial Intelligence Major),Seoul,Korea"}]},{"given":"Sang Hyun","family":"Yoo","sequence":"additional","affiliation":[{"name":"Kyungmin University,Dept. of Computer Software,Korea"}]}],"member":"263","reference":[{"key":"ref1","article-title":"LLM jailbreaking: threats & mitigation strategies","year":"2024"},{"key":"ref2","article-title":"Prompt injection attacks on LLMs","year":"2024"},{"key":"ref3","article-title":"An early categorization of prompt injection attacks on LLMs","author":"Rossi","year":"2024"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1145\/3663530.3665021"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1145\/3658644.3670388"},{"key":"ref6","article-title":"Exfiltration of personal information from ChatGPT via prompt injection","author":"Clarke","year":"2024"},{"key":"ref7","article-title":"Aligning language models to follow instructions","year":"2022","journal-title":"OpenAI"},{"key":"ref8","article-title":"Claude\u2019s constitution","year":"2023","journal-title":"Anthropic"},{"key":"ref9","article-title":"Claude 3.7 sonnet system card","year":"2024","journal-title":"Anthropic"},{"key":"ref10","article-title":"Jailbreaking leading safety-aligned LLMs with simple adaptive attacks","author":"Andriushchenko","year":"2024"},{"key":"ref11","article-title":"Jailbroken: how does LLM safety training fail?","volume-title":"Proc. 37th NeurIPS 2023","author":"Wei"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1145\/3663530.3665021"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1145\/3719027.3744836"},{"key":"ref14","article-title":"OWASP top 10 for llm applications","year":"2023","journal-title":"OWASP"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1145\/3689217.3690621"},{"key":"ref16","article-title":"Formalizing and benchmarking prompt injection attacks and defenses","volume-title":"Proc. 33rd USENIX Security Symposium","author":"Liu"},{"key":"ref17","article-title":"Azure OpenAI Service content filtering","year":"2025","journal-title":"Azure OpenAI"},{"key":"ref18","article-title":"Constitutional classifiers: defending against universal jailbreaks","year":"2025","journal-title":"Anthropic"},{"key":"ref19","article-title":"Attention tracker: detecting prompt injection attacks in LLMs","author":"Hung","year":"2024"},{"key":"ref20","article-title":"Constitutional classifiers: defending against universal jailbreaks","author":"Sharma","year":"2025"},{"key":"ref21","article-title":"Red teaming GPT-4V: are GPT-4V safe against uni\/multimodal jailbreak attacks?","author":"Qi","year":"2024"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-emnlp.272"},{"key":"ref23","article-title":"Benchmarking and defending against indirect prompt injection attacks on large language models","author":"Xie","year":"2023"}],"event":{"name":"2025 5th International Conference on Electrical, Computer and Energy Technologies (ICECET)","location":"Paris, France","start":{"date-parts":[[2025,7,3]]},"end":{"date-parts":[[2025,7,6]]}},"container-title":["2025 5th International Conference on Electrical, Computer and Energy Technologies (ICECET)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11471878\/11471697\/11472287.pdf?arnumber=11472287","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,15]],"date-time":"2026-04-15T19:23:43Z","timestamp":1776281023000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11472287\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,3]]},"references-count":23,"URL":"https:\/\/doi.org\/10.1109\/icecet63943.2025.11472287","relation":{},"subject":[],"published":{"date-parts":[[2025,7,3]]}}}