{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,2]],"date-time":"2026-01-02T07:48:38Z","timestamp":1767340118769,"version":"3.41.0"},"publisher-location":"Cham","reference-count":46,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031893629","type":"print"},{"value":"9783031893636","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-89363-6_18","type":"book-chapter","created":{"date-parts":[[2025,5,24]],"date-time":"2025-05-24T07:56:19Z","timestamp":1748073379000},"page":"312-330","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["The Hidden Dangers of Publicly Accessible LLMs: A Case Study on Gab AI"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-7221-6826","authenticated-orcid":false,"given":"Lakshika","family":"Vaishnav","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0007-3163-1445","authenticated-orcid":false,"given":"Sakshi","family":"Singh","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9551-9689","authenticated-orcid":false,"given":"Kimberly A.","family":"Cornell","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,25]]},"reference":[{"key":"18_CR1","unstructured":"Helbling, A., Phute, M., Hull, M., Chau, D.H.: LLM self defense: by self examination, LLMs know they are being tricked. arXiv preprint arXiv:2308.07308 (2023)"},{"issue":"1","key":"18_CR2","doi-asserted-by":"publisher","first-page":"27","DOI":"10.3390\/info15010027","volume":"15","author":"M Alawida","year":"2024","unstructured":"Alawida, M., Abu Shawar, B., Abiodun, O.I., Mehmood, A., Omolara, A.E., Al Hwaitat, A.K.: Unveiling the dark side of ChatGPT: exploring cyberattacks and enhancing user awareness. Information 15(1), 27 (2024)","journal-title":"Information"},{"key":"18_CR3","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown, T., et al.: Language models are few-shot learners. Adv. Neural. Inf. Process. Syst. 33, 1877\u20131901 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"18_CR4","doi-asserted-by":"crossref","unstructured":"Chen, Y., Arunasalam, A., Celik, Z.B.: Can large language models provide security & privacy advice? Measuring the ability of LLMs to refute misconceptions. In: Proceedings of the Annual Computer Security Applications Conference (ACSAC 2023), Austin, TX, USA, pp. 1\u20132. ACM (2023)","DOI":"10.1145\/3627106.3627196"},{"key":"18_CR5","unstructured":"ChatGPT: DAN Mode (DO ANYTHING NOW). https:\/\/plainenglish.io\/blog\/chatgpt-dan-mode-doanything-now. Accessed 20 June 2023"},{"key":"18_CR6","unstructured":"ChatGPT-Dan-Jailbreak. https:\/\/gist.github.com\/coolaj86\/6f4f7b30129b0251f61fa7baaa881516. Accessed 20 June 2023"},{"key":"18_CR7","unstructured":"Elia\u00e7\u0131k, E.: ChatGPT, D.A.N., Prompt: Unleash the real ChatGPT. Ghacks (2023). https:\/\/www.ghacks.net\/2023\/04\/30\/chatgpt-dan-prompt-unleash-the-real-chatgpt\/. Accessed 28 Apr 2023"},{"key":"18_CR8","doi-asserted-by":"crossref","unstructured":"Falade, P.V.: Decoding the threat landscape: ChatGPT, fraudgpt, and wormgpt in social engineering attacks. arXiv preprint arXiv:2310.05595 (2023)","DOI":"10.32628\/CSEIT2390533"},{"key":"18_CR9","doi-asserted-by":"crossref","unstructured":"Gupta, M., Akiri, C., Aryal, K., Parker, E., Praharaj, L.: From chatGPT to threatGPT: impact of generative AI in cybersecurity and privacy. IEEE Access (2023)","DOI":"10.1109\/ACCESS.2023.3300381"},{"key":"18_CR10","unstructured":"Penedo, G., et al.: The refinedweb dataset for falcon LLM: outperforming curated corpora with web data, and web data only. arXiv preprint arXiv:2306.01116 (2023)"},{"key":"18_CR11","doi-asserted-by":"crossref","unstructured":"Hadi, M.U., et al.: Large language models: a comprehensive survey of its applications, challenges, limitations, and future prospects (2023)","DOI":"10.36227\/techrxiv.23589741.v3"},{"key":"18_CR12","unstructured":"Hagendorff, T.: Machine psychology: investigating emergent capabilities and behavior in large language models using psychological methods. arXiv:2303.13988 (2023)"},{"key":"18_CR13","doi-asserted-by":"crossref","unstructured":"He, J., Vechev, M.: Large language models for code: security hardening and adversarial testing. In: Proceedings of the ICML 2023 Workshop on Deployable Generative AI, pp. 1\u20135 (2023)","DOI":"10.1145\/3576915.3623175"},{"key":"18_CR14","unstructured":"How to Jailbreak ChatGPT: Get it to Really do What You Want. https:\/\/www.digitaltrends.com\/computing\/how-to-jailbreak-chatgpt\/. Accessed 20 June 2023"},{"key":"18_CR15","unstructured":"Kelly, D.: WormGPT \u2013 the generative AI tool cybercriminals are using to launch business email compromise attacks. SlashNext. https:\/\/slashnext.com\/blog\/ai-based-cybercrime-tools-wormgpt-and-fraudgpt-could-be-the-tip-of-the-iceberg\/ (2023)"},{"key":"18_CR16","unstructured":"Kosinski, M., Forrest, A.: Prompt injection. IBM. https:\/\/www.ibm.com\/topics\/prompt-injection (2024)"},{"key":"18_CR17","doi-asserted-by":"crossref","unstructured":"Li, H., et al.: Multi-step jailbreaking privacy attacks on chatGPT. arXiv preprint arXiv:2304.05197 (2024)","DOI":"10.18653\/v1\/2023.findings-emnlp.272"},{"key":"18_CR18","unstructured":"Lin, Z., Cui, J., Liao, X., Wang, X.: Malla: demystifying real-world large language model integrated malicious services. arXiv preprint arXiv:2401.03315 (2024)"},{"issue":"1","key":"18_CR19","first-page":"1","volume":"2","author":"MS Rahaman","year":"2023","unstructured":"Rahaman, M.S., Ahsan, M.T., Anjum, N., Terano, H.J.R., Rahman, M.M.: From chatgpt-3 to GPT-4: a significant advancement in AI-driven NLP tools. J. Eng. Emerg. Technol. 2(1), 1\u201311 (2023)","journal-title":"J. Eng. Emerg. Technol."},{"key":"18_CR20","unstructured":"Monje, A., Monje, A., Hallman, R., Hallman, R., Cybenko, G., Cybenko, G.: Being a bad influence on the kids: malware generation in less than five minutes using ChatGPT. J. Cybersecur. Res. (2023)"},{"key":"18_CR21","unstructured":"OWASP. Oct OWASP Top 10 for LLM (2023). https:\/\/owasp.org\/www-project-top-10-for-large-language-model-applications\/assets\/PDF\/OWASP-Top-10-for-LLMs-2023-v1_1.pdf"},{"key":"18_CR22","unstructured":"Paria, S., Dasgupta, A., Bhunia, S.: Divas: an LLM-based endto-end framework for SOC security analysis and policy-based protection (2023)"},{"key":"18_CR23","unstructured":"Poireault, K.: The dark side of generative AI: five malicious LLMs found on the dark web. Infosecurity Europe (2023). https:\/\/www.infosecurityeurope.com\/en-gb\/blog\/threat-vectors\/generative-ai-dark-web-bots.html"},{"key":"18_CR24","doi-asserted-by":"crossref","unstructured":"Ganesh, P., Chang, H., Strobel, M., Shokri, R.: On the impact of machine learning randomness on group fairness. In: Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency, pp. 1789\u20131800 (2023)","DOI":"10.1145\/3593013.3594116"},{"key":"18_CR25","unstructured":"Qammar, A., Wang, H., Ding, J., Naouri, A., Daneshmand, M., Ning, H.: Chatbots to ChatGPT in a cybersecurity space: evolution, vulnerabilities, attacks, challenges, and future recommendations. arXiv preprint arXiv:2306.09255 (2023)"},{"key":"18_CR26","doi-asserted-by":"crossref","unstructured":"Shen, X., Chen, Z., Backes, M., Shen, Y., Zhang, Y.: \u201cDo anything now\u201d: characterizing and evaluating in-the-wild jailbreak prompts on large language models. arXiv preprint arXiv:2308.03825 (2023)","DOI":"10.1145\/3658644.3670388"},{"key":"18_CR27","unstructured":"Tong, M., Chen, K., Qi, Y., Zhang, J., Zhang, W., Yu, N.: PrivInfer: privacy-preserving inference for black-box large language model. arXiv preprint arXiv:2310.12214 (2023)"},{"key":"18_CR28","doi-asserted-by":"crossref","unstructured":"Teubner, T., Flath, C.M., Weinhardt, C., van der Aalst, W., Hinz, O.: Welcome to the era of chatGPT et al. the prospects of large language models. Business & Information Systems Engineering, pp. 1\u20137 (2023)","DOI":"10.1007\/s12599-023-00795-x"},{"issue":"7947","key":"18_CR29","doi-asserted-by":"publisher","first-page":"224","DOI":"10.1038\/d41586-023-00288-7","volume":"614","author":"EAM van Dis","year":"2023","unstructured":"van Dis, E.A.M., Bollen, J., Zuidema, W., van Rooij, R., Bockting, C.L.: ChatGPT: five priorities for research. Nature 614(7947), 224\u2013226 (2023). https:\/\/doi.org\/10.1038\/d41586-023-00288-7. PMID: 36737653","journal-title":"Nature"},{"key":"18_CR30","unstructured":"Unknown author. Charybdis worm: Many infections spread widely via Discord, Telegram, LAN, and more [Online forum post]. Hack Forums (n.d.). https:\/\/hackforums.net\/showthread.php?tid=6229200"},{"key":"18_CR31","unstructured":"White, J., et al.: A prompt pattern catalog to enhance prompt engineering with chatGPT. arXiv preprint arXiv:2302.11382 (2023)"},{"key":"18_CR32","doi-asserted-by":"crossref","unstructured":"Wu, J., Hooi, B.: Fake news in sheep\u2019s clothing: robust fake news detection against LLM-empowered style attacks. arXiv preprint arXiv:2310.10830 (2023)","DOI":"10.1145\/3637528.3671977"},{"key":"18_CR33","doi-asserted-by":"crossref","unstructured":"Sun, X., et al.: Defending against backdoor attacks in natural language generation. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 37, no. 4, pp. 5257\u20135265 (2023)","DOI":"10.1609\/aaai.v37i4.25656"},{"key":"18_CR34","doi-asserted-by":"crossref","unstructured":"Yang, J., et al.: Harnessing the power of LLMs in practice: a survey on chatGPT and beyond. ACM Trans. Knowl. Discov. Data (2023)","DOI":"10.1145\/3649506"},{"key":"18_CR35","doi-asserted-by":"crossref","unstructured":"Yao, Y., Duan, J., Xu, K., Cai, Y., Sun, Z., Zhang, Y.: A survey on large language model (LLM) security and privacy: the good, the bad, and the ugly. High-Confidence Comput. 100211 (2024)","DOI":"10.1016\/j.hcc.2024.100211"},{"key":"18_CR36","unstructured":"Liu, Y., et al.: Jailbreaking ChatGPT via prompt engineering: an empirical study. arXiv preprint arXiv:2305.13860 (2023)"},{"key":"18_CR37","unstructured":"Zhang, Y., Song, W., Ji, Z., Meng, N.: How well does LLM generate security tests?. arXiv preprint arXiv:2310.00710 (2023)"},{"key":"18_CR38","doi-asserted-by":"crossref","unstructured":"Pa, Y.M., Tanizaki, S., Kou, T., Van Eeten, M., Yoshioka, K., Matsumoto, T.: An attacker\u2019s dream? Exploring the capabilities of ChatGPT for developing malware. In: Proceedings of the 16th Cyber Security Experimentation and Test Workshop, pp. 10\u201318 (2023)","DOI":"10.1145\/3607505.3607513"},{"key":"18_CR39","doi-asserted-by":"crossref","unstructured":"Zannettou, S., et al.: What is gab: a bastion of free speech or an alt-right echo chamber. In: Companion Proceedings of the Web Conference 2018, pp. 1007\u20131014 (2018)","DOI":"10.1145\/3184558.3191531"},{"key":"18_CR40","unstructured":"Xi, Z., et al.: Defending pre-trained language models as few-shot learners against backdoor attacks. arXiv preprint arXiv:2309.13256 (2023)"},{"key":"18_CR41","unstructured":"Chowdhery, A., et al.: PaLM: scaling language modeling with pathways. J. Mach. Learn. Res. 24(240), 1\u2013113 (2023)"},{"key":"18_CR42","doi-asserted-by":"crossref","unstructured":"Motlagh, F.N., Hajizadeh, M., Majd, M., Najafi, P., Cheng, F., Meinel, C.: Large language models in cybersecurity: state-of-the-art. arXiv:2402.00891 (2024)","DOI":"10.5220\/0013377600003899"},{"key":"18_CR43","unstructured":"Shayegani, E., Mamun, M.A.A., Fu, Y., Zaree, P., Dong, Y., Abu-Ghazaleh, N.: Survey of vulnerabilities in large language models revealed by adversarial attacks. arXiv:2310.10844 (2023)"},{"key":"18_CR44","unstructured":"Vaswani, A., et al.: Attention is all you need. Adv. Neural Inform. Process. Syst. 30 (2017)"},{"key":"18_CR45","unstructured":"Quak, N.: How emerging technologies threaten our cybersecurity (white paper). Cyber. Threat Alliance ((2023, August 28))"},{"key":"18_CR46","unstructured":"Varkey, B.: Jailbreaking large language models: techniques, examples, prevention methods. Lakera (2023, September 19)"}],"container-title":["Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering","Digital Forensics and Cyber Crime"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-89363-6_18","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,24]],"date-time":"2025-05-24T07:56:29Z","timestamp":1748073389000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-89363-6_18"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031893629","9783031893636"],"references-count":46,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-89363-6_18","relation":{},"ISSN":["1867-8211","1867-822X"],"issn-type":[{"value":"1867-8211","type":"print"},{"value":"1867-822X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"25 May 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICDF2C","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Digital Forensics and Cyber Crime","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Dubrovnik","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Croatia","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"9 October 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"10 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icdf2c2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/icdf2c.eai-conferences.org\/2024\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}