{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,9]],"date-time":"2025-10-09T01:04:25Z","timestamp":1759971865419,"version":"build-2065373602"},"reference-count":97,"publisher":"Springer Science and Business Media LLC","issue":"11","license":[{"start":{"date-parts":[[2025,9,11]],"date-time":"2025-09-11T00:00:00Z","timestamp":1757548800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,9,11]],"date-time":"2025-09-11T00:00:00Z","timestamp":1757548800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"the National Key Research and Development Program of China","award":["2023YFC3303800"],"award-info":[{"award-number":["2023YFC3303800"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Cluster Comput"],"published-print":{"date-parts":[[2025,10]]},"DOI":"10.1007\/s10586-025-05482-y","type":"journal-article","created":{"date-parts":[[2025,9,11]],"date-time":"2025-09-11T12:28:28Z","timestamp":1757593708000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Securing social network user data in large language model deployments: challenges and best practices"],"prefix":"10.1007","volume":"28","author":[{"given":"Nasir Ahmad","family":"Jalali","sequence":"first","affiliation":[]},{"given":"Chen","family":"Hongsong","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,9,11]]},"reference":[{"key":"5482_CR1","first-page":"1","volume":"1","author":"BC Das","year":"2024","unstructured":"Das, B.C., Amini, M.H., Wu, Y.: Security and privacy challenges of large language models: a survey. ACM 1, 1\u201334 (2024)","journal-title":"ACM"},{"key":"5482_CR2","doi-asserted-by":"crossref","unstructured":"Song, C.H., Wu. J., Washington. C., et al: LLM-Planner: few-shot grounded planning for embodied agents with large language models, in 2023 IEEE\/CVF international conference on computer vision (ICCV), Paris, France, (2023)","DOI":"10.1109\/ICCV51070.2023.00280"},{"issue":"2","key":"5482_CR3","doi-asserted-by":"publisher","first-page":"230","DOI":"10.1109\/TETCI.2022.3141105","volume":"6","author":"J Duan","year":"2022","unstructured":"Duan, J., Yu, S., Li Tan, H., et al.: A survey of embodied AI: from simulators to research. IEEE Trans. Emerg. Topics Comput. Intell. 6(2), 230\u2013244 (2022)","journal-title":"IEEE Trans. Emerg. Topics Comput. Intell."},{"key":"5482_CR4","unstructured":"Yang, J: Large language models privacy and security, in Proceedings of the 2nd international conference on software engineering and machine learning, (2024)"},{"issue":"5","key":"5482_CR5","doi-asserted-by":"publisher","first-page":"1122","DOI":"10.1109\/JAS.2023.123618","volume":"10","author":"W Tianyu","year":"2023","unstructured":"Tianyu, W., He, S., Liu, J., et al.: A brief overview of ChatGPT: the history, status quo and potential future development. IEEE\/CAA J. Automat. Sinica 10(5), 1122\u20131136 (2023)","journal-title":"IEEE\/CAA J. Automat. Sinica"},{"key":"5482_CR6","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pdig.000","author":"J Gallifant","year":"2024","unstructured":"Gallifant, J., Fiske, A., Strekalova, Y.A.L., et al.: Peer review of GPT-4 technical report and system card. PLOS Digit Health (2024). https:\/\/doi.org\/10.1371\/journal.pdig.000","journal-title":"PLOS Digit Health"},{"key":"5482_CR7","doi-asserted-by":"crossref","unstructured":"Pan, X., Zhang, M., Ji, S., & Yang M.: Privacy risks of general-purpose language models, in 2020 IEEE Symposium on security and privacy, (2020)","DOI":"10.1109\/SP40000.2020.00095"},{"key":"5482_CR8","doi-asserted-by":"crossref","unstructured":"Bender, EM., Gebru, T., Mcmillan-Major A., et al: On the dangers of stochastic parrots: can language models be too big?, in FAccT \u201821: Proceedings of the 2021 ACM conference on fairness, accountability, and transparency, (2021)","DOI":"10.1145\/3442188.3445922"},{"key":"5482_CR9","doi-asserted-by":"crossref","unstructured":"Pan, X., Zhang, M., Ji, S., Yang M.: Privacy risk of general-purpose languge models, in 2020 IEEE symposium on security and privacy, (2020)","DOI":"10.1109\/SP40000.2020.00095"},{"key":"5482_CR10","unstructured":"Muhammad Usman H., et al: A survey on large language models: applications, challenges, limitations, and practical usage, Authorea Preprints, (2023)"},{"key":"5482_CR11","unstructured":"Li, Y., Wen, H., Wang, W. et al: Personal LLM agents: insights and survey about the capability, efficiency and security, arXiv:2401.05459v2 [cs.HC] 8 May 2024, (2024)"},{"key":"5482_CR12","unstructured":"Naveed, H., Ullah Khan, A., Qiu S. et al: A comprehensive overview of large language models,\u201d arXiv:2307.06435v9, vol. 9, pp. 1\u201346, 2023."},{"key":"5482_CR13","unstructured":"Guo, Z., Jin, R., Liu, C. et al: Evaluating large language models: a comprehensive sruvey, arXiv:2310.19736v3 [cs.CL] 25 Nov 2023, 2023."},{"key":"5482_CR14","unstructured":"Liu, Y., Yao, Y., Ton, J-F. et al: Trustworthy LLMs: a survey and guideline for evaluating large language models\u2019 Allignment, arXiv:2308.05374v2 [cs.AI] 21 Mar 2024, (2024)"},{"key":"5482_CR15","doi-asserted-by":"crossref","unstructured":"Subramani, N., Luccioni, A.S., Dodge J., Mitchell, M.: Detecting personal information in training corpora: an analysis, in Proceedings of the 3rd wrokshop on trustworthy natrural language processing (TrustNLP 2023), (2023)","DOI":"10.18653\/v1\/2023.trustnlp-1.18"},{"key":"5482_CR16","unstructured":"Shayegani, E., Md Abdullah AI Mamun, et al: Survey of vulnerabilities in large language models revealed by adversarial attacks, arXiv:2310.10844v1 [cs.CL] 16 Oct 2023, (2023)"},{"key":"5482_CR17","unstructured":"Schwinn, L., Dobre D., Gunnemann, S., Gidel, G.: Adversarial attacks and defenses in large language models: old and new threats, in 37th Conference on neural information processing system (NeurIPS2023) ICBINB Workshop, 2023."},{"key":"5482_CR18","unstructured":"Derner, E., Batistic K.: Beyond the safeguards: exploring the security risks of ChatGPT, arXiv:2305.08005v1 [cs.CR] 13 May 2023, pp. 1\u20138, (2023)"},{"key":"5482_CR19","doi-asserted-by":"crossref","unstructured":"Weidinger, L., Uesato, J., Rauh M., et al: Texonomy of risks posed by language models, in FAccT \u201822: Proceedings of the 2022 ACM conference on fairness, accountability, and transparency, (2022)","DOI":"10.1145\/3531146.3533088"},{"key":"5482_CR20","doi-asserted-by":"crossref","unstructured":"Lin, S., Hilton J., Evans, O.: TruthfulQA: Measuring how meodels mimic human falsehoods, in Proceedings of the 60th annual meeting of the association for computational linguistics, (2022)","DOI":"10.18653\/v1\/2022.acl-long.229"},{"key":"5482_CR21","doi-asserted-by":"publisher","DOI":"10.1145\/359730715:1-21","author":"R Navigli","year":"2023","unstructured":"Navigli, R., Conia, S., Ross, B.: Biases in large language models: origins, inventory, and discussion. ACM J. Data Inf. Quality (2023). https:\/\/doi.org\/10.1145\/359730715:1-21","journal-title":"ACM J. Data Inf. Quality"},{"key":"5482_CR22","doi-asserted-by":"crossref","unstructured":"Karpinska, M., Iyyer M., Large language models effectively leverage document-level context for literary translation, But Critical Errors Persist, arXiv preprint arXiv:2304.03245v3 [cs.CL] 22 May 2023, (2023)","DOI":"10.18653\/v1\/2023.wmt-1.41"},{"key":"5482_CR23","first-page":"177","volume":"24","author":"A Chowdher","year":"2023","unstructured":"Chowdher, A., Narang, S., Devlin, J., et al.: PaLM: scaling lanaguage modeling with pathways. J. Mach. Learn. Res. 24, 177\u2013188 (2023)","journal-title":"J. Mach. Learn. Res."},{"key":"5482_CR24","unstructured":"Biderman S., Schoelkopf, H., Anthony, Q. et al: Pythia: a suite for analzing large language models across training and scaling, in ICML\u201923: Proceeding of the 40th International Conference on Machine Learning, (2023)"},{"key":"5482_CR25","doi-asserted-by":"crossref","unstructured":"Chen, M., Xiao, C., Sun H et al: Combating security and privacy issues in the era of large languange models, in Proceedings of the 2024 conference of the North American chapter of the association for computational linguistics: human language technologies, 2024.","DOI":"10.18653\/v1\/2024.naacl-tutorials.2"},{"key":"5482_CR26","first-page":"102","volume":"2","author":"Wu Xiodong","year":"2024","unstructured":"Xiodong, Wu., Duan, R., Ni, J.: Unveiling security, privacy, and ethical concerns of ChatGPT. J. Inf. Intell. 2, 102\u2013115 (2024)","journal-title":"J. Inf. Intell."},{"key":"5482_CR27","unstructured":"Carlini, N., Tramer, F., Wallace, E., et al: Extracting training data from large language models, in Proceedings of the 30th USENIX Security Symposium, (2021)"},{"key":"5482_CR28","doi-asserted-by":"crossref","unstructured":"Si, W.M., Backes, M., Blackburn, J. et al: Why so Toxic?: measuring and triggering toxic behavior in open-domain chatbots, in CCS \u201822 Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, (2022)","DOI":"10.1145\/3548606.3560599"},{"key":"5482_CR29","unstructured":"Perez, F., Ribeiro, I.: Ignore priveious prompt: attacks techniques for language model, in ML Sefety workshop, 36th conference on neural information processing system (NeurIPS 2022), (2022)."},{"issue":"1","key":"5482_CR30","first-page":"1","volume":"15","author":"G Sebastian","year":"2023","unstructured":"Sebastian, G.: Do ChatGPT and Other AI Chatbots pose a cybersecurity risk? and exploratory study. Int. J. Sec. Privacy Pervas. Comput. 15(1), 1\u201311 (2023)","journal-title":"Int. J. Sec. Privacy Pervas. Comput."},{"issue":"3","key":"5482_CR31","doi-asserted-by":"publisher","first-page":"9","DOI":"10.1109\/MITP.2023.3275489","volume":"25","author":"N Kshetri","year":"2023","unstructured":"Kshetri, N.: Cybercrime and privacy threats of large language models. IT Professional 25(3), 9\u201313 (2023)","journal-title":"IT Professional"},{"key":"5482_CR32","first-page":"1","volume":"99","author":"EN Crothers","year":"2023","unstructured":"Crothers, E.N., Japkowicz, N., Viktor, H.L.: Machine-generated text: a comprehensive survey of threat models and detection methods. IEEE Access 99, 1\u201327 (2023)","journal-title":"IEEE Access"},{"key":"5482_CR33","doi-asserted-by":"crossref","unstructured":"Dong, Z., Zhou, Z., Yang, C., et al: Attacks, defenses and evaluations for LLMS conversations safety: a survy, in Proceedings of the 2024 conference of the North American chapter of the association for computational linguistics: human language technolgoy, (2024)","DOI":"10.18653\/v1\/2024.naacl-long.375"},{"key":"5482_CR34","doi-asserted-by":"crossref","unstructured":"Kang, D., Li, S., Stoica, L. et al: Exploiting programmatic behavior of LLMs: dual-use through standard security attacks, in 2024 IEEE security and privacy workshops (SPW), (2024)","DOI":"10.1109\/SPW63631.2024.00018"},{"key":"5482_CR35","doi-asserted-by":"publisher","first-page":"175","DOI":"10.1007\/s10462-024-10824-0","volume":"57","author":"X Huang","year":"2024","unstructured":"Huang, X., Ruan, W., Huang, W., et al.: A survey of satety and trustworthiness of large language models through the lens of verification and validation. Artif. Intell. Rev.. Intell. Rev. 57, 175 (2024)","journal-title":"Artif. Intell. Rev.. Intell. Rev."},{"key":"5482_CR36","doi-asserted-by":"crossref","unstructured":"Akter, M.S., Rahman, M.A., Rahman, M.M. et al: Authentic learning approach for data poisoning vulnerability in LLMs, in 2024 IEEE 48th annual computers, software, and applications conference (COMPSAC), Osaka, Japan, (2024)","DOI":"10.1109\/COMPSAC61105.2024.00210"},{"key":"5482_CR37","doi-asserted-by":"crossref","unstructured":"Song, C., Raghunathan, A.: Infromation leakage in embedding models, in In Proceedings of the 2020 ACM SIGSAC conference on computer and communications security (CCS \u201820), New York, NY, USA, (2020)","DOI":"10.1145\/3372297.3417270"},{"key":"5482_CR38","doi-asserted-by":"crossref","unstructured":"Brown, H., Lee, K, Mireshghallah, F.: What does it mean for a language model to preserve privacy?, in In 2022 ACM conference on fairness, accountability, and transparency (FAccT \u201822), New York, NY, USA, (2022)","DOI":"10.1145\/3531146.3534642"},{"key":"5482_CR39","unstructured":"Kirk, H.R., Jun, Y., Iqbal, H. et al: Bias out-of-the-box: an empirical analysis of intersectional occupational biases in popular generative language models, in 35th conference on neural information processing systems (NeurIPS 2021), (2021)"},{"key":"5482_CR40","doi-asserted-by":"publisher","first-page":"109698","DOI":"10.1016\/j.compeleceng.2024.109698","volume":"120","author":"H Kibriya","year":"2024","unstructured":"Kibriya, H., Khan, W.Z., Siddiqa, A., Khan, M.K.: Privacy issues in large language models: a survey. Computers Electr. Eng. 120, 109698 (2024)","journal-title":"Computers Electr. Eng."},{"key":"5482_CR41","unstructured":"Deng, J., Cheng, J., Sun, H., Zhang, Z., Huang, M.: Towards safer generative lanaguage models: a survey on safety risks, evaluations, and improvments, arXiv preprint arXiv:2302.09270."},{"key":"5482_CR42","doi-asserted-by":"crossref","unstructured":"Dhamala, J., Sun, T., Kumar, V. et al: BOLD: Dataset and metrics for measuring biases in open-ended language generation, in FAccT \u201821: Proceeding of the 2021 ACM conference on fairness, accountability, and transparency, Canada, (2021)","DOI":"10.1145\/3442188.3445924"},{"key":"5482_CR43","doi-asserted-by":"crossref","unstructured":"Ranade, P., Piplai, A., Mittal, S., Joshi, A.: Generating fake cyber threat intelligence using transformer-based models, in International joint conference on neural netwroks (IJCNN 2021), (2021)","DOI":"10.1109\/IJCNN52387.2021.9534192"},{"key":"5482_CR44","doi-asserted-by":"publisher","DOI":"10.2139\/ssrn.4425678","author":"S Addington","year":"2023","unstructured":"Addington, S.: ChatGPT: cyber security threats and countermeasures. SSRN Electronic J. (2023). https:\/\/doi.org\/10.2139\/ssrn.4425678","journal-title":"SSRN Electronic J."},{"key":"5482_CR45","unstructured":"Hazell J.: Spear phishing with large language models, Oxford Internt Institiute (Center for the Governance of AI), (2023)"},{"issue":"1","key":"5482_CR46","first-page":"1","volume":"15","author":"G Sebastian","year":"2023","unstructured":"Sebastian, G.: Do ChatGPT and other AI Chatbots pose a cybersecurity risk? an exploratory study. Int. J. Sec. Privacy Pervasive Comput. 15(1), 1\u201312 (2023)","journal-title":"Int. J. Sec. Privacy Pervasive Comput."},{"key":"5482_CR47","doi-asserted-by":"crossref","unstructured":"Chinnasamy, P., Vinodhini, B, Praveena, V., Vinothini, C., Ben Sujitha, B: Blockchain based Access Control and Data Sharing System for Smart Devices, in International E-conference on data analytics, intelligent systems and information security & ICDIIS , Pollachi, India, (2021)","DOI":"10.1088\/1742-6596\/1767\/1\/012056"},{"key":"5482_CR48","doi-asserted-by":"publisher","first-page":"1001","DOI":"10.1007\/s12652-021-02942-2","volume":"13","author":"P Chinnasamy","year":"2021","unstructured":"Chinnasamy, P., Deepalakshmi, P.: HCAC-EHR: hybrid cryptogrpahic access control for secure ehr retrieval in healthcare cloud. J. Ambient Intell. Humanized Comput. 13, 1001\u20131019 (2021)","journal-title":"J. Ambient Intell. Humanized Comput."},{"key":"5482_CR49","doi-asserted-by":"publisher","DOI":"10.1007\/s13735-024-00334-8","author":"P Kumar","year":"2024","unstructured":"Kumar, P.: Adversarial attacks and defenses for large language models (LLMs): methods, framework & challenges. Int. J. Multimedia Inf. Retrieval (2024). https:\/\/doi.org\/10.1007\/s13735-024-00334-8","journal-title":"Int. J. Multimedia Inf. Retrieval"},{"key":"5482_CR50","unstructured":"Rashid, H.: ChatGPT bug exposed payment details of paid users\u2014HACKREAD.Com, OpenAI, 24 March 2023. [Online]. Available: https:\/\/hackread.com\/chatgpt-bug-exposed-payment-details\/. [Accessed 25 9 2024]."},{"key":"5482_CR51","doi-asserted-by":"crossref","unstructured":"Zhang, H., Guo, Z., Zhu, H. et al: Jailbreak open-sourced large language models vai enforced decoding, in Proceeding of the 62nd Annual Meeting of the Association for Computaional Linguistics, 2024.","DOI":"10.18653\/v1\/2024.acl-long.299"},{"key":"5482_CR52","unstructured":"AJ ONeal, \u201cChatGPT-Dan-Jailbreak.md, https:\/\/gist.github.com\/coolaj86\/6f4f7b30129b0251f61fa7baaa881516, 27 9 2024."},{"key":"5482_CR53","unstructured":"Kim, E., \u201cAmazon warns employees not to share confidential information with ChatGPT after seeing cases where its answer \u2018closely matches existing marerial\u2019 from inside the company, https:\/\/www.businessinsider.com\/amazon-chatgpt-openai-warns-employees-not-share-confidential-information-microsoft-2023-1#:~:text=Amazon%20warns%20employees%20not%20to,material\u2019%20from%20inside%20the%20company, 31 9 2024."},{"key":"5482_CR54","unstructured":"Tilley, A., Kruppa, M.: Apple restricts employee use of ChatGPT, joining other companies wary of leaks, The Wall Street Journal - https:\/\/www.wsj.com\/articles\/apple-restricts-use-of-chatgpt-joining-other-companies-wary-of-leaks-d44d7d34, 28 9 (2024)"},{"key":"5482_CR55","unstructured":"Bushard, B.: Workers\u2019 ChatGPT use restricted at more banks including goldman, citigroup, https:\/\/www.forbes.com\/sites\/brianbushard\/2023\/02\/24\/workers-chatgpt-use-restricted-at-more-banks-including-goldman-citigroup\/?sh=3d767d496cf4, 28 9 2024."},{"key":"5482_CR56","doi-asserted-by":"crossref","unstructured":"Vemprala, N., Dietrich, G.: A social network analysis (SNA) study on data breach concerns over social media, in Proceedings of the 52nd Hawaii International Conference on System Sciences, (2019)","DOI":"10.24251\/HICSS.2019.862"},{"key":"5482_CR57","doi-asserted-by":"publisher","first-page":"103790","DOI":"10.1016\/j.cose.2024.103790","volume":"140","author":"G Jethava","year":"2024","unstructured":"Jethava, G., Pratap Rao, U.: Exploring security and trust mechanism in online social networks: an extensive review. Comptuer Sec. 140, 103790 (2024)","journal-title":"Comptuer Sec."},{"key":"5482_CR58","unstructured":"Dixon, H.: Data protection commission annual report, An Coimisiun um Chosaint Sonrai Data Protection Commission, 2021, 2022, (2023)"},{"key":"5482_CR59","unstructured":"Abrams, L.: 5.4 million Twitter users\u2019 stolen data leaked online\u2014more shared privately, https:\/\/www.bleepingcomputer.com\/news\/security\/54-million-twitter-users-stolen-data-leaked-online-more-shared-privately\/, 27 11 (2022)"},{"key":"5482_CR60","first-page":"100545","volume":"16","author":"D Barman","year":"2024","unstructured":"Barman, D., Guo, Z., Conlan, O.: The dark side of language models: exploring the potential of LLMs in multimedia disinformation generation and dissemination\u201d. Mach. Learn. Appl. 16, 100545 (2024)","journal-title":"Mach. Learn. Appl."},{"key":"5482_CR61","doi-asserted-by":"publisher","first-page":"493","DOI":"10.37394\/232015.2020.16.50","volume":"16","author":"D Galinec","year":"2020","unstructured":"Galinec, D., Luic, L.: Design of conceptual model for raising awareness of digital threats. WSEAS Trans. Environ. Dev. 16, 493\u2013504 (2020)","journal-title":"WSEAS Trans. Environ. Dev."},{"key":"5482_CR62","unstructured":"Ali, T., Kostakos, P.: HuntGPT: Integrating machine learning based anomaly detection and explianable AI with large language models (LLMs), arXiv preprint arXiv:2309.16021, 2023."},{"key":"5482_CR63","volume-title":"Communications-efficient LLM training for federated learning","author":"A Raje","year":"2024","unstructured":"Raje, A.: Communications-efficient LLM training for federated learning. Carnegie Mellon University, Pittsburgh (2024)"},{"key":"5482_CR64","unstructured":"Yu, S., Pablo Munoz, J., Jannesari, A.: Federated foundation models: privacy-preserving and collaborative learning for large models, ELRA Language Resources Association: CC BY-NC 4.0, pp. 7174-7184, (2024)"},{"key":"5482_CR65","first-page":"1","volume":"99","author":"Y Jiang","year":"2022","unstructured":"Jiang, Y., Jun Ko, B., Wang, S.: Model pruning enables efficient federated learning on edge devices. IEEE Trans. Neural Netw. Learn. Syst. 99, 1\u201313 (2022)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"5482_CR66","unstructured":"Karimireddy, S.P., Kale, S., Mohri, M. et al: SCAFFOLD: stochastic controlled averaging for federated learning, in Proceedings of the 37th International Conference on Machine Learning, New York, (2020)"},{"key":"5482_CR67","doi-asserted-by":"crossref","unstructured":"Chen, Y., Arunasalam, A., Berkay Celik, Z.: Can large language models provide security & privacy advice? measuring the ability of LLMs to fefute misconception, in In Annual Comptuer Security Applicatoins Conference (ACSAC \u201823), New York, USA, (2023)","DOI":"10.1145\/3627106.3627196"},{"issue":"4","key":"5482_CR68","first-page":"77","volume":"21","author":"M Penelova","year":"2021","unstructured":"Penelova, M.: Access control models. Cybernet. Inf. Technol. 21(4), 77\u2013105 (2021)","journal-title":"Cybernet. Inf. Technol."},{"issue":"2074","key":"5482_CR69","first-page":"1","volume":"14","author":"R Patil","year":"2024","unstructured":"Patil, R., Gudivada, V.: A review of current trends. techniques, and challenges in large language models (LLMs). Appl. Sci. 14(2074), 1\u201342 (2024)","journal-title":"Appl. Sci."},{"key":"5482_CR70","doi-asserted-by":"crossref","unstructured":"Jernite, Y., Nguyen, H., Biderman, S. et al: Data governance in the age of large-scale data-driven language technology, in in 2022 ACM Conference on Fairness, Accountability, and Transparency (FAccT \u201822), Seoul, Republic of Korea, (2022)","DOI":"10.1145\/3531146.3534637"},{"key":"5482_CR71","doi-asserted-by":"crossref","unstructured":"Hoory, S., Feder, A., Tendler, A., Cohen, A.: Learning and evaluating a differentially private pre-trained language model, in Proceedings of the third workshop on privacy in natural language processing, (2021)","DOI":"10.18653\/v1\/2021.findings-emnlp.102"},{"key":"5482_CR72","unstructured":"Duan, H., Dziedzic, A., Papernot, N. et al: Flocks of stochastic parrots: differentially private prompt learning for large language models, in NIPS \u201823: Proceedings of the 37th international conference on neural informaiton processing system, (2024)"},{"key":"5482_CR73","doi-asserted-by":"publisher","first-page":"15098","DOI":"10.1007\/s11227-023-05233-z","volume":"79","author":"TVT Doan","year":"2023","unstructured":"Doan, T.V.T., Messai, M.L., Gavin, G., Darmont, J.: A survey on implementations of homomorphic encryption schemes. J. Supercomput.Supercomput. 79, 15098\u201315139 (2023)","journal-title":"J. Supercomput.Supercomput."},{"key":"5482_CR74","doi-asserted-by":"publisher","first-page":"3759","DOI":"10.1007\/s40747-022-00756-z","volume":"9","author":"K Munjal","year":"2023","unstructured":"Munjal, K., Bhatia, R.: A systematic review of homomorphic encryption and its contributions in healthcare industry. Complex Intell. System 9, 3759\u20133786 (2023)","journal-title":"Complex Intell. System"},{"key":"5482_CR75","unstructured":"Hao, M., Li, H., Chen, H. et al: Iron: private inference on transformers, in 36th conference on neural information processing system (NeurIPS 2022), (2022)"},{"key":"5482_CR76","doi-asserted-by":"crossref","unstructured":"Dong, C., Weng, J., Liu, J-N. et al: Fusion: efficient and secure inference resilient to malicious servers, in Network and distributed system security (NDSS) Symposium , San Diego, CA, USA, (2023)","DOI":"10.14722\/ndss.2023.23199"},{"key":"5482_CR77","doi-asserted-by":"crossref","unstructured":"Pillai, S.E.V.S., Plimetla, K, \u201cEnhancing network privacy through secure multi-party computation in cloud environment,\u201d in 2024 International conference on integrated circuits and commmunication systems (ICICACS), Raichur, India, (2024)","DOI":"10.1109\/ICICACS60521.2024.10498662"},{"issue":"103142","key":"5482_CR78","first-page":"1","volume":"128","author":"C Gao","year":"2023","unstructured":"Gao, C., Jia, Yu.: SecureRC: a system for pirvacy-prserving relation classification using secure multi-party computation. Computer Sec. 128(103142), 1\u201311 (2023)","journal-title":"Computer Sec."},{"key":"5482_CR79","unstructured":"Rathee, D., Li, D., Stoica, I. et al: MPC-minimized secure LLM inference, arXiv: Prepring arXiv:2408.03561v1 [cs.CR] 7 Aug (2024)"},{"key":"5482_CR80","doi-asserted-by":"crossref","unstructured":"Gupta, K., Jawalkar, N., Mukherjee, A. et al: Sigma: Secure GPT Inference with function secret sharing,\u201d in Proceeding on Privacy Enhancing Technology, USA, CA, (2023)","DOI":"10.56553\/popets-2024-0107"},{"key":"5482_CR81","doi-asserted-by":"publisher","first-page":"100608","DOI":"10.1016\/j.cosrev.2023.100608","volume":"51","author":"AK Chattopadhyay","year":"2024","unstructured":"Chattopadhyay, A.K., Saha, S., Nag, A., Nandi, S.: Secret sharing: a comprehensive survey, taxonomy and applications. Computer Sci. Rev. 51, 100608 (2024)","journal-title":"Computer Sci. Rev."},{"key":"5482_CR82","doi-asserted-by":"publisher","first-page":"957","DOI":"10.1007\/s11277-023-10315-5","volume":"130","author":"D Pande","year":"2023","unstructured":"Pande, D., Singh Rawat, A., Deshmukh, M., Singh, M.: Single secret sharing scheme using chinese reminder theorem, modified shamir\u2019s scheme and XOR operation. Wireless Personal Commun 130, 957\u2013985 (2023)","journal-title":"Wireless Personal Commun"},{"key":"5482_CR83","doi-asserted-by":"crossref","unstructured":"Lukas, N., Salem, A., Sim, R. et al: Analyzing leakage of personally identifiable information in language models,\u201d in 2023 IEEE Symposium Security and Privacy (SP), (2023)","DOI":"10.1109\/SP46215.2023.10179300"},{"key":"5482_CR84","unstructured":"Zhou, X., Cao, S., Sun, X., Lo, D.: Large language model for vulnerability detection and repair: literature review and the road ahead,\u201d arXiv preprint arXiv:2404.02525v2 [cs.SE] 6 Apr (2024)."},{"key":"5482_CR85","unstructured":"Kim, S., Yun, S., Lee, H. et al: ProPILE:Probing privacy leakage in large language models,\u201d in 37th conference on neural information processing system (NeurIPS 2023), (2023)"},{"key":"5482_CR86","doi-asserted-by":"crossref","unstructured":"Chen, B., Paliwal, A., Yan, Q.: Jialbreaker in Jail: moving target defense for large language models,\u201d in MTD \u201823: Proceedings of the 20th ACM Workshop on Moving Target Defense, (2023)","DOI":"10.1145\/3605760.3623764"},{"key":"5482_CR87","unstructured":"Liu, Z., Wang, J., Dao, T. et al: Deja Vu: contextual sparsity for efficient LLMs at Inference Time,\u201d in Proceedings of the 40th international conference on machine learning, PMLR, (2023)"},{"key":"5482_CR88","unstructured":"Mireshghallah, N., Kim, H., Zhou, X. et al: Can LLMs keep a secret? testing privacy implications of language models via contextual integrity theory,\u201d in International Conference on Learning Representations, Vienna, Austria, (2024)."},{"key":"5482_CR89","doi-asserted-by":"crossref","unstructured":"Huang, J., Shao, H., Chang, K.C-C.: Are large pre-trained language models leaking your personal information,\u201d in Findings of the association for computational linguistics: EMNLP 2022, Abu Dhabi, United Arab Emirates, (2022)","DOI":"10.18653\/v1\/2022.findings-emnlp.148"},{"issue":"4","key":"5482_CR90","first-page":"5549","volume":"14","author":"P Wang","year":"2023","unstructured":"Wang, P., Zhu, J., Ma, Q.: Private data protection in social networks based on blockchain. Int. J. Adv. Netw. Appl. 14(4), 5549\u20135555 (2023)","journal-title":"Int. J. Adv. Netw. Appl."},{"key":"5482_CR91","first-page":"1","volume":"21","author":"S Dhall","year":"2021","unstructured":"Dhall, S., Dwivedi, A.D., Pal, S.K., et al.: Blockchain-based framework for reducing fake or vicious news spread on social media\/massaging platforms. ACM Trans. Asian Low-Resource Language Inf. Proc. 21, 1\u201333 (2021)","journal-title":"ACM Trans. Asian Low-Resource Language Inf. Proc."},{"key":"5482_CR92","doi-asserted-by":"publisher","DOI":"10.3390\/app12136567","author":"M Ali Hisseine","year":"2022","unstructured":"Ali Hisseine, M., Chen, D., Yang, X.: The application of blockchain in social media: a systemeatic literature review. Appl. Sci. (2022). https:\/\/doi.org\/10.3390\/app12136567","journal-title":"Appl. Sci."},{"key":"5482_CR93","doi-asserted-by":"crossref","unstructured":"Weidinger, L., Uesato, J., Rauh, M. et al: Taxonomy of risks posed by language models, in FAccT \u201822: Proceedings of the 20222 ACM conference on fairness, accountability, and transparency, (2022)","DOI":"10.1145\/3531146.3533088"},{"issue":"1","key":"5482_CR94","doi-asserted-by":"publisher","first-page":"1","DOI":"10.26833\/ijeg.987605","volume":"8","author":"C Avci","year":"2023","unstructured":"Avci, C., Budak, M., Yagmur, N., Bektas Balcik, F.: Camparison between random forest and support vector machine algorithm for LULC classification. Int. J. Eng. Geosci. 8(1), 1\u201310 (2023)","journal-title":"Int. J. Eng. Geosci."},{"key":"5482_CR95","first-page":"e01739","volume":"21","author":"A Gatera","year":"2023","unstructured":"Gatera, A., Kuraduseng, M., Bajpai, G., et al.: Comparison of random forest and support vector machine regression models for forecasting road accidents. Sci. African 21, e01739 (2023)","journal-title":"Sci. African"},{"key":"5482_CR96","doi-asserted-by":"publisher","first-page":"5983","DOI":"10.1038\/s41598-023-33215-x","volume":"13","author":"F Maite Siemers","year":"2023","unstructured":"Maite Siemers, F., Bajorath, J.: Differences in learning characteristics between support vector machine and random forest models for compound classification revealed by Shapley value analysis\u201d. Sci. Rep. 13, 5983 (2023)","journal-title":"Sci. Rep."},{"key":"5482_CR97","unstructured":"Kavzoglu, T., Bilucan, G., Teke, A.: Comparison of support vector machines, random forest and decision tree methods for classification of sentinel-2a image using different band combinations,\u201d in 41st Asian Conference on Remote Sensing (ACRS 2020), (2020)"}],"container-title":["Cluster Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10586-025-05482-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10586-025-05482-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10586-025-05482-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,8]],"date-time":"2025-10-08T17:31:38Z","timestamp":1759944698000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10586-025-05482-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,11]]},"references-count":97,"journal-issue":{"issue":"11","published-print":{"date-parts":[[2025,10]]}},"alternative-id":["5482"],"URL":"https:\/\/doi.org\/10.1007\/s10586-025-05482-y","relation":{},"ISSN":["1386-7857","1573-7543"],"issn-type":[{"type":"print","value":"1386-7857"},{"type":"electronic","value":"1573-7543"}],"subject":[],"published":{"date-parts":[[2025,9,11]]},"assertion":[{"value":"17 November 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 March 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 May 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 September 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"717"}}