{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T07:05:53Z","timestamp":1764572753176,"version":"3.46.0"},"reference-count":35,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2025,11,19]],"date-time":"2025-11-19T00:00:00Z","timestamp":1763510400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,11,19]],"date-time":"2025-11-19T00:00:00Z","timestamp":1763510400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Key Laboratory of Wireless Sensor Networks in University of Sichuan Province","award":["WSN2022001","WSN2022001","WSN2022001","WSN2022001","WSN2022001","WSN2022001"],"award-info":[{"award-number":["WSN2022001","WSN2022001","WSN2022001","WSN2022001","WSN2022001","WSN2022001"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int. J. Inf. Secur."],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1007\/s10207-025-01159-0","type":"journal-article","created":{"date-parts":[[2025,11,19]],"date-time":"2025-11-19T15:23:13Z","timestamp":1763565793000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["GSSA-AT: a textual adversarial defense method based on synonym substitution"],"prefix":"10.1007","volume":"24","author":[{"given":"Min","family":"Zhu","sequence":"first","affiliation":[]},{"given":"HuanZhou","family":"Li","sequence":"additional","affiliation":[]},{"given":"ZhangGuo","family":"Tang","sequence":"additional","affiliation":[]},{"given":"HanCheng","family":"Long","sequence":"additional","affiliation":[]},{"given":"Hao","family":"Yan","sequence":"additional","affiliation":[]},{"given":"Jian","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,11,19]]},"reference":[{"key":"1159_CR1","doi-asserted-by":"crossref","unstructured":"Seker, A., Bandel, E., Bareket, D., Brusilovsky, I., Greenfeld, R., Tsarfaty, R.: Alephbert: Language model pre-training and evaluation from sub-word to sentence level. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 46\u201356 (2022)","DOI":"10.18653\/v1\/2022.acl-long.4"},{"key":"1159_CR2","doi-asserted-by":"crossref","unstructured":"Hu, J., Hayashi, H., Cho, K., Neubig, G.: Deep: Denoising entity pre-training for neural machine translation. In: 60th Annual Meeting of the Association for Computational Linguistics, ACL 2022, pp. 1753\u20131766 (2022)","DOI":"10.18653\/v1\/2022.acl-long.123"},{"key":"1159_CR3","first-page":"20","volume":"1050","author":"IJ Goodfellow","year":"2015","unstructured":"Goodfellow, I.J., Shlens, J., Szegedy, C.: Explaining and harnessing adversarial examples. stat 1050, 20 (2015)","journal-title":"stat"},{"key":"1159_CR4","doi-asserted-by":"crossref","unstructured":"Chen, Y., Gao, H., Cui, G., Qi, F., Huang, L., Liu, Z., Sun, M.: Why should adversarial perturbations be imperceptible? rethink the research paradigm in adversarial nlp. In: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp. 11222\u201311237 (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.771"},{"key":"1159_CR5","doi-asserted-by":"crossref","unstructured":"Zheng, R., Dou, S., Zhou, Y., Liu, Q., Gui, T., Zhang, Q., Wei, Z., Huang, X.-J., Zhang, M.: Detecting adversarial samples through sharpness of loss landscape. In: Findings of the Association for Computational Linguistics: ACL 2023, pp. 11282\u201311298 (2023)","DOI":"10.18653\/v1\/2023.findings-acl.717"},{"key":"1159_CR6","doi-asserted-by":"crossref","unstructured":"Maheshwary, R., Maheshwary, S., Pudi, V.: A strong baseline for query efficient attacks in a black box setting. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 8396\u20138409 (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.661"},{"key":"1159_CR7","doi-asserted-by":"crossref","unstructured":"Wang, T., Wang, X., Qin, Y., Packer, B., Li, K., Chen, J., Beutel, A., Chi, E.: Cat-gen: Improving robustness in nlp models via controlled adversarial text generation. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 5141\u20135146 (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.417"},{"key":"1159_CR8","doi-asserted-by":"crossref","unstructured":"Hu, X., Liu, G., Zheng, B., Zhao, L., Wang, Q., Zhang, Y., Du, M.: Fasttextdodger: Decision-based adversarial attack against black-box nlp models with extremely high efficiency. IEEE Transactions on Information Forensics and Security, 2398\u20132411 (2024)","DOI":"10.1109\/TIFS.2024.3350376"},{"key":"1159_CR9","doi-asserted-by":"crossref","unstructured":"Peng, H., Guo, S., Zhao, D., Zhang, X., Han, J., Ji, S., Yang, X., Zhong, M.-H.: Textcheater: A query-efficient textual adversarial attack in the hard-label setting. IEEE Transactions on Dependable and Secure Computing, 3901\u20133916 (2024)","DOI":"10.1109\/TDSC.2023.3339802"},{"key":"1159_CR10","doi-asserted-by":"crossref","unstructured":"Jia, R., Raghunathan, A., G\u00f6ksel, K., Liang, P.: Certified robustness to adversarial word substitutions. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 4129\u20134142 (2019)","DOI":"10.18653\/v1\/D19-1423"},{"key":"1159_CR11","doi-asserted-by":"crossref","unstructured":"Wang, Y., Yang, Y., He, D., He, K.: Robustness-aware word embedding improves certified robustness to adversarial word substitutions. In: Findings of the Association for Computational Linguistics: ACL 2023, pp. 673\u2013687 (2023)","DOI":"10.18653\/v1\/2023.findings-acl.42"},{"key":"1159_CR12","unstructured":"Zhao, H., Ma, C., Dong, X., Luu, A.T., Deng, Z.-H., Zhang, H.: Certified robustness against natural language attacks by causal intervention. In: International Conference on Machine Learning, pp. 26958\u201326970 (2022)"},{"key":"1159_CR13","unstructured":"Zheng, R., Zhou, Y., Xi, Z., Gui, T., Zhang, Q., Huang, X.: Subspace defense: Discarding adversarial perturbations by learning a subspace for clean signals. In: International Conference on Language Resources and Evaluation (2024)"},{"key":"1159_CR14","doi-asserted-by":"crossref","unstructured":"Yang, Y., Liu, X., He, K.: Fast adversarial training against textual adversarial attacks. ArXiv (2024)","DOI":"10.18653\/v1\/2025.findings-naacl.43"},{"key":"1159_CR15","unstructured":"Wang, X., Hao, J., Yang, Y., He, K.: Natural language adversarial defense through synonym encoding. In: Uncertainty in Artificial Intelligence, pp. 823\u2013833 (2021)"},{"key":"1159_CR16","doi-asserted-by":"crossref","unstructured":"Jones, E., Jia, R., Raghunathan, A., Liang, P.: Robust encodings: A framework for combating adversarial typos. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 2752\u20132765 (2020)","DOI":"10.18653\/v1\/2020.acl-main.245"},{"key":"1159_CR17","doi-asserted-by":"crossref","unstructured":"Mozes, M., Stenetorp, P., Kleinberg, B., Griffin, L.: Frequency-guided word substitutions for detecting textual adversarial examples. In: Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, pp. 171\u2013186 (2021)","DOI":"10.18653\/v1\/2021.eacl-main.13"},{"key":"1159_CR18","unstructured":"Zhu, C., Cheng, Y., Gan, Z., Sun, S., Goldstein, T., Liu, J.: Freelb: Enhanced adversarial training for natural language understanding. In: International Conference on Learning Representations (2020)"},{"key":"1159_CR19","doi-asserted-by":"publisher","first-page":"119110","DOI":"10.1016\/j.eswa.2022.119110","volume":"214","author":"J Liu","year":"2023","unstructured":"Liu, J., Jin, H., Xu, G., Lin, M., Wu, T., Nour, M., Alenezi, F., Alhudhaif, A., Polat, K.: Aliasing black box adversarial attack with joint self-attention distribution and confidence probability. Expert Syst. Appl. 214, 119110 (2023)","journal-title":"Expert Syst. Appl."},{"key":"1159_CR20","doi-asserted-by":"crossref","unstructured":"Li, J., Ji, S., Du, T., Li, B., Wang, T.: Textbugger: Generating adversarial text against real-world applications. In: 26th Annual Network and Distributed System Security Symposium (2019)","DOI":"10.14722\/ndss.2019.23138"},{"key":"1159_CR21","doi-asserted-by":"crossref","unstructured":"Mrk\u0161ic, N., OS\u00e9aghdha, D., Thomson, B., Ga\u0161ic, M., Rojas-Barahona, L., Su, P.-H., Vandyke, D., Wen, T.-H., Young, S.: Counter-fitting word vectors to linguistic constraints. In: Proceedings of NAACL-HLT, pp. 142\u2013148 (2016)","DOI":"10.18653\/v1\/N16-1018"},{"key":"1159_CR22","unstructured":"Kannan, H., Kurakin, A., Goodfellow, I.: Adversarial logit pairing. arXiv preprint arXiv:1803.06373 (2018)"},{"key":"1159_CR23","unstructured":"Zhang, H., Yu, Y., Jiao, J., Xing, E., El\u00a0Ghaoui, L., Jordan, M.: Theoretically principled trade-off between robustness and accuracy. In: International Conference on Machine Learning, pp. 7472\u20137482 (2019)"},{"key":"1159_CR24","unstructured":"Song, C., He, K., Wang, L., Hopcroft, J.E.: Improving the generalization of adversarial training with domain adaptation. In: International Conference on Learning Representations (2019)"},{"key":"1159_CR25","unstructured":"Ding, G.W., Sharma, Y., Lui, K.Y.C., Huang, R.: Mma training: Direct input space margin maximization through adversarial training. In: International Conference on Learning Representations (2020)"},{"key":"1159_CR26","unstructured":"Zhang, X., Zhao, J., LeCun, Y.: Character-level convolutional networks for text classification. Advances in neural information processing systems 28 (2015)"},{"key":"1159_CR27","unstructured":"Yang, Y., Wang, X., He, K.: Robust textual embedding against word-level adversarial attacks. In: Proceedings of the Thirty-Eighth Conference on Uncertainty in Artificial Intelligence, pp. 2214\u20132224 (2022)"},{"key":"1159_CR28","unstructured":"Mikolov, T., Chen, K., Corrado, G.S., Dean, J.: Efficient estimation of word representations in vector space. In: International Conference on Learning Representations (2013)"},{"key":"1159_CR29","unstructured":"Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization. In: International Conference on Learning Representations, Conference Track Proceedings (2015)"},{"key":"1159_CR30","doi-asserted-by":"crossref","unstructured":"Ren, S., Deng, Y., He, K., Che, W.: Generating natural language adversarial examples through probability weighted word saliency. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 1085\u20131097 (2019)","DOI":"10.18653\/v1\/P19-1103"},{"issue":"4","key":"1159_CR31","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3576923","volume":"41","author":"C Wu","year":"2023","unstructured":"Wu, C., Zhang, R., Guo, J., De Rijke, M., Fan, Y., Cheng, X.: Prada: Practical black-box adversarial attacks against neural ranking models. ACM Transactions on Information Systems 41(4), 1\u201327 (2023)","journal-title":"ACM Transactions on Information Systems"},{"key":"1159_CR32","doi-asserted-by":"crossref","unstructured":"Mosca, E., Agarwal, S., Rando, J., Groh, G.L.: \"that is a suspicious reaction!\": Interpreting logits variation to detect nlp adversarial attacks. In: Annual Meeting of the Association for Computational Linguistics (2022)","DOI":"10.18653\/v1\/2022.acl-long.538"},{"issue":"2","key":"1159_CR33","doi-asserted-by":"publisher","first-page":"395","DOI":"10.1162\/coli_a_00476","volume":"49","author":"J Zeng","year":"2023","unstructured":"Zeng, J., Xu, J., Zheng, X., Huang, X.: Certified robustness to text adversarial attacks by randomized [mask]. Computational Linguistics 49(2), 395\u2013427 (2023)","journal-title":"Computational Linguistics"},{"issue":"1","key":"1159_CR34","doi-asserted-by":"publisher","first-page":"289","DOI":"10.1007\/s00521-023-08946-7","volume":"36","author":"J Huang","year":"2022","unstructured":"Huang, J., Chen, L.: Defense against adversarial attacks via textual embeddings based on semantic associative field. Neural Comput. Appl. 36(1), 289\u2013301 (2022)","journal-title":"Neural Comput. Appl."},{"key":"1159_CR35","unstructured":"Wang, Y., Zou, D., Yi, J., Bailey, J., Ma, X., Gu, Q.: Improving adversarial robustness requires revisiting misclassified examples. In: International Conference on Learning Representations (2019)"}],"container-title":["International Journal of Information Security"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10207-025-01159-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10207-025-01159-0\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10207-025-01159-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T07:02:36Z","timestamp":1764572556000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10207-025-01159-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,19]]},"references-count":35,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2025,12]]}},"alternative-id":["1159"],"URL":"https:\/\/doi.org\/10.1007\/s10207-025-01159-0","relation":{},"ISSN":["1615-5262","1615-5270"],"issn-type":[{"type":"print","value":"1615-5262"},{"type":"electronic","value":"1615-5270"}],"subject":[],"published":{"date-parts":[[2025,11,19]]},"assertion":[{"value":"3 September 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 November 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 November 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflicts of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of Interest"}},{"value":"The authors declare no competing interests.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"241"}}