{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,24]],"date-time":"2025-11-24T15:20:20Z","timestamp":1763997620093,"version":"3.45.0"},"reference-count":46,"publisher":"MDPI AG","issue":"12","license":[{"start":{"date-parts":[[2025,11,23]],"date-time":"2025-11-23T00:00:00Z","timestamp":1763856000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100006769","name":"Russian Science Foundation","doi-asserted-by":"publisher","award":["25-71-10012"],"award-info":[{"award-number":["25-71-10012"]}],"id":[{"id":"10.13039\/501100006769","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Algorithms"],"abstract":"<jats:p>Cybersecurity remains a key challenge in the development of intelligent telecommunications systems and the Internet of Things (IoT). The growing destructive impact of the digital environment, coupled with high-performance computing (HPC), requires the development of effective countermeasures to ensure the security of the digital space. Traditional approaches to detecting destructive content are primarily limited to static text analysis, which ignores the temporal dynamics and evolution of destructive impact scenarios. This is critical for monitoring tasks in the digital environment, where threats rapidly evolve. To overcome this limitation, this study proposes a hybrid architecture, Hyb-TKAN, based on adaptive algorithms that account for the temporal component and nonlinear dependencies. This approach enables not only the classification of destructive messages but also the analysis of their development and transformation over time. Unlike existing studies, which focus on individual aspects of aggressive content, the model utilizes multilayered data analysis to identify hidden relationships and nonlinear patterns in destructive messages. The integration of these components ensures high adaptability and accuracy of text processing. The presented approach was implemented in a multi-class classification task with evaluation based on real text data. The obtained results demonstrate improved classification accuracy. In the Experimental Analysis Section, the results are compared with the closest modern analogs, confirming the relevance and competitiveness of the proposed hybrid neural network.<\/jats:p>","DOI":"10.3390\/a18120735","type":"journal-article","created":{"date-parts":[[2025,11,24]],"date-time":"2025-11-24T14:59:04Z","timestamp":1763996344000},"page":"735","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["A Hybrid Neural Network Transformer for Detecting and Classifying Destructive Content in Digital Space"],"prefix":"10.3390","volume":"18","author":[{"given":"Aleksandr","family":"Chechkin","sequence":"first","affiliation":[{"name":"Department of Mathematics and Data Analysis, Financial University Under the Government of the Russian Federation, Leningradsky Ave., 49\/2, 125167 Moscow, Russia"}]},{"given":"Ekaterina","family":"Pleshakova","sequence":"additional","affiliation":[{"name":"MIREA\u2014Russian Technological University, 78 Vernadsky Avenue, 119454 Moscow, Russia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0446-0552","authenticated-orcid":false,"given":"Sergey","family":"Gataullin","sequence":"additional","affiliation":[{"name":"MIREA\u2014Russian Technological University, 78 Vernadsky Avenue, 119454 Moscow, Russia"},{"name":"Social Modeling Lab, Central Economics and Mathematics Institute, Russian Academy of Sciences, Nakhimovsky Pr., 47, 117418 Moscow, Russia"}]}],"member":"1968","published-online":{"date-parts":[[2025,11,23]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"54608","DOI":"10.1109\/ACCESS.2024.3389497","article-title":"GPT (Generative Pre-Trained Transformer)\u2014A Comprehensive Review on Enabling Technologies, Potential Applications, Emerging Challenges, and Future Directions","volume":"12","author":"Yenduri","year":"2024","journal-title":"IEEE Access"},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"429","DOI":"10.1007\/s11416-024-00529-x","article-title":"Next Gen Cybersecurity Paradigm Towards Artificial General Intelligence: Russian Market Challenges and Future Global Technological Trends","volume":"20","author":"Pleshakova","year":"2024","journal-title":"J. Comput. Virol. Hacking Tech."},{"key":"ref_3","first-page":"83951","article-title":"Phased Consistency Models","volume":"37","author":"Wang","year":"2024","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_4","unstructured":"Zheng, J., Hu, M., Fan, Z., Wang, C., Ding, C., Tao, D., and Cham, T.J. (2025, November 20). Trajectory Consistency Distillation. CoRR 2024. Available online: https:\/\/openreview.net\/forum?id=aDJXCgfkf4."},{"key":"ref_5","unstructured":"Geng, Z., Pokle, A., Luo, W., Lin, J., and Kolter, J.Z. (2024). Consistency Models Made Easy. arXiv."},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"19525","DOI":"10.1109\/TNNLS.2025.3592935","article-title":"A Survey of Attacks on Large Vision\u2013Language Models: Resources, Advances, and Future Trends","volume":"36","author":"Liu","year":"2025","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"ref_7","unstructured":"Zhang, P., Zeng, G., Wang, T., and Lu, W. (2024). TinyLLaMA: An Open-Source Small Language Model. arXiv."},{"key":"ref_8","doi-asserted-by":"crossref","unstructured":"Schick, T., and Sch\u00fctze, H. (2021, January 6\u201311). It\u2019s Not Just Size That Matters: Small Language Models Are Also Few-Shot Learners. Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Online.","DOI":"10.18653\/v1\/2021.naacl-main.185"},{"key":"ref_9","doi-asserted-by":"crossref","unstructured":"Magister, L.C., Mallinson, J., Adamek, J., Malmi, E., and Severyn, A. (2023, January 9\u201314). Teaching Small Language Models to Reason. Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), Toronto, ON, Canada.","DOI":"10.18653\/v1\/2023.acl-short.151"},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Wang, F., Zhang, Z., Zhang, X., Wu, Z., Mo, T., Lu, Q., Wang, W., Li, R., Xu, J., and Tang, X. (2024). A Comprehensive Survey of Small Language Models in the Era of Large Language Models: Techniques, Enhancements, Applications, Collaboration with LLMs, and Trustworthiness. ACM Transactions on Intelligent Systems and Technology, ACM.","DOI":"10.1145\/3768165"},{"key":"ref_11","unstructured":"Hu, S., Tu, Y., Han, X., He, C., Cui, G., Long, X., Zheng, Z., Fang, Y., Huang, Y., and Zhao, W. (2024, January 7\u20139). MiniCPM: Unveiling the Potential of Small Language Models with Scalable Training Strategies. Proceedings of the First Conference on Language Modeling, Philadelphia, PA, USA."},{"key":"ref_12","first-page":"4819","article-title":"Weak-to-Strong Search: Align Large Language Models via Searching over Small Language Models","volume":"37","author":"Zhou","year":"2024","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_13","unstructured":"Kim, M.J., Pertsch, K., Karamcheti, S., Xiao, T., Balakrishna, A., Nair, S., Rafailov, R., Foster, E., Lam, G., and Sanketi, P. (2025). OpenVLA: An Open-Source Vision-Language-Action Model. arXiv."},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"3988","DOI":"10.1109\/LRA.2025.3544909","article-title":"TinyVLA: Towards Fast, Data-Efficient Vision-Language-Action Models for Robotic Manipulation","volume":"10","author":"Wen","year":"2025","journal-title":"IEEE Robot. Autom. Lett."},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Salazar, J., Liang, D., Nguyen, T.Q., and Kirchhoff, K. (2020, January 5\u201310). Masked Language Model Scoring. Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online.","DOI":"10.18653\/v1\/2020.acl-main.240"},{"key":"ref_16","unstructured":"Bao, H., Dong, L., Wei, F., Wang, W., Yang, N., Liu, X., Wang, Y., Gao, J., Piao, S., and Zhou, M. (2020, January 13\u201318). UNILMv2: Pseudo-Masked Language Models for Unified Language Model Pre-Training. Proceedings of the 37th International Conference on Machine Learning, Virtual."},{"key":"ref_17","first-page":"130136","article-title":"Simple and Effective Masked Diffusion Language Models","volume":"37","author":"Sahoo","year":"2024","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_18","unstructured":"Arriola, M., Chiu, J., Gokaslan, A., Kuleshov, V., Marroquin, E., Rush, A., Sahoo, S., and Schiff, Y. (2023, January 2\u20133). Segment Anything. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Paris, France."},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Shen, Z., and Jiao, R. (2024). Segment Anything Model for Medical Image Segmentation: Current Applications and Future Directions. Comput. Biol. Med., 171.","DOI":"10.1016\/j.compbiomed.2024.108238"},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"102918","DOI":"10.1016\/j.media.2023.102918","article-title":"Segment Anything Model for Medical Image Analysis: An Experimental Study","volume":"89","author":"Mazurowski","year":"2023","journal-title":"Med. Image Anal."},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Ren, S., Luzi, F., Lahrichi, S., Kassaw, K., Collins, L.M., Bradbury, K., and Malof, J.M. (2024, January 3\u20138). Segment Anything, from Space?. Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA.","DOI":"10.1109\/WACV57701.2024.00817"},{"key":"ref_22","first-page":"29914","article-title":"Segment Anything in High Quality","volume":"36","author":"Ke","year":"2023","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_23","first-page":"8815","article-title":"SAMRS: Scaling-Up Remote Sensing Segmentation Dataset with Segment Anything Model","volume":"36","author":"Wang","year":"2023","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"20","DOI":"10.32362\/2500-316X-2024-12-6-20-25","article-title":"Analysis and selection of the structure of a multiprocessor computing system according to the performance criterion","volume":"12","author":"Petushkov","year":"2024","journal-title":"Russ. Technol. J."},{"key":"ref_25","unstructured":"Liu, Z., Wang, Y., Vaidya, S., Ruehle, F., Halverson, J., Solja\u010di\u0107, M., Hou, T.Y., and Tegmark, M. (2024). KAN: Kolmogorov-Arnold Networks. arXiv."},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"5","DOI":"10.33276\/DE-2024-02-01","article-title":"Data Economy, Neural Network Training, and Multidimensional Geometry","volume":"3","author":"Kozyrev","year":"2024","journal-title":"Digit. Econ."},{"key":"ref_27","doi-asserted-by":"crossref","first-page":"93259","DOI":"10.1109\/ACCESS.2025.3572583","article-title":"Emotion Recognition Through Advanced Signal Fusion and Kolmogorov-Arnold Networks","volume":"13","author":"Thant","year":"2025","journal-title":"IEEE Access"},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"Ping, J., Xu, B., Wang, X., Zhang, W., Gao, Z., and Song, A. (2025, January 7\u201310). KAN-GCNN: EEG-Based Emotion Recognition with a Kolmogorov-Arnold Network-Enhanced Graph Convolutional Neural Network. Proceedings of the 5th International Conference on Robotics and Control Engineering, Link\u00f6ping, Sweden.","DOI":"10.1145\/3733774.3735333"},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Ghosh, S., Saha, S., and Jana, N.D. (2025, January 6\u201311). KANGAN-AVSS: Kolmogorov-Arnold Network Based Generative Adversarial Networks for Audio\u2013Visual Speech Synthesis. Proceedings of the ICASSP 2025\u2014IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Hyderabad, India.","DOI":"10.1109\/ICASSP49660.2025.10890863"},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Wang, Z., Zainal, A., Siraj, M.M., Ghaleb, F.A., Hao, X., and Han, S. (2025). An Intrusion Detection Model Based on Convolutional Kolmogorov-Arnold Networks. Sci. Rep., 15.","DOI":"10.1038\/s41598-024-85083-8"},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Zheng, J., Cao, M., and Zhang, C. (2025). ICKAN: A Deep Musical Instrument Classification Model Incorporating Kolmogorov-Arnold Network. Sci. Rep., 15.","DOI":"10.1038\/s41598-025-09493-y"},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"Alsayed, A., Li, C., Abdalsalam, M., and Fat\u2019hAlalim, A. (2025). A Hybrid Model for Arabic Character Recognition Using CNN and Kolmogorov-Arnold Networks (KANs). Multimed. Tools Appl., 1\u201324.","DOI":"10.1007\/s11042-025-20934-8"},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"160855","DOI":"10.1109\/ACCESS.2025.3605171","article-title":"KAN-MID: A Kolmogorov-Arnold Networks-Based Framework for Malicious URL and Intrusion Detection in IoT Systems","volume":"13","author":"Rizk","year":"2025","journal-title":"IEEE Access"},{"key":"ref_34","unstructured":"Pleshakova, E.S., and Gataullin, S.T. (2025). KAN-BiLSTM: Hybrid Neural Network Model with Multi-Domain Attention for Threat Analysis in Digital Space, Rospatent Federal Service on Intellectual Property. (In Russian)."},{"key":"ref_35","doi-asserted-by":"crossref","first-page":"107831","DOI":"10.1016\/j.neunet.2025.107831","article-title":"KKANs: Kurkova-Kolmogorov-Arnold Networks and Their Learning Dynamics","volume":"191","author":"Toscano","year":"2025","journal-title":"Neural Netw."},{"key":"ref_36","unstructured":"Devlin, J., Chang, M.W., Lee, K., and Toutanova, K. (2019, January 2\u20137). BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding. Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA."},{"key":"ref_37","unstructured":"Sanh, V., Debut, L., Chaumond, J., and Wolf, T. (2019). DistilBERT: A Distilled Version of BERT\u2014Smaller, Faster, Cheaper and Lighter. arXiv."},{"key":"ref_38","unstructured":"Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., and Stoyanov, V. (2019). RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv."},{"key":"ref_39","unstructured":"Yang, Z., Dai, Z., Yang, Y., Carbonell, J., Salakhutdinov, R.R., and Le, Q.V. (2019, January 8\u201314). XLNet: Generalized Autoregressive Pretraining for Language Understanding. Proceedings of the 33rd Conference on Neural Information Processing Systems (NeurIPS 2019), Vancouver, BC, Canada."},{"key":"ref_40","unstructured":"Lan, Z., Chen, M., Goodman, S., Gimpel, K., Sharma, P., and Soricut, R. (2019). ALBERT: A Lite BERT for Self-Supervised Learning of Language Representations. arXiv."},{"key":"ref_41","unstructured":"Clark, K., Luong, M.T., Le, Q.V., and Manning, C.D. (2020). ELECTRA: Pre-Training Text Encoders as Discriminators Rather Than Generators. arXiv."},{"key":"ref_42","unstructured":"He, P., Gao, J., and Chen, W. (2021). DeBERTaV3: Improving DeBERTa Using ELECTRA-Style Pre-Training with Gradient-Disentangled Embedding Sharing. arXiv."},{"key":"ref_43","doi-asserted-by":"crossref","unstructured":"Jin, Z., and Zhang, Y. (2025). A Graph Neural Network-Based Context-Aware Framework for Sentiment Analysis Classification in Chinese Microblogs. Mathematics, 13.","DOI":"10.3390\/math13060997"},{"key":"ref_44","doi-asserted-by":"crossref","unstructured":"Vaca-Rubio, C.J., Blanco, L., Pereira, R., and Caus, M. (2024). Kolmogorov-Arnold Networks (KANs) for Time Series Analysis. arXiv.","DOI":"10.1109\/GCWkshp64532.2024.11100692"},{"key":"ref_45","doi-asserted-by":"crossref","unstructured":"Chen, K.L., and Ding, J.J. (2025, January 18\u201321). Kolmogorov-Arnold Networks with Trainable Activation Functions for Data Regression and Classification. Proceedings of the 2025 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Fukuoka, Japan.","DOI":"10.1109\/ICAIIC64266.2025.10920715"},{"key":"ref_46","doi-asserted-by":"crossref","first-page":"100287","DOI":"10.1109\/ACCESS.2025.3577059","article-title":"G-KAN: Graph Kolmogorov-Arnold Network for Node Classification Using Contrastive Learning","volume":"13","author":"Yuan","year":"2025","journal-title":"IEEE Access"}],"container-title":["Algorithms"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1999-4893\/18\/12\/735\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,24]],"date-time":"2025-11-24T15:17:27Z","timestamp":1763997447000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1999-4893\/18\/12\/735"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,23]]},"references-count":46,"journal-issue":{"issue":"12","published-online":{"date-parts":[[2025,12]]}},"alternative-id":["a18120735"],"URL":"https:\/\/doi.org\/10.3390\/a18120735","relation":{},"ISSN":["1999-4893"],"issn-type":[{"value":"1999-4893","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,11,23]]}}}