{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,26]],"date-time":"2026-03-26T04:06:10Z","timestamp":1774497970469,"version":"3.50.1"},"reference-count":87,"publisher":"MDPI AG","issue":"3","license":[{"start":{"date-parts":[[2025,2,20]],"date-time":"2025-02-20T00:00:00Z","timestamp":1740009600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62376033"],"award-info":[{"award-number":["62376033"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["BDCC"],"abstract":"<jats:p>Large language models (LLMs) have demonstrated remarkable capabilities in text generation, which also raise numerous concerns about their potential misuse, especially in educational exercises and academic writing. Accurately identifying and tracing the origins of LLM-generated content is crucial for accountability and transparency, ensuring the responsible use of LLMs in educational and academic environments. Previous methods utilize binary classifiers to discriminate whether a piece of text was written by a human or generated by a specific LLM or employ multi-class classifiers to trace the source LLM from a fixed set. These methods, however, are restricted to one or several pre-specified LLMs and cannot generalize to new LLMs, which are continually emerging. This study formulates source LLM tracing in a class-incremental learning (CIL) fashion, where new LLMs continually emerge, and a model incrementally learns to identify new LLMs without forgetting old ones. A training-free continual learning method is further devised for the task, the idea of which is to continually extract prototypes for emerging LLMs, using a frozen encoder, and then to perform origin tracing via prototype matching after a delicate decorrelation process. For evaluation, two datasets are constructed, one in English and one in Chinese. These datasets simulate a scenario where six LLMs emerge over time and are used to generate student essays, and an LLM detector has to incrementally expand its recognition scope as new LLMs appear. Experimental results show that the proposed method achieves an average accuracy of 97.04% on the English dataset and 91.23% on the Chinese dataset. These results validate the feasibility of continual origin tracing of LLM-generated text and verify its effectiveness in detecting cheating in student coursework.<\/jats:p>","DOI":"10.3390\/bdcc9030050","type":"journal-article","created":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T03:47:46Z","timestamp":1740109666000},"page":"50","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["On Continually Tracing Origins of LLM-Generated Text and Its Application in Detecting Cheating in Student Coursework"],"prefix":"10.3390","volume":"9","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6102-3407","authenticated-orcid":false,"given":"Quan","family":"Wang","sequence":"first","affiliation":[{"name":"School of Computer Science, Beijing University of Posts and Telecommunications, Beijing 100876, China"}]},{"given":"Haoran","family":"Li","sequence":"additional","affiliation":[{"name":"School of Computer Science, Beijing University of Posts and Telecommunications, Beijing 100876, China"}]}],"member":"1968","published-online":{"date-parts":[[2025,2,20]]},"reference":[{"key":"ref_1","unstructured":"OpenAI (2022, November 30). ChatGPT: Optimizing Language Models for Dialogue. Available online: https:\/\/chatgpt.r4wand.eu.org."},{"key":"ref_2","unstructured":"OpenAI (2023). GPT-4 Technical Report. arXiv."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"135","DOI":"10.3390\/nanomanufacturing3020009","article-title":"Good Practices for Scientific Article Writing with ChatGPT and Other Artificial Intelligence Language Models","volume":"3","year":"2023","journal-title":"Nanomanufacturing"},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"2855","DOI":"10.1007\/s11280-023-01166-y","article-title":"An Empirical Study of Pre-Trained Language Models in Simple Knowledge Graph Question Answering","volume":"26","author":"Hu","year":"2023","journal-title":"World Wide Web"},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"1813","DOI":"10.1108\/APJML-10-2023-994","article-title":"Written by ChatGPT, Illustrated by Midjourney: Generative AI for Content Marketing","volume":"35","author":"Wahid","year":"2023","journal-title":"Asia Pac. J. Mark. Logist."},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Jiang, B., Tan, Z., Nirmal, A., and Liu, H. (2024, January 18\u201320). Disinformation Detection: An Evolving Challenge in the Age of LLMs. Proceedings of the 2024 SIAM International Conference on Data Mining, Houston, TX, USA.","DOI":"10.1137\/1.9781611978032.50"},{"key":"ref_7","unstructured":"Panda, A., Choquette-Choo, C.A., Zhang, Z., Yang, Y., and Mittal, P. (2024). Teach LLMs to Phish: Stealing Private Information from Language Models. arXiv."},{"key":"ref_8","first-page":"7","article-title":"Academic Integrity Considerations of AI Large Language Models in the Post-Pandemic Era: ChatGPT and Beyond","volume":"20","author":"Perkins","year":"2023","journal-title":"J. Univ. Teach. Learn. Pract."},{"key":"ref_9","doi-asserted-by":"crossref","unstructured":"Lo, C.K. (2023). What is the Impact of ChatGPT on Education? A Rapid Review of the Literature. Educ. Sci., 13.","DOI":"10.3390\/educsci13040410"},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Rahman, M.M., and Watanobe, Y. (2023). ChatGPT for Education and Research: Opportunities, Threats, and Strategies. Appl. Sci., 13.","DOI":"10.20944\/preprints202303.0473.v1"},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"50","DOI":"10.1145\/3624725","article-title":"The Science of Detecting LLM-Generated Text","volume":"67","author":"Tang","year":"2024","journal-title":"Commun. ACM"},{"key":"ref_12","unstructured":"Gehrmann, S., Strobelt, H., and Rush, A. (August, January 28). GLTR: Statistical Detection and Visualization of Generated Text. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: System Demonstrations, Florence, Italy."},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Zhu, B., Yuan, L., Cui, G., Chen, Y., Fu, C., He, B., Deng, Y., Liu, Z., Sun, M., and Gu, M. (2023, January 6\u201310). Beat LLMs at Their Own Game: Zero-Shot LLM-Generated Text Detection via Querying ChatGPT. Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, Singapore.","DOI":"10.18653\/v1\/2023.emnlp-main.463"},{"key":"ref_14","unstructured":"Mitchell, E., Lee, Y., Khazatsky, A., Manning, C.D., and Finn, C. (2023, January 23\u201329). DetectGPT: Zero-Shot Machine-Generated Text Detection Using Probability Curvature. Proceedings of the 40th International Conference on Machine Learning, Honolulu, HI, USA."},{"key":"ref_15","unstructured":"Bao, G., Zhao, Y., Teng, Z., Yang, L., and Zhang, Y. (2024, January 7\u201311). Fast-DetectGPT: Efficient Zero-Shot Detection of Machine-Generated Text via Conditional Probability Curvature. Proceedings of the 12th International Conference on Learning Representations, Vienna, Austria."},{"key":"ref_16","unstructured":"Zellers, R., Holtzman, A., Rashkin, H., Bisk, Y., Farhadi, A., Roesner, F., and Choi, Y. (2019, January 8\u201314). Defending Against Neural Fake News. Proceedings of the 33rd Conference on Neural Information Processing Systems, Vancouver, BC, Canada."},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Ippolito, D., Duckworth, D., Callison-Burch, C., and Eck, D. (2020, January 5\u201310). Automatic Detection of Generated Text is Easiest when Humans are Fooled. Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online.","DOI":"10.18653\/v1\/2020.acl-main.164"},{"key":"ref_18","doi-asserted-by":"crossref","unstructured":"Zhong, W., Tang, D., Xu, Z., Wang, R., Duan, N., Zhou, M., Wang, J., and Yin, J. (2020, January 16\u201320). Neural Deepfake Detection with Factual Structure of Text. Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, Online.","DOI":"10.18653\/v1\/2020.emnlp-main.193"},{"key":"ref_19","unstructured":"Wang, Q., Zhang, L., Guo, Z., and Mao, Z. (2024, January 20\u201325). IDEATE: Detecting AI-Generated Text Using Internal and External Factual Structures. Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation, Turin, Italy."},{"key":"ref_20","unstructured":"Li, L., Wang, P., Ren, K., Sun, T., and Qiu, X. (2023). Origin Tracing and Detecting of LLMs. arXiv."},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Wu, K., Pang, L., Shen, H., Cheng, X., and Chua, T.S. (2023, January 6\u201310). LLMDet: A Third Party Large Language Models Generated Text Detection Tool. Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2023, Singapore.","DOI":"10.18653\/v1\/2023.findings-emnlp.139"},{"key":"ref_22","unstructured":"Soto, R.A.R., Koch, K., Khan, A., Chen, B.Y., Bishop, M., and Andrews, N. (2024, January 7\u201311). Few-Shot Detection of Machine-Generated Text using Style Representations. Proceedings of the 12th International Conference on Learning Representations, Vienna, Austria."},{"key":"ref_23","first-page":"3366","article-title":"A Continual Learning Survey: Defying Forgetting in Classification Tasks","volume":"44","author":"Aljundi","year":"2022","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"5513","DOI":"10.1109\/TPAMI.2022.3213473","article-title":"Class-Incremental Learning: Survey and Performance Evaluation on Image Classification","volume":"45","author":"Masana","year":"2023","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Mensink, T., Verbeek, J., Perronnin, F., and Csurka, G. (2012, January 7\u201313). Metric Learning for Large Scale Image Classification: Generalizing to New Classes at Near-Zero Cost. Proceedings of the 12th European Conference on Computer Vision, Florence, Italy.","DOI":"10.1007\/978-3-642-33709-3_35"},{"key":"ref_26","unstructured":"Snell, J., Swersky, K., and Zemel, R. (2017, January 4\u20139). Prototypical Networks for Few-shot Learning. Proceedings of the 31st Conference on Neural Information Processing Systems, Long Beach, CA, USA."},{"key":"ref_27","unstructured":"Devlin, J., Chang, M.W., Lee, K., and Toutanova, K. (2019, January 2\u20137). BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding. Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA."},{"key":"ref_28","unstructured":"Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L., and Stoyanov, V. (2019). RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv."},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Wang, Y., Mansurov, J., Ivanov, P., Su, J., Shelmanov, A., Tsvigun, A., Mohammed Afzal, O., Mahmoud, T., Puccetti, G., and Arnold, T. (2024, January 11\u201316). M4GT-Bench: Evaluation Benchmark for Black-Box Machine-Generated Text Detection. Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics, Bangkok, Thailand.","DOI":"10.18653\/v1\/2024.acl-long.218"},{"key":"ref_30","unstructured":"Gomez, A., and Zhang, I. (2022, June 14). Introducing: Cohere For AI. Available online: https:\/\/cohere.com\/blog\/introducing-cohere-for-ai."},{"key":"ref_31","unstructured":"(2022). BigScience. BLOOM: A 176B-Parameter Open-Access Multilingual Language Model. arXiv."},{"key":"ref_32","unstructured":"Conover, M., Hayes, M., Mathur, A., Xie, J., Wan, J., Shah, S., Ghodsi, A., Wendell, P., Zaharia, M., and Xin, R. (2023, April 12). Free Dolly: Introducing the World\u2019s First Truly Open Instruction-Tuned LLM. Available online: https:\/\/www.databricks.com\/blog\/2023\/04\/12\/dolly-first-open-commercially-viable-instruction-tuned-llm."},{"key":"ref_33","unstructured":"Li, Y., Zhang, Y., Zhao, Z., Shen, L., Liu, W., Mao, W., and Zhang, H. (2022, January 12\u201317). CSL: A Large-scale Chinese Scientific Literature Dataset. Proceedings of the 29th International Conference on Computational Linguistics, Gyeongju, Republic of Korea."},{"key":"ref_34","unstructured":"OpenAI (2024, May 13). Hello GPT-4o. Available online: https:\/\/openai.com\/index\/hello-gpt-4o\/."},{"key":"ref_35","unstructured":"Baidu (2023, March 24). ERNIE Bot: Baidu\u2019s Knowledge-Enhanced Large Language Model Built on Full AI Stack Technology. Available online: http:\/\/research.baidu.com\/Blog\/index-view?id=183."},{"key":"ref_36","unstructured":"ByteDance (2024, September 23). ByteDance Doubao LLM Offers Trillions of Tokens for Free Use. Available online: https:\/\/team.doubao.com\/en\/blog\/bytedance-doubao-llm-offers-trillions-of-tokens-for-free-use?view_from=blog."},{"key":"ref_37","unstructured":"Alibaba (2023). Qwen Technical Report. arXiv."},{"key":"ref_38","unstructured":"Frieder, S., Pinchetti, L., Griffiths, R.R., Salvatori, T., Lukasiewicz, T., Petersen, P., and Berner, J. (2023, January 10\u201316). Mathematical Capabilities of ChatGPT. Proceedings of the 37th Conference on Neural Information Processing Systems, New Orleans, LA, USA."},{"key":"ref_39","doi-asserted-by":"crossref","unstructured":"Kung, T.H., Cheatham, M., Medenilla, A., Sillos, C., De Leon, L., Elepa\u00f1o, C., Madriaga, M., Aggabao, R., Diaz-Candido, G., and Maningo, J. (2023). Performance of ChatGPT on USMLE: Potential for AI-Assisted Medical Education Using Large Language Models. PLoS Digit. Health, 2.","DOI":"10.1371\/journal.pdig.0000198"},{"key":"ref_40","doi-asserted-by":"crossref","first-page":"1092","DOI":"10.1126\/science.abq1158","article-title":"Competition-Level Code Generation with AlphaCode","volume":"378","author":"Li","year":"2022","journal-title":"Science"},{"key":"ref_41","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3672459","article-title":"Self-Collaboration Code Generation via ChatGPT","volume":"33","author":"Dong","year":"2024","journal-title":"ACM Trans. Softw. Eng. Methodol."},{"key":"ref_42","doi-asserted-by":"crossref","unstructured":"Sobania, D., Briesch, M., Hanna, C., and Petke, J. (2023, January 16\u201316). An Analysis of the Automatic Bug Fixing Performance of ChatGPT. Proceedings of the 2023 IEEE\/ACM International Workshop on Automated Program Repair, Melbourne, Australia.","DOI":"10.1109\/APR59189.2023.00012"},{"key":"ref_43","doi-asserted-by":"crossref","unstructured":"Jalil, S., Rafi, S., LaToza, T.D., Moran, K., and Lam, W. (2023, January 16\u201320). ChatGPT and Software Testing Education: Promises & Perils. Proceedings of the 2023 IEEE International Conference on Software Testing, Verification and Validation Workshops, Dublin, Ireland.","DOI":"10.1109\/ICSTW58534.2023.00078"},{"key":"ref_44","first-page":"387","article-title":"ChatGPT Goes to Law School","volume":"71","author":"Choi","year":"2021","journal-title":"J. Leg. Educ."},{"key":"ref_45","doi-asserted-by":"crossref","first-page":"103662","DOI":"10.1016\/j.frl.2023.103662","article-title":"ChatGPT for (Finance) Research: The Bananarama Conjecture","volume":"53","author":"Dowling","year":"2023","journal-title":"Financ. Res. Lett."},{"key":"ref_46","doi-asserted-by":"crossref","unstructured":"Malinka, K., Peres\u00edni, M., Firc, A., Hujn\u00e1k, O., and Janus, F. (2023, January 8\u201312). On the Educational Impact of ChatGPT: Is Artificial Intelligence Ready to Obtain a University Degree?. Proceedings of the 2023 Conference on Innovation and Technology in Computer Science Education, Turku, Finland.","DOI":"10.1145\/3587102.3588827"},{"key":"ref_47","doi-asserted-by":"crossref","unstructured":"Susnjak, T., and McIntosh, T.R. (2024). ChatGPT: The End of Online Exam Integrity?. Educ. Sci., 14.","DOI":"10.3390\/educsci14060656"},{"key":"ref_48","doi-asserted-by":"crossref","unstructured":"Borji, A. (2023). A Categorical Archive of ChatGPT Failures. arXiv.","DOI":"10.21203\/rs.3.rs-2895792\/v1"},{"key":"ref_49","unstructured":"Rawte, V., Sheth, A., and Das, A. (2023). A Survey of Hallucination in Large Foundation Models. arXiv."},{"key":"ref_50","doi-asserted-by":"crossref","first-page":"90","DOI":"10.1111\/bjet.13370","article-title":"Practical and Ethical Challenges of Large Language Models in Education: A Systematic Scoping Review","volume":"55","author":"Yan","year":"2024","journal-title":"Br. J. Educ. Technol."},{"key":"ref_51","doi-asserted-by":"crossref","unstructured":"Li, Y., Li, Q., Cui, L., Bi, W., Wang, Z., Wang, L., Yang, L., Shi, S., and Zhang, Y. (2024, January 11\u201316). MAGE: Machine-Generated Text Detection in the Wild. Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics, Bangkok, Thailand.","DOI":"10.18653\/v1\/2024.acl-long.3"},{"key":"ref_52","unstructured":"Solaiman, I., Brundage, M., Clark, J., Askell, A., Herbert-Voss, A., Wu, J., Radford, A., Krueger, G., Kim, J.W., and Kreps, S. (2019). Release Strategies and the Social Impacts of Language Models. arXiv."},{"key":"ref_53","doi-asserted-by":"crossref","unstructured":"Verma, V., Fleisig, E., Tomlin, N., and Klein, D. (2024, January 16\u201321). Ghostbuster: Detecting Text Ghostwritten by Large Language Models. Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Mexico City, Mexico.","DOI":"10.18653\/v1\/2024.naacl-long.95"},{"key":"ref_54","doi-asserted-by":"crossref","unstructured":"Uchendu, A., Le, T., Shu, K., and Lee, D. (2020, January 16\u201320). Authorship Attribution for Neural Text Generation. Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, Online.","DOI":"10.18653\/v1\/2020.emnlp-main.673"},{"key":"ref_55","doi-asserted-by":"crossref","unstructured":"Bhattacharjee, A., Kumarage, T., Moraffah, R., and Liu, H. (2023, January 1\u20134). ConDA: Contrastive Domain Adaptation for AI-generated Text Detection. Proceedings of the 13th International Joint Conference on Natural Language Processing and the 3rd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics, Nusa Dua, Bali.","DOI":"10.18653\/v1\/2023.ijcnlp-main.40"},{"key":"ref_56","doi-asserted-by":"crossref","unstructured":"Rodriguez, J.D., Hay, T., Gros, D., Shamsi, Z., and Srinivasan, R. (2022, January 10\u201315). Cross-Domain Detection of GPT-2-Generated Technical Text. Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Seattle, WA, USA.","DOI":"10.18653\/v1\/2022.naacl-main.88"},{"key":"ref_57","doi-asserted-by":"crossref","unstructured":"Liu, X., Zhang, Z., Wang, Y., Pu, H., Lan, Y., and Shen, C. (2023, January 6\u201310). CoCo: Coherence-Enhanced Machine-Generated Text Detection under Low Resource with Contrastive Learning. Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, Singapore.","DOI":"10.18653\/v1\/2023.emnlp-main.1005"},{"key":"ref_58","doi-asserted-by":"crossref","unstructured":"Liu, S., Liu, X., Wang, Y., Cheng, Z., Li, C., Zhang, Z., Lan, Y., and Shen, C. (2024, January 11\u201316). Does DetectGPT Fully Utilize Perturbation? Bridging Selective Perturbation to Fine-Tuned Contrastive Learning Detector would be Better. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics, Bangkok, Thailand.","DOI":"10.18653\/v1\/2024.acl-long.103"},{"key":"ref_59","doi-asserted-by":"crossref","unstructured":"Hashimoto, T.B., Zhang, H., and Liang, P. (2019, January 2\u20137). Unifying Human and Statistical Evaluation for Natural Language Generation. Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA.","DOI":"10.18653\/v1\/N19-1169"},{"key":"ref_60","unstructured":"Lavergne, T., Urvoy, T., and Yvon, F. (2008, January 21\u201325). Detecting Fake Content with Relative Entropy Scoring. Proceedings of the 2008 International Conference on Uncovering Plagiarism, Authorship and Social Software Misuse, Patras, Greece."},{"key":"ref_61","unstructured":"Yang, X., Cheng, W., Wu, Y., Petzold, L.R., Wang, W.Y., and Chen, H. (2024, January 7\u201311). DNA-GPT: Divergent N-Gram Analysis for Training-Free Detection of GPT-Generated Text. Proceedings of the 12th International Conference on Learning Representations, Vienna, Austria."},{"key":"ref_62","doi-asserted-by":"crossref","unstructured":"Chen, Y., Kang, H., Zhai, V., Li, L., Singh, R., and Raj, B. (2023, January 6\u201310). Token Prediction as Implicit Classification to Identify LLM-Generated Text. Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, Singapore.","DOI":"10.18653\/v1\/2023.emnlp-main.810"},{"key":"ref_63","doi-asserted-by":"crossref","first-page":"1185","DOI":"10.1038\/s42256-022-00568-3","article-title":"Three Types of Incremental Learning","volume":"4","author":"Tuytelaars","year":"2022","journal-title":"Nat. Mach. Intell."},{"key":"ref_64","doi-asserted-by":"crossref","first-page":"5362","DOI":"10.1109\/TPAMI.2024.3367329","article-title":"A Comprehensive Survey of Continual Learning: Theory, Method and Application","volume":"46","author":"Wang","year":"2024","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_65","unstructured":"Aljundi, R., Lin, M., Goujaud, B., and Bengio, Y. (2019, January 8\u201314). Gradient based Sample Selection for Online Continual Learning. Proceedings of the 33rd Conference on Neural Information Processing Systems, Vancouver, BC, Canada."},{"key":"ref_66","unstructured":"Chaudhry, A., Ranzato, M., Rohrbach, M., and Elhoseiny, M. (2019, January 6\u20139). Efficient Lifelong Learning with A-GEM. Proceedings of the 7th International Conference on Learning Representations, New Orleans, LA, USA."},{"key":"ref_67","doi-asserted-by":"crossref","first-page":"3521","DOI":"10.1073\/pnas.1611835114","article-title":"Overcoming Catastrophic Forgetting in Neural Networks","volume":"114","author":"Kirkpatrick","year":"2017","journal-title":"Proc. Natl. Acad. Sci. USA"},{"key":"ref_68","doi-asserted-by":"crossref","first-page":"2935","DOI":"10.1109\/TPAMI.2017.2773081","article-title":"Learning without Forgetting","volume":"40","author":"Li","year":"2018","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_69","doi-asserted-by":"crossref","unstructured":"Hu, Z., Li, Y., Lyu, J., Gao, D., and Vasconcelos, N. (2023, January 18\u201322). Dense Network Expansion for Class Incremental Learning. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada.","DOI":"10.1109\/CVPR52729.2023.01141"},{"key":"ref_70","doi-asserted-by":"crossref","unstructured":"Zhou, D.W., Sun, H.L., Ye, H.J., and Zhan, D.C. (2024, January 17\u201321). Expandable Subspace Ensemble for Pre-Trained Model-Based Class-Incremental Learning. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA.","DOI":"10.1109\/CVPR52733.2024.02223"},{"key":"ref_71","unstructured":"McDonnell, M.D., Gong, D., Parvaneh, A., Abbasnejad, E., and van den Hengel, A. (2023, January 10\u201316). RanPAC: Random Projections and Pre-Trained Models for Continual Learning. Proceedings of the 37th Conference on Neural Information Processing Systems, New Orleans, LA, USA."},{"key":"ref_72","doi-asserted-by":"crossref","unstructured":"Zhou, D.W., Cai, Z.W., Ye, H.J., Zhan, D.C., and Liu, Z. (2024). Revisiting Class-Incremental Learning with Pre-Trained Models: Generalizability and Adaptivity are All You Need. Int. J. Comput. Vis.","DOI":"10.1007\/s11263-024-02218-0"},{"key":"ref_73","doi-asserted-by":"crossref","unstructured":"McDonnell, M.D., McKilliam, R.G., and de Chazal, P. (2016, January 24\u201329). On the Importance of Pair-wise Feature Correlations for Image Classification. Proceedings of the 2016 International Joint Conference on Neural Networks, Vancouver, BC, Canada.","DOI":"10.1109\/IJCNN.2016.7727483"},{"key":"ref_74","doi-asserted-by":"crossref","first-page":"905","DOI":"10.1109\/TSMCB.2005.847744","article-title":"Incremental Linear Discriminant Analysis for Classification of Data Streams","volume":"35","author":"Pang","year":"2005","journal-title":"IEEE Trans. Syst. Man, Cybern. Part B (Cybernetics)"},{"key":"ref_75","doi-asserted-by":"crossref","unstructured":"Hayes, T.L., and Kanan, C. (2020, January 14\u201319). Lifelong Machine Learning with Deep Streaming Linear Discriminant Analysis. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Online.","DOI":"10.1109\/CVPRW50498.2020.00118"},{"key":"ref_76","doi-asserted-by":"crossref","unstructured":"Panos, A., Kobe, Y., Reino, D.O., Aljundi, R., and Turner, R.E. (2023, January 2\u20136). First Session Adaptation: A Strong Replay-Free Baseline for Class-Incremental Learning. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Paris, France.","DOI":"10.1109\/ICCV51070.2023.01725"},{"key":"ref_77","unstructured":"Murphy, K.P. (2012). Machine Learning: A Probabilistic Perspective, MIT Press."},{"key":"ref_78","unstructured":"Hinton, G., Vinyals, O., and Dean, J. (2015). Distilling the Knowledge in a Neural Network. arXiv."},{"key":"ref_79","unstructured":"Houlsby, N., Giurgiu, A., Jastrzebski, S., Morrone, B., de Laroussilhe, Q., Gesmundo, A., Attariyan, M., and Gelly, S. (2019, January 10\u201315). Parameter-Efficient Transfer Learning for NLP. Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA."},{"key":"ref_80","doi-asserted-by":"crossref","first-page":"3504","DOI":"10.1109\/TASLP.2021.3124365","article-title":"Pre-Training with Whole Word Masking for Chinese BERT","volume":"29","author":"Cui","year":"2021","journal-title":"IEEE\/ACM Trans. Audio Speech Lang. Process."},{"key":"ref_81","unstructured":"Loshchilov, I., and Hutter, F. (2019, January 6\u20139). Decoupled Weight Decay Regularization. Proceedings of the 7th International Conference on Learning Representations, New Orleans, LA, USA."},{"key":"ref_82","doi-asserted-by":"crossref","unstructured":"Ma, S., and Wang, Q. (2024, January 12\u201316). Zero-Shot Detection of LLM-Generated Text Using Token Cohesiveness. Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, Miami, FL, USA.","DOI":"10.18653\/v1\/2024.emnlp-main.971"},{"key":"ref_83","doi-asserted-by":"crossref","unstructured":"Wang, P., Li, L., Ren, K., Jiang, B., Zhang, D., and Qiu, X. (2023, January 6\u201310). SeqXGPT: Sentence-Level AI-Generated Text Detection. Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, Singapore.","DOI":"10.18653\/v1\/2023.emnlp-main.73"},{"key":"ref_84","doi-asserted-by":"crossref","unstructured":"Peng, X., Zhou, Y., He, B., Sun, L., and Sun, Y. (2023, January 6\u201310). Hidding the Ghostwriters: An Adversarial Evaluation of AI-Generated Student Essay Detection. Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, Singapore.","DOI":"10.18653\/v1\/2023.emnlp-main.644"},{"key":"ref_85","unstructured":"Krishna, K., Song, Y., Karpinska, M., Wieting, J., and Iyyer, M. (2023, January 10\u201316). Paraphrasing Evades Detectors of AI-Generated Text, but Retrieval is an Effective Defense. Proceedings of the 37th Conference on Neural Information Processing Systems, New Orleans, LA, USA."},{"key":"ref_86","unstructured":"Hu, X., Chen, P.Y., and Ho, T.Y. (2023, January 10\u201316). RADAR: Robust AI-Text Detection via Adversarial Learning. Proceedings of the 37th Conference on Neural Information Processing Systems, New Orleans, LA, USA."},{"key":"ref_87","doi-asserted-by":"crossref","first-page":"102869","DOI":"10.1016\/j.compcom.2024.102869","article-title":"Detecting the Anomalies: Exploring Implications of Qualitative Research in Identifying AI-Generated Text for AI-Assisted Composition Instruction","volume":"73","author":"Garib","year":"2024","journal-title":"Comput. Compos."}],"container-title":["Big Data and Cognitive Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2504-2289\/9\/3\/50\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,9]],"date-time":"2025-10-09T16:39:12Z","timestamp":1760027952000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2504-2289\/9\/3\/50"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,2,20]]},"references-count":87,"journal-issue":{"issue":"3","published-online":{"date-parts":[[2025,3]]}},"alternative-id":["bdcc9030050"],"URL":"https:\/\/doi.org\/10.3390\/bdcc9030050","relation":{},"ISSN":["2504-2289"],"issn-type":[{"value":"2504-2289","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,2,20]]}}}