{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T14:21:28Z","timestamp":1773325288913,"version":"3.50.1"},"publisher-location":"Singapore","reference-count":26,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819570713","type":"print"},{"value":"9789819570720","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-7072-0_9","type":"book-chapter","created":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T00:09:01Z","timestamp":1773274141000},"page":"132-148","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Multi-Stage Variance-Controlled Gradient Updates: Toward Robust Continual Learning"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-9587-555X","authenticated-orcid":false,"given":"Ji","family":"Feng","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0005-6682-6731","authenticated-orcid":false,"given":"Qingjun","family":"Zhang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0008-5049-7672","authenticated-orcid":false,"given":"Yongqiang","family":"Xu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0005-6300-2251","authenticated-orcid":false,"given":"Peilin","family":"Li","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8582-4302","authenticated-orcid":false,"given":"Degang","family":"Yang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,4,1]]},"reference":[{"key":"9_CR1","doi-asserted-by":"crossref","unstructured":"Wang, J., Dong, D., Shou, L., Chen, K., Chen, G.: Effective continual learning for text classification with lightweight snapshots. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a037, pp. 10122\u201310130 (2023)","DOI":"10.1609\/aaai.v37i8.26206"},{"key":"9_CR2","doi-asserted-by":"crossref","unstructured":"McCloskey, M., Cohen, N.J.: Catastrophic interference in connectionist networks: the sequential learning problem. In: Psychology of Learning and Motivation, vol.\u00a024, pp. 109\u2013165. Elsevier (1989)","DOI":"10.1016\/S0079-7421(08)60536-8"},{"key":"9_CR3","doi-asserted-by":"publisher","unstructured":"Yang, H., Zhang, Y., Xu, J., Lu, H., Heng, P., Lam, W.: Unveiling the generalization power of fine-tuned large language models. In: Duh, K., G\u00f3mez-Adorno, H., Bethard, S. (eds.) Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), NAACL 2024, Mexico City, Mexico, June 16-21, 2024, pp. 884\u2013899. Association for Computational Linguistics (2024). https:\/\/doi.org\/10.18653\/V1\/2024.NAACL-LONG.51, https:\/\/doi.org\/10.18653\/v1\/2024.naacl-long.51","DOI":"10.18653\/V1\/2024.NAACL-LONG.51"},{"key":"9_CR4","doi-asserted-by":"publisher","unstructured":"Shi, H., Xu, Z., Wang, H., Qin, W., Wang, W., Wang, Y., Wang, H.: Continual learning of large language models: a comprehensive survey. CoRR abs\/2404.16789 (2024). https:\/\/doi.org\/10.48550\/ARXIV.2404.16789","DOI":"10.48550\/ARXIV.2404.16789"},{"key":"9_CR5","doi-asserted-by":"publisher","unstructured":"Wu, T., Luo, L., Li, Y., Pan, S., Vu, T., Haffari, G.: Continual learning for large language models: a survey. CoRR abs\/2402.01364 (2024). https:\/\/doi.org\/10.48550\/ARXIV.2402.01364","DOI":"10.48550\/ARXIV.2402.01364"},{"key":"9_CR6","doi-asserted-by":"publisher","unstructured":"Scialom, T., Chakrabarty, T., Muresan, S.: Fine-tuned language models are continual learners. In: Goldberg, Y., Kozareva, Z., Zhang, Y. (eds.) Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, Abu Dhabi, United Arab Emirates, December 7-11, 2022, pp. 6107\u20136122. Association for Computational Linguistics (2022). https:\/\/doi.org\/10.18653\/V1\/2022.EMNLP-MAIN.410, https:\/\/doi.org\/10.18653\/v1\/2022.emnlp-main.410","DOI":"10.18653\/V1\/2022.EMNLP-MAIN.410"},{"key":"9_CR7","doi-asserted-by":"crossref","unstructured":"Wang, Z., Liu, Y., Ji, T., Wang, X., Wu, Y., Jiang, C., Chao, Y., Han, Z., Wang, L., Shao, X., et\u00a0al.: Rehearsal-free continual language learning via efficient parameter isolation. In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 10933\u201310946 (2023)","DOI":"10.18653\/v1\/2023.acl-long.612"},{"key":"9_CR8","doi-asserted-by":"publisher","unstructured":"Gururangan, S., Lewis, M., Holtzman, A., Smith, N.A., Zettlemoyer, L.: Demix layers: disentangling domains for modular language modeling. In: Carpuat, M., de\u00a0Marneffe, M., Ru\u00edz, I.V.M. (eds.) Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL 2022, Seattle, WA, United States, July 10-15, 2022, pp. 5557\u20135576. Association for Computational Linguistics (2022). https:\/\/doi.org\/10.18653\/V1\/2022.NAACL-MAIN.407, https:\/\/doi.org\/10.18653\/v1\/2022.naacl-main.407","DOI":"10.18653\/V1\/2022.NAACL-MAIN.407"},{"issue":"2","key":"9_CR9","first-page":"3","volume":"1","author":"EJ Hu","year":"2022","unstructured":"Hu, E.J., Shen, Y., Wallis, P., Allen-Zhu, Z., Li, Y., Wang, S., Wang, L., Chen, W., et al.: Lora: Low-rank adaptation of large language models. ICLR 1(2), 3 (2022)","journal-title":"ICLR"},{"key":"9_CR10","doi-asserted-by":"publisher","unstructured":"Zheng, J., Qiu, S., Ma, Q.: Learn or recall? revisiting incremental learning with pre-trained language models. In: Ku, L., Martins, A., Srikumar, V. (eds.) Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2024, Bangkok, Thailand, August 11-16, 2024, pp. 14848\u201314877. Association for Computational Linguistics (2024). https:\/\/doi.org\/10.18653\/V1\/2024.ACL-LONG.794, https:\/\/doi.org\/10.18653\/v1\/2024.acl-long.794","DOI":"10.18653\/V1\/2024.ACL-LONG.794"},{"key":"9_CR11","unstructured":"Zhu, D., et al.: Model tailor: Mitigating catastrophic forgetting in multi-modal large language models. In: Forty-first International Conference on Machine Learning, ICML 2024, Vienna, Austria, July 21-27, 2024. OpenReview.net (2024). https:\/\/openreview.net\/forum?id=piujJIF3zs"},{"key":"9_CR12","doi-asserted-by":"publisher","unstructured":"Touvron, H., et al.: Llama: open and efficient foundation language models. CoRR abs\/2302.13971 (2023). https:\/\/doi.org\/10.48550\/ARXIV.2302.13971, https:\/\/doi.org\/10.48550\/arXiv.2302.13971","DOI":"10.48550\/ARXIV.2302.13971"},{"key":"9_CR13","doi-asserted-by":"crossref","unstructured":"Du, W., et al.: Unlocking continual learning abilities in language models. In: Al-Onaizan, Y., Bansal, M., Chen, Y. (eds.) Findings of the Association for Computational Linguistics: EMNLP 2024, Miami, Florida, USA, November 12-16, 2024, pp. 6503\u20136522. Association for Computational Linguistics (2024). https:\/\/aclanthology.org\/2024.findings-emnlp.379","DOI":"10.18653\/v1\/2024.findings-emnlp.379"},{"key":"9_CR14","unstructured":"Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21, 140:1\u2013140:67 (2020). https:\/\/jmlr.org\/papers\/v21\/20-074.html"},{"key":"9_CR15","unstructured":"Grattafiori, A., et\u00a0al.: The llama 3 herd of models. arXiv preprint arXiv:2407.21783 (2024)"},{"key":"9_CR16","doi-asserted-by":"publisher","unstructured":"Jiang, T., et al.: Mora: High-rank updating for parameter-efficient fine-tuning. CoRR abs\/2405.12130 (2024). https:\/\/doi.org\/10.48550\/ARXIV.2405.12130","DOI":"10.48550\/ARXIV.2405.12130"},{"key":"9_CR17","doi-asserted-by":"publisher","unstructured":"Wang, X., et al.: Orthogonal subspace learning for language model continual learning. In: Bouamor, H., Pino, J., Bali, K. (eds.) Findings of the Association for Computational Linguistics: EMNLP 2023, Singapore, December 6-10, 2023, pp. 10658\u201310671. Association for Computational Linguistics (2023). https:\/\/doi.org\/10.18653\/V1\/2023.FINDINGS-EMNLP.715","DOI":"10.18653\/V1\/2023.FINDINGS-EMNLP.715"},{"key":"9_CR18","unstructured":"Qin, C., Joty, S.R.: LFPT5: A unified framework for lifelong few-shot language learning based on prompt tuning of T5. In: The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. OpenReview.net (2022). https:\/\/openreview.net\/forum?id=HCRVf71PMF"},{"key":"9_CR19","unstructured":"Zhang, X., Zhao, J.J., LeCun, Y.: Character-level convolutional networks for text classification. In: Cortes, C., Lawrence, N.D., Lee, D.D., Sugiyama, M., Garnett, R. (eds.) Advances in Neural Information Processing Systems 28: Annual Conference on Neural Information Processing Systems 2015, December 7-12, 2015, Montreal, Quebec, Canada, pp. 649\u2013657 (2015). https:\/\/proceedings.neurips.cc\/paper\/2015\/hash\/250cf8b51c773f3f8dc8b4be867a9a02-Abstract.html"},{"key":"9_CR20","unstructured":"Razdaibiedina, A., Mao, Y., Hou, R., Khabsa, M., Lewis, M., Almahairi, A.: Progressive prompts: Continual learning for language models. In: The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023. OpenReview.net (2023). https:\/\/openreview.net\/forum?id=UJTgQBc91_"},{"key":"9_CR21","unstructured":"Wei, Y., Wang, Z., Liu, J., Ding, Y., Zhang, L.: Magicoder: Empowering code generation with oss-instruct. In: Forty-first International Conference on Machine Learning, ICML 2024, Vienna, Austria, July 21-27, 2024. OpenReview.net (2024). https:\/\/openreview.net\/forum?id=XUeoOBid3x"},{"key":"9_CR22","unstructured":"Biderman, D., et al.: Lora learns less and forgets less. Trans. Mach. Learn. Res. 2024 (2024). https:\/\/openreview.net\/forum?id=aloEru2qCG"},{"key":"9_CR23","unstructured":"Clark, P., Cowhey, I., Etzioni, O., Khot, T., Sabharwal, A., Schoenick, C., Tafjord, O.: Think you have solved question answering? try arc, the AI2 reasoning challenge. CoRR abs\/1803.05457 (2018). http:\/\/arxiv.org\/abs\/1803.05457"},{"key":"9_CR24","doi-asserted-by":"publisher","unstructured":"Zellers, R., Holtzman, A., Bisk, Y., Farhadi, A., Choi, Y.: Hellaswag: Can a machine really finish your sentence? In: Korhonen, A., Traum, D.R., M\u00e0rquez, L. (eds.) Proceedings of the 57th Conference of the Association for Computational Linguistics, ACL 2019, Florence, Italy, July 28- August 2, 2019, Volume 1: Long Papers, pp. 4791\u20134800. Association for Computational Linguistics (2019). https:\/\/doi.org\/10.18653\/V1\/P19-1472","DOI":"10.18653\/V1\/P19-1472"},{"key":"9_CR25","doi-asserted-by":"publisher","unstructured":"Sakaguchi, K., Bras, R.L., Bhagavatula, C., Choi, Y.: Winogrande: an adversarial winograd schema challenge at scale. Commun. ACM 64(9), 99\u2013106 (2021). https:\/\/doi.org\/10.1145\/3474381","DOI":"10.1145\/3474381"},{"key":"9_CR26","doi-asserted-by":"publisher","first-page":"10088","DOI":"10.52202\/075280-0441","volume":"36","author":"T Dettmers","year":"2023","unstructured":"Dettmers, T., Pagnoni, A., Holtzman, A., Zettlemoyer, L.: Qlora: efficient finetuning of quantized llms. Adv. Neural. Inf. Process. Syst. 36, 10088\u201310115 (2023)","journal-title":"Adv. Neural. Inf. Process. Syst."}],"container-title":["Lecture Notes in Computer Science","PRICAI 2025: Trends in Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-7072-0_9","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T00:09:06Z","timestamp":1773274146000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-7072-0_9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9789819570713","9789819570720"],"references-count":26,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-7072-0_9","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"1 April 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRICAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Pacific Rim International Conference on Artificial Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Wellington","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"New Zealand","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17 November 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21 November 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"pricai2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.pricai.org\/2025\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}