{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,8]],"date-time":"2026-03-08T02:24:17Z","timestamp":1772936657957,"version":"3.50.1"},"reference-count":85,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T00:00:00Z","timestamp":1772841600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T00:00:00Z","timestamp":1772841600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Pioneer R&D Program of Zhejiang Province","award":["2024C01024"],"award-info":[{"award-number":["2024C01024"]}]},{"name":"Open Research Project of National Key Laboratory of Science and Technology on Space-Born Intelligent Information Processing","award":["TJ-02-22-01"],"award-info":[{"award-number":["TJ-02-22-01"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Comput Vis"],"published-print":{"date-parts":[[2026,4]]},"DOI":"10.1007\/s11263-026-02746-x","type":"journal-article","created":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T08:12:34Z","timestamp":1772871154000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Few-shot Class-Incremental Learning via Generative Co-Memory Regularization"],"prefix":"10.1007","volume":"134","author":[{"given":"Kexin","family":"Bao","sequence":"first","affiliation":[]},{"given":"Yong","family":"Li","sequence":"additional","affiliation":[]},{"given":"Dan","family":"Zeng","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5293-310X","authenticated-orcid":false,"given":"Shiming","family":"Ge","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,3,7]]},"reference":[{"key":"2746_CR1","doi-asserted-by":"crossref","unstructured":"Agarwal, A., Banerjee, B., Cuzzolin, F.,& Chaudhuri, S.(2022). Semantics-driven generative replay for few-shot class incremental learning. In: ACM Int. Conf. Multimedia (MM), pp. 5246\u20135254","DOI":"10.1145\/3503161.3548160"},{"key":"2746_CR2","doi-asserted-by":"crossref","unstructured":"Ahmed, N., Kukleva, A.,& Schiele, B.(2024). Orco: Towards better generalization via orthogonality and contrast for few-shot class-incremental learning. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 28762\u201328771","DOI":"10.1109\/CVPR52733.2024.02717"},{"key":"2746_CR3","unstructured":"Ahn, H., Cha, S., Lee, D.,& Moon, T.(2019). Uncertainty-based continual learning with adaptive regularization. In: Proc. Adv. Neural Inform. Process. Syst. (NeurIPS), pp. 4394\u20134404"},{"key":"2746_CR4","unstructured":"Aky\u00fcrek, A.F., Aky\u00fcrek, E., Wijaya, D., & Andreas, J.(2022). Subspace regularizers for few-shot class incremental learning. In: Proc. Int. Conf. Learn. Represent. (ICLR) . https:\/\/openreview.net\/forum?id=boJy41J-tnQ"},{"key":"2746_CR5","doi-asserted-by":"crossref","unstructured":"Aljundi, R., Babiloni, F., Elhoseiny, M., and others.(2018). Memory aware synapses: Learning what (not) to forget. In: Eur. Conf. Comput. Vis. (ECCV), pp. 144\u2013161","DOI":"10.1007\/978-3-030-01219-9_9"},{"key":"2746_CR6","doi-asserted-by":"crossref","unstructured":"Bai, J., Yuan, A., Xiao, Z., and others.(2022). Class incremental learning with few-shots based on linear programming for hyperspectral image classification. IEEE Trans. Cybern. (TC),52(6), 5474\u20135485","DOI":"10.1109\/TCYB.2020.3032958"},{"key":"2746_CR7","doi-asserted-by":"crossref","unstructured":"Bansal, A., Nanduri, A., Castillo, C.D., and others.(2017). Umdfaces: An annotated face dataset for training deep networks. In: IEEE International Joint Conference on Biometrics (IJCB), pp. 464\u2013473","DOI":"10.1109\/BTAS.2017.8272731"},{"key":"2746_CR8","unstructured":"Brown, T., Mann, B., Ryder, N., and others.(2020). Language models are few-shot learners. In: Proc. Adv. Neural Inform. Process. Syst. (NeurIPS), pp. 1877\u20131901"},{"key":"2746_CR9","unstructured":"Chen, K., & Lee, C.(2021). Incremental few-shot learning via vector quantization in deep embedded space. In: Proc. Int. Conf. Learn. Represent. (ICLR) . https:\/\/openreview.net\/forum?id=3SV-ZePhnZM"},{"key":"2746_CR10","doi-asserted-by":"crossref","unstructured":"Chi, Z., Gu, L., Liu, H., and others.(2022). Metafscil: A meta-learning approach for few-shot class incremental learning. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 14146\u201314155","DOI":"10.1109\/CVPR52688.2022.01377"},{"key":"2746_CR11","doi-asserted-by":"crossref","unstructured":"Cui, Y., Deng, W., Xu, X., and others. (2023). Uncertainty-guided semi-supervised few-shot class-incremental learning with knowledge distillation. IEEE Trans. Multimedia (TMM),25, 6422\u20136435.","DOI":"10.1109\/TMM.2022.3208743"},{"key":"2746_CR12","doi-asserted-by":"crossref","unstructured":"Cui, Y., Yu, Z., Peng, W., and others. (2024). Rethinking few-shot class-incremental learning with open-set hypothesis in hyperbolic geometry. IEEE Trans. Multimedia (TMM),26, 5897\u20135910.","DOI":"10.1109\/TMM.2023.3340550"},{"key":"2746_CR13","doi-asserted-by":"crossref","unstructured":"D\u2019Alessandro, M., Alonso, A., Calabr\u00e9s, E.,& Galar, M.(2023). Multimodal parameter-efficient few-shot class incremental learning. In: Proc. IEEE\/CVF Int. Conf. Comput. Vis. (ICCV) Workshop, pp. 3393\u20133403","DOI":"10.1109\/ICCVW60793.2023.00364"},{"key":"2746_CR14","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., and others.(2009). ImageNet: A large-scale hierarchical image database. In: Proc. IEEE Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 248\u2013255","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"2746_CR15","doi-asserted-by":"crossref","unstructured":"Deng, Y.,& Xiang, X.(2024). Expanding hyperspherical space for few-shot class-incremental learning. In: IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 1956\u20131965","DOI":"10.1109\/WACV57701.2024.00197"},{"key":"2746_CR16","doi-asserted-by":"crossref","unstructured":"Dhar, P., Singh, R.V., Peng, K., and others. (2019). Learning without memorizing. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 5138\u20135146","DOI":"10.1109\/CVPR.2019.00528"},{"key":"2746_CR17","doi-asserted-by":"crossref","unstructured":"Dong, J., Li, H., Cong, Y., and others.(2024). No one left behind: Real-world federated class-incremental learning. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI),46(4), 2054\u20132070","DOI":"10.1109\/TPAMI.2023.3334213"},{"key":"2746_CR18","doi-asserted-by":"crossref","unstructured":"Dong, J., Wang, L., Fang, Z., and others.(2022). Federated class-incremental learning. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 10154\u201310163","DOI":"10.1109\/CVPR52688.2022.00992"},{"key":"2746_CR19","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., and others.(2021). An image is worth 16x16 words: Transformers for image recognition at scale. In: Proc. Int. Conf. Learn. Represent. (ICLR) . https:\/\/openreview.net\/forum?id=YicbFdNTTy"},{"key":"2746_CR20","doi-asserted-by":"crossref","unstructured":"Guo, C., Zhao, Q., Lyu, S., and others.(2023). Decision boundary optimization for few-shot class-incremental learning. In: Proc. IEEE\/CVF Int. Conf. Comput. Vis. (ICCV) Workshop, pp. 3493\u20133503","DOI":"10.1109\/ICCVW60793.2023.00376"},{"key":"2746_CR21","doi-asserted-by":"crossref","unstructured":"He, K., Chen, X., Xie, S., and others.(2022). Masked autoencoders are scalable vision learners. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 15979\u201315988","DOI":"10.1109\/CVPR52688.2022.01553"},{"key":"2746_CR22","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S.,& Sun, J.(2016). Deep residual learning for image recognition. In: Proc. IEEE Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"2746_CR23","unstructured":"Hendrycks, D., Mu, N., Cubuk, E.D., and others.(2020). Augmix: A simple data processing method to improve robustness and uncertainty. In: Proc. Int. Conf. Learn. Represent. (ICLR) . https:\/\/openreview.net\/forum?id=S1gmrxHFvB"},{"key":"2746_CR24","doi-asserted-by":"crossref","unstructured":"Hersche, M., Karunaratne, G., Cherubini, G., and others. (2022). Constrained few-shot class-incremental learning. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 9047\u20139057","DOI":"10.1109\/CVPR52688.2022.00885"},{"key":"2746_CR25","doi-asserted-by":"crossref","unstructured":"Hou, S., Pan, X., Loy, C.C., and others.(2019). Learning a unified classifier incrementally via rebalancing. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 831\u2013839","DOI":"10.1109\/CVPR.2019.00092"},{"key":"2746_CR26","doi-asserted-by":"crossref","unstructured":"Iscen, A., Zhang, J., Lazebnik, S., & Schmid, C.(2020). Memory-efficient incremental learning through feature adaptation. In: Eur. Conf. Comput. Vis. (ECCV), pp. 699\u2013715","DOI":"10.1007\/978-3-030-58517-4_41"},{"key":"2746_CR27","doi-asserted-by":"crossref","unstructured":"Ji, Z., Hou, Z., Liu, X., and others. (2023). Memorizing complementation network for few-shot class-incremental learning. IEEE Trans. Image Process. (TIP),32, 937\u2013948.","DOI":"10.1109\/TIP.2023.3236160"},{"key":"2746_CR28","unstructured":"Khandelwal, A.(2023). MASIL: towards maximum separable class representation for few shot class incremental learning. In: Topological, Algebraic and Geometric Learning Workshops, pp. 519\u2013533"},{"key":"2746_CR29","unstructured":"Kim, D., Han, D., Seo, J., &Moon, J.(2023). Warping the space: Weight space rotation for class-incremental few-shot learning. In: Proc. Int. Conf. Learn. Represent. (ICLR) . https:\/\/openreview.net\/pdf?id=kPLzOfPfA2l"},{"key":"2746_CR30","doi-asserted-by":"crossref","unstructured":"Kirkpatrick, J., Pascanu, R., Rabinowitz, N., and others.(2017). Overcoming catastrophic forgetting in neural networks. Proc. Nat. Aca. of Sci. (PNAS),114(13), 3521\u20133526","DOI":"10.1073\/pnas.1611835114"},{"key":"2746_CR31","doi-asserted-by":"crossref","unstructured":"Kiyasseh, D., Zhu, T.,& Clifton, D.(2021). A clinical deep learning framework for continually learning from cardiac signals across diseases, time, modalities, and institutions. Nat. Commun.12, 4221","DOI":"10.1038\/s41467-021-24483-0"},{"key":"2746_CR32","unstructured":"Krizhevsky, A.(2009). Learning multiple layers of features from tiny images. Technical report, University of Toronto"},{"key":"2746_CR33","doi-asserted-by":"crossref","unstructured":"Kundu, J., Venkatesh, R., Venkat, N., and others.(2020). Class-incremental domain adaptation. In: Eur. Conf. Comput. Vis. (ECCV), pp. 53\u201369","DOI":"10.1007\/978-3-030-58601-0_4"},{"key":"2746_CR34","first-page":"2579","volume":"9","author":"VDM Laurens","year":"2008","unstructured":"Laurens, V. D. M., & Hinton, G. (2008). Visualizing data using t-SNE. In:J. Mach. Learn. Res. (JMLR),9, 2579\u20132605.","journal-title":"In:J. Mach. Learn. Res. (JMLR)"},{"key":"2746_CR35","doi-asserted-by":"crossref","unstructured":"Lee, J., Hong, H.G., Joo, D.,& Kim, J.(2020). Continual learning with extended kronecker-factored approximate curvature. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 8998\u20139007","DOI":"10.1109\/CVPR42600.2020.00902"},{"key":"2746_CR36","doi-asserted-by":"crossref","unstructured":"Lin, H., Feng, S., Li, X., and others.(2023). Anchor assisted experience replay for online class-incremental learning. IEEE Trans. Circuit Syst. Video Technol. (TCSVT),33(5), 2217\u20132232","DOI":"10.1109\/TCSVT.2022.3219605"},{"key":"2746_CR37","doi-asserted-by":"crossref","unstructured":"Lin, H., Han, G., Ma, J., and others.(2023). Supervised masked knowledge distillation for few-shot transformers. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 19649\u201319659","DOI":"10.1109\/CVPR52729.2023.01882"},{"key":"2746_CR38","doi-asserted-by":"crossref","unstructured":"Liu, H., Gu, L., Chi, Z., and others.(2022). Few-shot class-incremental learning via entropy-regularized data-free replay. In: Eur. Conf. Comput. Vis. (ECCV), pp. 146\u2013162","DOI":"10.1007\/978-3-031-20053-3_9"},{"key":"2746_CR39","doi-asserted-by":"crossref","unstructured":"Liu, B., Yang, B., Xie, L., and others.(2023). Learnable distribution calibration for few-shot class-incremental learning. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI),45(10), 12699\u201312706","DOI":"10.1109\/TPAMI.2023.3273291"},{"key":"2746_CR40","doi-asserted-by":"crossref","unstructured":"Long, Y., Xia, G.-S., Li, S., and others. (2021). On creating benchmark dataset for aerial image interpretation: Reviews, guidances, and million-aid. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens.,14, 4205\u20134230.","DOI":"10.1109\/JSTARS.2021.3070368"},{"key":"2746_CR41","doi-asserted-by":"crossref","unstructured":"Luo, X., Chen, W., Liang, Z., and others. (2023). Adversarial style discrepancy minimization for unsupervised domain adaptation. Neural Networks,157, 216\u2013225.","DOI":"10.1016\/j.neunet.2022.10.015"},{"key":"2746_CR42","doi-asserted-by":"crossref","unstructured":"Ostapenko, O., Puscas, M.M., Klein, T., and others.(2019). Learning to remember: A synaptic plasticity driven framework for continual learning. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 11321\u201311329","DOI":"10.1109\/CVPR.2019.01158"},{"key":"2746_CR43","doi-asserted-by":"crossref","unstructured":"Park, K.-H., Song, K., & Park, G.-M.(2024). Pre-trained vision and language transformers are few-shot incremental learners. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 23881\u201323890","DOI":"10.1109\/CVPR52733.2024.02254"},{"key":"2746_CR44","doi-asserted-by":"crossref","unstructured":"Peng, C., Zhao, K., Wang, T.,& Lovell, B.C.(2022). Few-shot class-incremental learning from an open-set perspective. In: Eur. Conf. Comput. Vis. (ECCV), pp. 382\u2013397","DOI":"10.1007\/978-3-031-19806-9_22"},{"key":"2746_CR45","doi-asserted-by":"crossref","unstructured":"Qi, H., Brown, M.,& Lowe, D.(2018). Low-shot learning with imprinted weights. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 5822\u20135830","DOI":"10.1109\/CVPR.2018.00610"},{"key":"2746_CR46","doi-asserted-by":"crossref","unstructured":"Qiang, W., Li, J., Su, B., and others. (2023). Meta attention-generation network for cross-granularity few-shot learning. Int. J. Comput. Vis. (IJCV),131(5), 1211\u20131233.","DOI":"10.1007\/s11263-023-01760-7"},{"key":"2746_CR47","doi-asserted-by":"crossref","unstructured":"Qiu, W., Fu, S., Zhang, J., and others.(2023). Semantic-visual guided transformer for few-shot class-incremental learning. In: Int. Conf. Multimedia and Expo (ICME), pp. 2885\u20132890","DOI":"10.1109\/ICME55011.2023.00490"},{"key":"2746_CR48","doi-asserted-by":"crossref","unstructured":"Rebuffi, S., Kolesnikov, A., Sperl, G.,& Lampert, C.H.(2017). iCaRL: Incremental classifier and representation learning. In: Proc. IEEE Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 5533\u20135542","DOI":"10.1109\/CVPR.2017.587"},{"key":"2746_CR49","doi-asserted-by":"publisher","first-page":"125","DOI":"10.1007\/s11263-007-0075-7","volume":"77","author":"DA Ross","year":"2008","unstructured":"Ross, D. A., Lim, J., Lin, R., & Yang, M. (2008). Incremental learning for robust visual tracking. Int. J. Comput. Vis. (IJCV),77, 125\u2013141.","journal-title":"Int. J. Comput. Vis. (IJCV)"},{"key":"2746_CR50","doi-asserted-by":"crossref","unstructured":"Silva-Rodr\u00edguez, J., Hajimiri, S., Ayed, I.B.,& Dolz, J.(2024). A closer look at the few-shot adaptation of large vision-language models. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 1\u201310","DOI":"10.1109\/CVPR52733.2024.02235"},{"key":"2746_CR51","unstructured":"Snell, J., Swersky, K.,& Zemel, R.(2017). Prototypical networks for few-shot learning. In: Proc. Adv. Neural Inform. Process. Syst. (NeurIPS), pp. 4080\u20134090"},{"key":"2746_CR52","unstructured":"Steiner, A., Kolesnikov, A., Zhai, X., and others. (2022). How to train your vit? data, augmentation, and regularization in vision transformers. Trans. Mach. Learn. Res. (TMLR),2022, 1\u201316."},{"key":"2746_CR53","doi-asserted-by":"crossref","unstructured":"Sung, Y., Cho, J., &Bansal, M.(2022). VL-ADAPTER: parameter-efficient transfer learning for vision-and-language tasks. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 5217\u20135227","DOI":"10.1109\/CVPR52688.2022.00516"},{"key":"2746_CR54","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TGRS.2023.3248601","volume":"61","author":"Y Tai","year":"2023","unstructured":"Tai, Y., Tan, Y., Xiong, S., & Tian, J. (2023). Mine-distill-prototypes for complete few-shot class-incremental learning in image classification. IEEE Trans. Geosci. Remote Sensing (TGRS),61, 1\u201313.","journal-title":"IEEE Trans. Geosci. Remote Sensing (TGRS)"},{"key":"2746_CR55","doi-asserted-by":"crossref","unstructured":"Tang, Y., Peng, Y., Meng, J., & Zheng, W.(2024). Rethinking few-shot class-incremental learning: Learning from yourself. In: Eur. Conf. Comput. Vis. (ECCV), pp. 108\u2013128","DOI":"10.1007\/978-3-031-73030-6_7"},{"key":"2746_CR56","doi-asserted-by":"crossref","unstructured":"Tao, X., Hong, X., Chang, X., and others.(2020). Few-shot class-incremental learning. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 12180\u201312189","DOI":"10.1109\/CVPR42600.2020.01220"},{"key":"2746_CR57","doi-asserted-by":"crossref","unstructured":"Triki, A.R., Aljundi, R., Blaschko, M.B.,& Tuytelaars, T.(2017). Encoder based lifelong learning. In: Proc. IEEE Int. Conf. Comput. Vis. (ICCV), pp. 1329\u20131337","DOI":"10.1109\/ICCV.2017.148"},{"key":"2746_CR58","unstructured":"Tsimpoukelli, M., Menick, J.L., Cabi, S., and others.(2021). Multimodal few-shot learning with frozen language models. In: Proc. Adv. Neural Inform. Process. Syst. (NeurIPS), pp. 200\u2013212"},{"key":"2746_CR59","unstructured":"Vinyals, O., Blundell, C., Lillicrap, T., and others. (2016). Matching networks for one shot learning. In: Proc. Adv. Neural Inform. Process. Syst. (NeurIPS), pp. 3637\u20133645"},{"key":"2746_CR60","unstructured":"Wah, C., Branson, S., Welinder, P., and others. (2011). The caltech-ucsd birds-200-2011 dataset. California Institute of Technology: Technical report."},{"key":"2746_CR61","doi-asserted-by":"crossref","unstructured":"Wang, J., Lan, C., Liu, C., and others.(2021). Generalizing to unseen domains: A survey on domain generalization. In: Proc. Int. Joint Conf. Artif. Intell. (IJCAI), pp. 4627\u20134635","DOI":"10.24963\/ijcai.2021\/628"},{"key":"2746_CR62","doi-asserted-by":"crossref","unstructured":"Wang, Z., Yang, E., Shen, L.,& Huang, H.(2025). A comprehensive survey of forgetting in deep learning beyond continual learning. IEEE Trans.Pattern Anal. Mach. Intell. (TPAMI),47(3), 1464\u20131483","DOI":"10.1109\/TPAMI.2024.3498346"},{"key":"2746_CR63","doi-asserted-by":"crossref","unstructured":"Wang, X., Zhang, S., Cen, J., and others.(2024). Clip-guided prototype modulating for few-shot action recognition. Int. J. Comput. Vis. (IJCV),132(6), 1899\u20131912","DOI":"10.1007\/s11263-023-01917-4"},{"key":"2746_CR64","doi-asserted-by":"crossref","unstructured":"Wang, W., Zhang, L., Fu, S., and others. (2024). Gradient guided multiscale feature collaboration networks for few-shot class-incremental remote sensing scene classification. IEEE Trans. Geosci. Remote Sensing (TGRS),62, 1\u201312.","DOI":"10.1109\/TGRS.2024.3369083"},{"key":"2746_CR65","doi-asserted-by":"crossref","unstructured":"Wang, Q., Zhou, D., Zhang, Y., Zhan, D., & Ye, H.(2023). Few-shot class-incremental learning via training-free prototype calibration. In: Proc. Adv. Neural Inform. Process. Syst. (NeurIPS), pp. 15060\u201315076","DOI":"10.52202\/075280-0661"},{"key":"2746_CR66","unstructured":"Wu, C., Herranz, L., Liu, X., and others.(2018). Memory replay gans: Learning to generate images from new categories without forgetting. In: Proc. Adv. Neural Inform. Process. Syst. (NeurIPS), pp. 5966\u20135976"},{"key":"2746_CR67","doi-asserted-by":"crossref","unstructured":"Wu, R., Wang, M., Li, Z., and others. (2024). Few-shot stereo matching with high domain adaptability based on adaptive recursive network. Int. J. Comput. Vis. (IJCV),132(5), 1484\u20131501.","DOI":"10.1007\/s11263-023-01953-0"},{"key":"2746_CR68","doi-asserted-by":"crossref","unstructured":"Yang, X., Han, M., Luo, Y., and others. (2023). Two-stream prototype learning network for few-shot face recognition under occlusions. IEEE Trans. Multimedia (TMM),25, 1555\u20131563.","DOI":"10.1109\/TMM.2023.3253054"},{"key":"2746_CR69","unstructured":"Yang, B., Lin, M., Zhang, Y., and others. (2023). Dynamic support network for few-shot class incremental learning. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI),45(3), 2945\u20132951."},{"key":"2746_CR70","unstructured":"Yang, Y., Yuan, H., Li, X., and others.(2023). Neural collapse inspired feature-classifier alignment for few-shot class-incremental learning. In: Proc. Int. Conf. Learn. Represent. (ICLR) . https:\/\/openreview.net\/forum?id=y5W8tpojhtJ"},{"key":"2746_CR71","unstructured":"Zenke, F., Poole, B., &Ganguli, S.(2017). Continual learning through synaptic intelligence. In: Proc. Int. Conf. Mach. Learn. (ICML), pp. 3987\u20133995"},{"key":"2746_CR72","doi-asserted-by":"crossref","unstructured":"Zhai, J., Liu, X., Bagdanov, A., and others. (2023). Masked autoencoders are efficient class incremental learners. In: Proc. IEEE\/CVF Int. Conf. Comput. Vis. (ICCV), pp. 19047\u201319056","DOI":"10.1109\/ICCV51070.2023.01750"},{"key":"2746_CR73","doi-asserted-by":"crossref","unstructured":"Zhang, C.-L., Luo, J.-H., Wei, X.-S., & Wu, J.(2018). In defense of fully connected layers in visual representation transfer. In: Pacific-Rim Conference on Multimedia (PCM), pp. 807\u2013817","DOI":"10.1007\/978-3-319-77383-4_79"},{"key":"2746_CR74","doi-asserted-by":"crossref","unstructured":"Zhang, C., Song, N., Lin, G., and others. (2021). Few-shot incremental learning with continually evolved classifiers. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 12455\u201312464","DOI":"10.1109\/CVPR46437.2021.01227"},{"key":"2746_CR75","doi-asserted-by":"crossref","unstructured":"Zhang, K., Zhang, C., Li, S., and others.(2022). Student network learning via evolutionary knowledge distillation. IEEE Trans. Circuit Syst. Video Technol. (TCSVT),32(4), 2251\u20132263","DOI":"10.1109\/TCSVT.2021.3090902"},{"key":"2746_CR76","doi-asserted-by":"crossref","unstructured":"Zhao, H., Fu, Y., Kang, M., and others. (2024). Mgsvf: Multi-grained slow vs. fast framework for few-shot class-incremental learning. IEEE Trans.Pattern Anal. Mach. Intell. (TPAMI),46(3), 1576\u20131588.","DOI":"10.1109\/TPAMI.2021.3133897"},{"key":"2746_CR77","doi-asserted-by":"crossref","unstructured":"Zhao, L., Lu, J., Cheng, Z., and others.(2023). Rethinking self-supervision for few-shot class-incremental learning. In: Int. Conf. Multimedia and Expo (ICME), pp. 726\u2013731","DOI":"10.1109\/ICME55011.2023.00130"},{"key":"2746_CR78","doi-asserted-by":"crossref","unstructured":"Zhao, L., Lu, J., Xu, Y., and others.(2023). Few-shot class-incremental learning via class-aware bilateral distillation. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 11838\u201311847","DOI":"10.1109\/CVPR52729.2023.01139"},{"key":"2746_CR79","unstructured":"Zheng, J., Li, X., & Lucey, S.(2024). Convolutional initialization for data-efficient vision transformers. arXiv preprint"},{"key":"2746_CR80","doi-asserted-by":"crossref","unstructured":"Zhou, D., Cai, Z.-W., Ye, H., and others.(2024). Revisiting class-incremental learning with pre-trained models: Generalizability and adaptivity are all you need. Int. J. Comput. Vis. (IJCV),133(3), 1012\u20131032","DOI":"10.1007\/s11263-024-02218-0"},{"key":"2746_CR81","doi-asserted-by":"crossref","unstructured":"Zhou, D., Sun, H., Ye, H., & Zhan, D.(2024). Expandable subspace ensemble for pre-trained model-based class-incremental learning. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 23554\u201323564","DOI":"10.1109\/CVPR52733.2024.02223"},{"key":"2746_CR82","doi-asserted-by":"crossref","unstructured":"Zhou, D., Wang, F., Ye, H., and others.(2022). Forward compatible few-shot class-incremental learning. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 9036\u20139046","DOI":"10.1109\/CVPR52688.2022.00884"},{"key":"2746_CR83","doi-asserted-by":"crossref","unstructured":"Zhou, D., Ye, H., Ma, L., and others.(2023). Few-shot class-incremental learning by sampling multi-phase tasks. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI),45(11), 12816\u201312831","DOI":"10.1109\/TPAMI.2022.3200865"},{"key":"2746_CR84","doi-asserted-by":"crossref","unstructured":"Zhu, K., Cao, Y., Zhai, W., and others.(2021). Self-promoted prototype refinement for few-shot class-incremental learning. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 6801\u20136810","DOI":"10.1109\/CVPR46437.2021.00673"},{"key":"2746_CR85","doi-asserted-by":"crossref","unstructured":"Zhuang, H., Weng, Z., He, R., and others. (2023). GKEAL: gaussian kernel embedded analytic learning for few-shot class incremental task. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recog. (CVPR), pp. 7746\u20137755","DOI":"10.1109\/CVPR52729.2023.00748"}],"container-title":["International Journal of Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-026-02746-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11263-026-02746-x","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-026-02746-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T08:12:50Z","timestamp":1772871170000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11263-026-02746-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,3,7]]},"references-count":85,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2026,4]]}},"alternative-id":["2746"],"URL":"https:\/\/doi.org\/10.1007\/s11263-026-02746-x","relation":{},"ISSN":["0920-5691","1573-1405"],"issn-type":[{"value":"0920-5691","type":"print"},{"value":"1573-1405","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,3,7]]},"assertion":[{"value":"25 August 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"1 January 2026","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 March 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"164"}}