{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T08:56:21Z","timestamp":1743065781345,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":42,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819720910"},{"type":"electronic","value":"9789819720927"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-97-2092-7_15","type":"book-chapter","created":{"date-parts":[[2024,3,29]],"date-time":"2024-03-29T06:02:18Z","timestamp":1711692138000},"page":"302-315","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Isolation and\u00a0Integration: A Strong Pre-trained Model-Based Paradigm for\u00a0Class-Incremental Learning"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0009-8202-6186","authenticated-orcid":false,"given":"Wei","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Yuan","family":"Xie","sequence":"additional","affiliation":[]},{"given":"Zhizhong","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Xin","family":"Tan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,3,30]]},"reference":[{"key":"15_CR1","doi-asserted-by":"crossref","unstructured":"Aljundi, R., Babiloni, F., Elhoseiny, M., Rohrbach, M., Tuytelaars, T.: Memory aware synapses: learning what (not) to forget. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 139\u2013154 (2018)","DOI":"10.1007\/978-3-030-01219-9_9"},{"key":"15_CR2","first-page":"15920","volume":"33","author":"P Buzzega","year":"2020","unstructured":"Buzzega, P., Boschini, M., Porrello, A., Abati, D., Calderara, S.: Dark experience for general continual learning: a strong, simple baseline. Adv. Neural. Inf. Process. Syst. 33, 15920\u201315930 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"15_CR3","unstructured":"Chen, S., et al.: Adaptformer: adapting vision transformers for scalable visual recognition. Adv. Neural. Inf. Process. Syst. 35, 16664\u201316678 (2022)"},{"key":"15_CR4","doi-asserted-by":"crossref","unstructured":"Dhar, P., Singh, R.V., Peng, K.C., Wu, Z., Chellappa, R.: Learning without memorizing. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5138\u20135146 (2019)","DOI":"10.1109\/CVPR.2019.00528"},{"key":"15_CR5","unstructured":"Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"15_CR6","doi-asserted-by":"crossref","unstructured":"Douillard, A., Ram\u00e9, A., Couairon, G., Cord, M.: Dytox: transformers for continual learning with dynamic token expansion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9285\u20139295 (2022)","DOI":"10.1109\/CVPR52688.2022.00907"},{"issue":"7","key":"15_CR7","doi-asserted-by":"publisher","first-page":"1755","DOI":"10.1162\/08997660260028700","volume":"14","author":"RM French","year":"2002","unstructured":"French, R.M., Chater, N.: Using noise to compute error surfaces in connectionist networks: a novel means of reducing catastrophic forgetting. Neural Comput. 14(7), 1755\u20131769 (2002)","journal-title":"Neural Comput."},{"key":"15_CR8","doi-asserted-by":"crossref","unstructured":"Han, X., et al.: Pre-trained models: past, present and future. AI Open 2, 225\u2013250 (2021)","DOI":"10.1016\/j.aiopen.2021.08.002"},{"key":"15_CR9","doi-asserted-by":"crossref","unstructured":"Hendrycks, D., et al.: The many faces of robustness: a critical analysis of out-of-distribution generalization. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 8340\u20138349 (2021)","DOI":"10.1109\/ICCV48922.2021.00823"},{"key":"15_CR10","unstructured":"Houlsby, N., et al.: Parameter-efficient transfer learning for NLP. In: International Conference on Machine Learning, pp. 2790\u20132799. PMLR (2019)"},{"key":"15_CR11","doi-asserted-by":"crossref","unstructured":"Kirkpatrick, J., et al.: Overcoming catastrophic forgetting in neural networks. Proc. Natl. Acad. Sci. 114(13), 3521\u20133526 (2017)","DOI":"10.1073\/pnas.1611835114"},{"key":"15_CR12","unstructured":"Krizhevsky, A., Hinton, G., et al.: Learning multiple layers of features from tiny images (2009)"},{"key":"15_CR13","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 25 (2012)"},{"issue":"12","key":"15_CR14","doi-asserted-by":"publisher","first-page":"2935","DOI":"10.1109\/TPAMI.2017.2773081","volume":"40","author":"Z Li","year":"2017","unstructured":"Li, Z., Hoiem, D.: Learning without forgetting. IEEE Trans. Pattern Anal. Mach. Intell. 40(12), 2935\u20132947 (2017)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"15_CR15","unstructured":"Van\u00a0der Maaten, L., Hinton, G.: Visualizing data using t-sne. J. Mach. Learn. Res. 9(11) (2008)"},{"key":"15_CR16","unstructured":"Nguyen, C.V., Achille, A., Lam, M., Hassner, T., Mahadevan, V., Soatto, S.: Toward understanding catastrophic forgetting in continual learning. arXiv preprint arXiv:1908.01091 (2019)"},{"key":"15_CR17","doi-asserted-by":"publisher","unstructured":"Prabhu, A., Torr, P.H.S., Dokania, P.K.: GDumb: a simple approach that questions our progress in continual learning. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) Computer Vision \u2013 ECCV 2020. LNCS, vol. 12347, pp. 524\u2013540. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58536-5_31","DOI":"10.1007\/978-3-030-58536-5_31"},{"key":"15_CR18","unstructured":"Radford, A., et al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748\u20138763. PMLR (2021)"},{"key":"15_CR19","doi-asserted-by":"crossref","unstructured":"Rebuffi, S.A., Kolesnikov, A., Sperl, G., Lampert, C.H.: icarl: Incremental classifier and representation learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2001\u20132010 (2017)","DOI":"10.1109\/CVPR.2017.587"},{"key":"15_CR20","unstructured":"Rusu, A.A., et al.: Progressive neural networks. arXiv preprint arXiv:1606.04671 (2016)"},{"key":"15_CR21","unstructured":"Serra, J., Suris, D., Miron, M., Karatzoglou, A.: Overcoming catastrophic forgetting with hard attention to the task. In: International Conference on Machine Learning, pp. 4548\u20134557. PMLR (2018)"},{"key":"15_CR22","doi-asserted-by":"crossref","unstructured":"Smith, J.S., et al.: Coda-prompt: continual decomposed attention-based prompting for rehearsal-free continual learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11909\u201311919 (2023)","DOI":"10.1109\/CVPR52729.2023.01146"},{"key":"15_CR23","unstructured":"Sun, H.L., Zhou, D.W., Ye, H.J., Zhan, D.C.: Pilot: a pre-trained model-based continual learning toolbox. arXiv preprint arXiv:2309.07117 (2023)"},{"key":"15_CR24","unstructured":"Wah, C., Branson, S., Welinder, P., Perona, P., Belongie, S.: The caltech-UCSD birds-200-2011 dataset (2011)"},{"key":"15_CR25","doi-asserted-by":"publisher","unstructured":"Wang, F.Y., Zhou, D.W., Ye, H.J., Zhan, D.C.: FOSTER: feature boosting and compression for class-incremental learning. In: Avidan, S., Brostow, G., Cisse, M., Farinella, G.M., Hassner, T. (eds.) Computer Vision, ECCV 2022. LNCS, vol. 13685, pp. 398\u2013414. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19806-9_23","DOI":"10.1007\/978-3-031-19806-9_23"},{"key":"15_CR26","unstructured":"Wang, L., et al.: Afec: active forgetting of negative transfer in continual learning. Adv. Neural. Inf. Process. Syst. 34, 22379\u201322391 (2021)"},{"key":"15_CR27","doi-asserted-by":"publisher","unstructured":"Wang, L., Zhang, X., Li, Q., Zhu, J., Zhong, Y.: CoSCL: cooperation of\u00a0small continual learners is stronger than a\u00a0big one. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) Computer Vision, ECCV 2022. LNCS, vol. 13686, pp. 254\u2013271. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19809-0_15","DOI":"10.1007\/978-3-031-19809-0_15"},{"key":"15_CR28","unstructured":"Wang, L., et al.: Memory replay with data compression for continual learning. arXiv preprint arXiv:2202.06592 (2022)"},{"key":"15_CR29","doi-asserted-by":"crossref","unstructured":"Wang, R., et al.: Attriclip: a non-incremental learner for incremental knowledge learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3654\u20133663 (2023)","DOI":"10.1109\/CVPR52729.2023.00356"},{"key":"15_CR30","first-page":"5682","volume":"35","author":"Y Wang","year":"2022","unstructured":"Wang, Y., Huang, Z., Hong, X.: S-prompts learning with pre-trained transformers: an Occam\u2019s razor for domain incremental learning. Adv. Neural. Inf. Process. Syst. 35, 5682\u20135695 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"15_CR31","doi-asserted-by":"publisher","unstructured":"Wang, Z., et al.: DualPrompt: complementary prompting for\u00a0rehearsal-free continual learning. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) Computer Vision, ECCV 2022. LNCS, vol. 13686, pp. 631\u2013648. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19809-0_36","DOI":"10.1007\/978-3-031-19809-0_36"},{"key":"15_CR32","doi-asserted-by":"crossref","unstructured":"Wang, Z., et al.: Learning to prompt for continual learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 139\u2013149 (2022)","DOI":"10.1109\/CVPR52688.2022.00024"},{"key":"15_CR33","doi-asserted-by":"crossref","unstructured":"Welling, M.: Herding dynamical weights to learn. In: Proceedings of the 26th Annual International Conference on Machine Learning, pp. 1121\u20131128 (2009)","DOI":"10.1145\/1553374.1553517"},{"key":"15_CR34","doi-asserted-by":"crossref","unstructured":"Wu, Y., et al.: Large scale incremental learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 374\u2013382 (2019)","DOI":"10.1109\/CVPR.2019.00046"},{"key":"15_CR35","doi-asserted-by":"crossref","unstructured":"Yan, S., Xie, J., He, X.: Der: dynamically expandable representation for class incremental learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3014\u20133023 (2021)","DOI":"10.1109\/CVPR46437.2021.00303"},{"key":"15_CR36","doi-asserted-by":"crossref","unstructured":"Zhang, G., Wang, L., Kang, G., Chen, L., Wei, Y.: SLCA: slow learner with classifier alignment for continual learning on a pre-trained model. arXiv preprint arXiv:2303.05118 (2023)","DOI":"10.1109\/ICCV51070.2023.01754"},{"key":"15_CR37","doi-asserted-by":"crossref","unstructured":"Zheng, Z., Ma, M., Wang, K., Qin, Z., Yue, X., You, Y.: Preventing zero-shot transfer degradation in continual learning of vision-language models. arXiv preprint arXiv:2303.06628 (2023)","DOI":"10.1109\/ICCV51070.2023.01752"},{"key":"15_CR38","unstructured":"Zhou, D.W., Wang, Q.W., Qi, Z.H., Ye, H.J., Zhan, D.C., Liu, Z.: Deep class-incremental learning: a survey. arXiv preprint arXiv:2302.03648 (2023)"},{"key":"15_CR39","unstructured":"Zhou, D.W., Wang, Q.W., Ye, H.J., Zhan, D.C.: A model or 603 exemplars: towards memory-efficient class-incremental learning. arXiv preprint arXiv:2205.13218 (2022)"},{"key":"15_CR40","doi-asserted-by":"crossref","unstructured":"Zhou, D.W., Ye, H.J., Zhan, D.C.: Co-transport for class-incremental learning. In: Proceedings of the 29th ACM International Conference on Multimedia, pp. 1645\u20131654 (2021)","DOI":"10.1145\/3474085.3475306"},{"key":"15_CR41","doi-asserted-by":"crossref","unstructured":"Zhou, D.W., Ye, H.J., Zhan, D.C., Liu, Z.: Revisiting class-incremental learning with pre-trained models: generalizability and adaptivity are all you need. arXiv preprint arXiv:2303.07338 (2023)","DOI":"10.1007\/s11263-024-02218-0"},{"key":"15_CR42","unstructured":"Zhou, D.W., Zhang, Y., Ning, J., Ye, H.J., Zhan, D.C., Liu, Z.: Learning without forgetting for vision-language models. arXiv preprint arXiv:2305.19270 (2023)"}],"container-title":["Lecture Notes in Computer Science","Computational Visual Media"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-97-2092-7_15","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,15]],"date-time":"2024-11-15T07:06:21Z","timestamp":1731654381000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-97-2092-7_15"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9789819720910","9789819720927"],"references-count":42,"URL":"https:\/\/doi.org\/10.1007\/978-981-97-2092-7_15","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"30 March 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"CVM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Computational Visual Media","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Wellington","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"New Zealand","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"10 April 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12 April 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"cvm2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CVM submission system","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"212","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"34","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"16% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}