{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,24]],"date-time":"2026-02-24T18:15:23Z","timestamp":1771956923409,"version":"3.50.1"},"publisher-location":"Cham","reference-count":78,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031732089","type":"print"},{"value":"9783031732096","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,11,1]],"date-time":"2024-11-01T00:00:00Z","timestamp":1730419200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,1]],"date-time":"2024-11-01T00:00:00Z","timestamp":1730419200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-73209-6_18","type":"book-chapter","created":{"date-parts":[[2024,10,31]],"date-time":"2024-10-31T15:02:57Z","timestamp":1730386977000},"page":"306-324","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":11,"title":["Weighted Ensemble Models Are Strong Continual Learners"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2479-0564","authenticated-orcid":false,"given":"Imad Eddine","family":"Marouf","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0008-2395-8111","authenticated-orcid":false,"given":"Subhankar","family":"Roy","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4274-8298","authenticated-orcid":false,"given":"Enzo","family":"Tartaglione","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6927-8930","authenticated-orcid":false,"given":"St\u00e9phane","family":"Lathuili\u00e8re","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,1]]},"reference":[{"issue":"2","key":"18_CR1","doi-asserted-by":"publisher","first-page":"73","DOI":"10.1016\/j.tins.2004.12.003","volume":"28","author":"WC Abraham","year":"2005","unstructured":"Abraham, W.C., Robins, A.: Memory retention-the synaptic stability versus plasticity dilemma. Trends Neurosci. 28(2), 73\u201378 (2005)","journal-title":"Trends Neurosci."},{"key":"18_CR2","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"144","DOI":"10.1007\/978-3-030-01219-9_9","volume-title":"Computer Vision \u2013 ECCV 2018","author":"R Aljundi","year":"2018","unstructured":"Aljundi, R., Babiloni, F., Elhoseiny, M., Rohrbach, M., Tuytelaars, T.: Memory aware synapses: learning what (not) to forget. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11207, pp. 144\u2013161. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01219-9_9"},{"key":"18_CR3","unstructured":"Amari, S.I.: Neural learning in structured parameter spaces - natural riemannian gradient. In: Mozer, M., Jordan, M., Petsche, T. (eds.) Advances in Neural Information Processing Systems. vol.\u00a09. MIT Press (1996)"},{"key":"18_CR4","unstructured":"Buzzega, P., Boschini, M., Porrello, A., Abati, D., Calderara, S.: Dark experience for general continual learning: a strong, simple baseline. In: Advances in Neural Information Processing Systems, vol. 33 (2020)"},{"key":"18_CR5","doi-asserted-by":"crossref","unstructured":"Chen*, X., Xie*, S., He, K.: An empirical study of training self-supervised vision transformers. arXiv preprint arXiv:2104.02057 (2021)","DOI":"10.1109\/ICCV48922.2021.00950"},{"key":"18_CR6","unstructured":"Chizat, L., Oyallon, E., Bach, F.: On lazy training in differentiable programming (2020)"},{"key":"18_CR7","unstructured":"Choshen, L., Venezian, E., Slonim, N., Katz, Y.: Fusing finetuned models for better pretraining (2022)"},{"key":"18_CR8","doi-asserted-by":"crossref","unstructured":"Dhar, P., Singh, R.V., Peng, K.C., Wu, Z., Chellappa, R.: Learning without memorizing. In: IEEE Conference on Computer Vision and Pattern Recognition (2019)","DOI":"10.1109\/CVPR.2019.00528"},{"key":"18_CR9","doi-asserted-by":"crossref","unstructured":"Dogucu, M., Johnson, A., Ott, M.: Bayesrules: datasets and supplemental functions from bayes rules! book (2021), r package version 0.0.2.9000","DOI":"10.32614\/CRAN.package.bayesrules"},{"key":"18_CR10","unstructured":"Dyer, E., Lewkowycz, A., Ramasesh, V.: Effect of scale on catastrophic forgetting in neural networks. In: International Conference on Learning Representations (2022)"},{"key":"18_CR11","doi-asserted-by":"crossref","unstructured":"d\u2019Ascoli, S., Touvron, H., Leavitt, M.L., Morcos, A.S., Biroli, G., Sagun, L.: Convit: improving vision transformers with soft convolutional inductive biases. In: International Conference of Machine Learning (2021)","DOI":"10.1088\/1742-5468\/ac9830"},{"key":"18_CR12","unstructured":"Entezari, R., Sedghi, H., Saukh, O., Neyshabur, B.: The role of permutation invariance in linear mode connectivity of neural networks. In: International Conference on Learning Representations (2022)"},{"key":"18_CR13","unstructured":"Foret, P., Kleiner, A., Mobahi, H., Neyshabur, B.: Sharpness-aware minimization for efficiently improving generalization. In: International Conference on Learning Representations (2021)"},{"key":"18_CR14","unstructured":"Frankle, J., Dziugaite, G.K., Roy, D., Carbin, M.: Linear mode connectivity and the lottery ticket hypothesis. In: International Conference on Machine Learning (2020)"},{"key":"18_CR15","doi-asserted-by":"crossref","unstructured":"French, R.M.: Catastrophic forgetting in connectionist networks. Trends Cogn. Sci. 3(4), 128\u2013135 (1999)","DOI":"10.1016\/S1364-6613(99)01294-2"},{"key":"18_CR16","doi-asserted-by":"crossref","unstructured":"Friedman, J., Hastie, T., Tibshirani, R., et\u00a0al.: The elements of statistical learning. Springer series in statistics New York (2001)","DOI":"10.1007\/978-0-387-21606-5"},{"key":"18_CR17","doi-asserted-by":"crossref","unstructured":"He, K., Chen, X., Xie, S., Li, Y., Doll\u00e1r, P., Girshick, R.: Masked autoencoders are scalable vision learners (2021)","DOI":"10.1109\/CVPR52688.2022.01553"},{"key":"18_CR18","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: IEEE Conference on Computer Vision and Pattern Recognition (2015)","DOI":"10.1109\/CVPR.2016.90"},{"key":"18_CR19","doi-asserted-by":"crossref","unstructured":"Hendrycks, D., et\u00a0al.: The many faces of robustness: a critical analysis of out-of-distribution generalization. In: International Conference of Computer Vision (ICCV) (2021)","DOI":"10.1109\/ICCV48922.2021.00823"},{"key":"18_CR20","unstructured":"Ilharco, G., et al.: Patching open-vocabulary models by interpolating weights (2022)"},{"key":"18_CR21","unstructured":"Izmailov, P., Podoprikhin, D., Garipov, T., Vetrov, D., Wilson, A.G.: Averaging weights leads to wider optima and better generalization. In: Conference on Uncertainty in Artificial Intelligence (UAI) (2018)"},{"key":"18_CR22","unstructured":"Janson, P., Zhang, W., Aljundi, R., Elhoseiny, M.: A simple baseline that questions the use of pretrained-models in continual learning (2023)"},{"key":"18_CR23","unstructured":"Kairouz, P., et al.: Advances and open problems in federated learning (2021)"},{"key":"18_CR24","doi-asserted-by":"crossref","unstructured":"Kim, D., Han, B.: On the stability-plasticity dilemma of class-incremental learning. In: IEEE Conference on Computer Vision and Pattern Recognition (2023)","DOI":"10.1109\/CVPR52729.2023.01934"},{"key":"18_CR25","doi-asserted-by":"crossref","unstructured":"Kirkpatrick, J., et\u00a0al.: Overcoming catastrophic forgetting in neural networks. PNAS 114(13), 3521\u20133526 (2017)","DOI":"10.1073\/pnas.1611835114"},{"key":"18_CR26","unstructured":"Kolesnikov, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. In: International Conference on Learning Representations (2021)"},{"key":"18_CR27","doi-asserted-by":"crossref","unstructured":"Krause, J., Stark, M., Deng, J., Fei-Fei, L.: 3d object representations for fine-grained categorization. In: Proceedings of the IEEE International Conference on Computer Vision Workshops (2013)","DOI":"10.1109\/ICCVW.2013.77"},{"key":"18_CR28","unstructured":"Krizhevsky, A., Hinton, G., et\u00a0al.: Learning multiple layers of features from tiny images (2009)"},{"key":"18_CR29","unstructured":"Lakshminarayanan, B., Pritzel, A., Blundell, C.: Simple and scalable predictive uncertainty estimation using deep ensembles. In: Advances in Neural Information Processing Systems (2017)"},{"issue":"3","key":"18_CR30","doi-asserted-by":"publisher","first-page":"50","DOI":"10.1109\/MSP.2020.2975749","volume":"37","author":"T Li","year":"2020","unstructured":"Li, T., Sahu, A.K., Talwalkar, A., Smith, V.: Federated learning: challenges, methods, and future directions. IEEE Sig. Process. Mag. 37(3), 50\u201360 (2020)","journal-title":"IEEE Sig. Process. Mag."},{"issue":"12","key":"18_CR31","doi-asserted-by":"publisher","first-page":"2935","DOI":"10.1109\/TPAMI.2017.2773081","volume":"40","author":"Z Li","year":"2017","unstructured":"Li, Z., Hoiem, D.: Learning without forgetting. TPAMI 40(12), 2935\u20132947 (2017)","journal-title":"TPAMI"},{"issue":"1","key":"18_CR32","doi-asserted-by":"publisher","first-page":"15","DOI":"10.1109\/TPAMI.2018.2876413","volume":"42","author":"Z Liao","year":"2020","unstructured":"Liao, Z., Drummond, T., Reid, I., Carneiro, G.: Approximate fisher information matrix to characterize the training of deep neural networks. IEEE Trans. Pattern Anal. Mach. Intell. 42(1), 15\u201326 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"18_CR33","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"issue":"5","key":"18_CR34","doi-asserted-by":"publisher","first-page":"5513","DOI":"10.1109\/TPAMI.2022.3213473","volume":"45","author":"M Masana","year":"2022","unstructured":"Masana, M., Liu, X., Twardowski, B., Menta, M., Bagdanov, A.D., van de Weijer, J.: Class-incremental learning: survey and performance evaluation on image classification. IEEE Trans. Pattern Anal. Mach. Intell. 45(5), 5513\u20135533 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"18_CR35","first-page":"17703","volume":"35","author":"MS Matena","year":"2022","unstructured":"Matena, M.S., Raffel, C.A.: Merging models with fisher-weighted averaging. Adv. Neural Inform. Process. Syst. 35, 17703\u201317716 (2022)","journal-title":"Adv. Neural Inform. Process. Syst."},{"key":"18_CR36","unstructured":"McDonnell, M.D., Gong, D., Parvaneh, A., Abbasnejad, E., van\u00a0den Hengel, A.: Ranpac: random projections and pre-trained models for continual learning. In: Advances in Neural Information Processing Systems vol. 36 (2024)"},{"key":"18_CR37","unstructured":"McMahan, H.B., Moore, E., Ramage, D., Hampson, S.: y Arcas. Communication-efficient learning of deep networks from decentralized data, B.A. (2023)"},{"key":"18_CR38","unstructured":"Mehta, S.V., Patil, D., Chandar, S., Strubell, E.: An empirical investigation of the role of pre-training in lifelong learning (2023)"},{"key":"18_CR39","unstructured":"Mirzadeh, S.I., Farajtabar, M., Gorur, D., Pascanu, R., Ghasemzadeh, H.: Linear mode connectivity in multitask and continual learning (2020)"},{"key":"18_CR40","doi-asserted-by":"crossref","unstructured":"Murata, K., Ito, S., Ohara, K.: Learning and transforming general representations to break down stability-plasticity dilemma. In: Proceedings of the Asian Conference on Computer Vision (2022)","DOI":"10.1007\/978-3-031-26351-4_33"},{"key":"18_CR41","unstructured":"Neyshabur, B., Sedghi, H., Zhang, C.: What is being transferred in transfer learning? In: Advances in Neural Information Processing Systems (2020)"},{"key":"18_CR42","unstructured":"Nichol, A., Achiam, J., Schulman, J.: On first-order meta-learning algorithms. arXiv preprint arXiv:1803.02999 (2018)"},{"key":"18_CR43","unstructured":"Oquab, M., et al.: Dinov2: learning robust visual features without supervision (2023)"},{"key":"18_CR44","unstructured":"Ovadia, Y., et al.: Can you trust your model\u2019s uncertainty? evaluating predictive uncertainty under dataset shift. In: Advances in Neural Information Processing Systems (2019)"},{"key":"18_CR45","doi-asserted-by":"crossref","unstructured":"Panos, A., Kobe, Y., Reino, D.O., Aljundi, R., Turner, R.E.: First session adaptation: a strong replay-free baseline for class-incremental learning. In: International Conference on Computer Vision (2023)","DOI":"10.1109\/ICCV51070.2023.01725"},{"key":"18_CR46","unstructured":"Pascanu, R., Bengio, Y.: Revisiting natural gradient for deep networks (2014)"},{"key":"18_CR47","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"524","DOI":"10.1007\/978-3-030-58536-5_31","volume-title":"Computer Vision \u2013 ECCV 2020","author":"A Prabhu","year":"2020","unstructured":"Prabhu, A., Torr, P.H.S., Dokania, P.K.: GDumb: a simple approach that questions our progress in continual learning. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12347, pp. 524\u2013540. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58536-5_31"},{"key":"18_CR48","unstructured":"Radford, A., et\u00a0al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, PMLR (2021)"},{"key":"18_CR49","unstructured":"Ramasesh, V.V., Lewkowycz, A., Dyer, E.: Effect of scale on catastrophic forgetting in neural networks. In: Proceedings of the International Conference on Learning Representations (2021)"},{"key":"18_CR50","unstructured":"Ram\u00e9, A., Kirchmeyer, M., Rahier, T., Rakotomamonjy, A., Gallinari, P., Cord, M.: Diverse weight averaging for out-of-distribution generalization (2023)"},{"key":"18_CR51","unstructured":"Ridnik, T., Ben-Baruch, E., Noy, A., Zelnik-Manor, L.: Imagenet-21k pretraining for the masses (2021)"},{"key":"18_CR52","unstructured":"Rusu, A.A., et al.: Progressive neural networks. arXiv preprint arXiv:1606.04671 (2016)"},{"key":"18_CR53","unstructured":"Rype\u015b\u0107, G., Cygert, S., Khan, V., Trzcinski, T., Zieli\u0144ski, B.M., Twardowski, B.: Divide and not forget: ensemble of selectively trained experts in continual learning. In: International Conference on Learning Representations (2023)"},{"key":"18_CR54","unstructured":"Schuhmann, C., et al.: Laion-400 m: open dataset of clip-filtered 400 million image-text pairs. arXiv preprint arXiv:2111.02114 (2021)"},{"key":"18_CR55","unstructured":"Serra, J., Suris, D., Miron, M., Karatzoglou, A.: Overcoming catastrophic forgetting with hard attention to the task. In: International Conference on Machine Learning (2018)"},{"key":"18_CR56","first-page":"5708","volume":"34","author":"A Soen","year":"2021","unstructured":"Soen, A., Sun, K.: On the variance of the fisher information for deep learning. Adv. Neural Inform. Process. Syst. 34, 5708\u20135719 (2021)","journal-title":"Adv. Neural Inform. Process. Syst."},{"issue":"4","key":"18_CR57","doi-asserted-by":"publisher","first-page":"889","DOI":"10.1198\/106186005X78800","volume":"14","author":"JC Spall","year":"2005","unstructured":"Spall, J.C.: Monte carlo computation of the fisher information matrix in nonstandard settings. J. Comput. Graph. Stat. 14(4), 889\u2013909 (2005)","journal-title":"J. Comput. Graph. Stat."},{"key":"18_CR58","doi-asserted-by":"crossref","unstructured":"Spall, J.C.: Improved methods for monte carlo estimation of the fisher information matrix. In: 2008 American Control Conference (2008)","DOI":"10.1109\/ACC.2008.4586850"},{"key":"18_CR59","unstructured":"Stickland, A.C., Murray, I.: Diverse ensembles improve calibration. In: International Conference on Machine Learning (ICML) Workshop on Uncertainty and Robustness in Deep Learning (2020)"},{"key":"18_CR60","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z.: Rethinking the inception architecture for computer vision. In: IEEE Conference on Computer Vision and Pattern Recognition (2016)","DOI":"10.1109\/CVPR.2016.308"},{"key":"18_CR61","unstructured":"van\u00a0de Ven, G.M., Tolias, A.S.: Three scenarios for continual learning (2019)"},{"key":"18_CR62","doi-asserted-by":"crossref","unstructured":"Villa, A., et al.: Pivot: prompting for video continual learning. arXiv preprint arXiv:2212.04842 (2022)","DOI":"10.1109\/CVPR52729.2023.02319"},{"key":"18_CR63","unstructured":"Wah, C., Branson, S., Welinder, P., et\u00a0al.: The caltech-ucsd birds-200-2011 dataset (2011)"},{"key":"18_CR64","doi-asserted-by":"crossref","unstructured":"Wang, L., Yang, K., Li, C., Hong, L., Li, Z., Zhu, J.: Ordisco: effective and efficient usage of incremental unlabeled data for semi-supervised continual learning. In: IEEE Conference on Computer Vision and Pattern Recognition (2021)","DOI":"10.1109\/CVPR46437.2021.00534"},{"issue":"12","key":"18_CR65","doi-asserted-by":"publisher","first-page":"1356","DOI":"10.1038\/s42256-023-00747-w","volume":"5","author":"L Wang","year":"2023","unstructured":"Wang, L., et al.: Incorporating neuro-inspired adaptability for continual learning in artificial intelligence. Nat. Mach. Intell. 5(12), 1356\u20131368 (2023)","journal-title":"Nat. Mach. Intell."},{"key":"18_CR66","doi-asserted-by":"crossref","unstructured":"Wang, L., Zhang, X., Li, Q., Zhu, J., Zhong, Y.: Coscl: cooperation of small continual learners is stronger than a big one (2022)","DOI":"10.1007\/978-3-031-19809-0_15"},{"key":"18_CR67","unstructured":"Wang, L., Zhang, X., Su, H., Zhu, J.: A comprehensive survey of continual learning: theory, method and application (2023)"},{"key":"18_CR68","unstructured":"Wang, Y., Huang, Z., Hong, X.: S-prompts learning with pre-trained transformers: an occam\u2019s razor for domain incremental learning. arXiv preprint arXiv:2207.12819 (2022)"},{"key":"18_CR69","doi-asserted-by":"publisher","unstructured":"Wang, Z., et al.: DualPrompt: complementary prompting for rehearsal-free continual learning. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) Computer Vision - ECCV 2022, ECCV 2022, LNCS, vol. 13686, pp. 631\u2013648. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19809-0_36","DOI":"10.1007\/978-3-031-19809-0_36"},{"key":"18_CR70","doi-asserted-by":"crossref","unstructured":"Wang, Z., et al.: Learning to prompt for continual learning. In: Conference on Computer Vision and Pattern Recognition (2022)","DOI":"10.1109\/CVPR52688.2022.00024"},{"key":"18_CR71","unstructured":"Wortsman, M., et\u00a0al.: Model soups: averaging weights of multiple fine-tuned models improves accuracy without increasing inference time. In: International Conference on Machine Learning, PMLR (2022)"},{"key":"18_CR72","doi-asserted-by":"crossref","unstructured":"Wortsman, M., et al.: Robust fine-tuning of zero-shot models. In: IEEE Conference on Computer Vision and Pattern Recognition (2022)","DOI":"10.1109\/CVPR52688.2022.00780"},{"key":"18_CR73","doi-asserted-by":"crossref","unstructured":"Wu, T.Y., et al.: Class-incremental learning with strong pre-trained models. In: IEEE Conference on Computer Vision and Pattern Recognition (2022)","DOI":"10.1109\/CVPR52688.2022.00938"},{"key":"18_CR74","doi-asserted-by":"crossref","unstructured":"Wu, Y., et al.: Large scale incremental learning. In: IEEE Conference on Computer Vision and Pattern Recognition (2019)","DOI":"10.1109\/CVPR.2019.00046"},{"key":"18_CR75","doi-asserted-by":"crossref","unstructured":"Yang, B., et al.: Continual object detection via prototypical task correlation guided gating mechanism. In: IEEE Conference on Computer Vision and Pattern Recognition (2022)","DOI":"10.1109\/CVPR52688.2022.00904"},{"key":"18_CR76","unstructured":"Zenke, F., Poole, B., Ganguli, S.: Continual learning through synaptic intelligence. In: International Conference on Machine Learning (2017)"},{"key":"18_CR77","doi-asserted-by":"crossref","unstructured":"Zhang, G., Wang, L., Kang, G., Chen, L., Wei, Y.: Slca: slow learner with classifier alignment for continual learning on a pre-trained model. In: International Conference on Computer Vision (2023)","DOI":"10.1109\/ICCV51070.2023.01754"},{"key":"18_CR78","doi-asserted-by":"crossref","unstructured":"Zhou, D.W., Ye, H.J., Zhan, D.C., Liu, Z.: Revisiting class-incremental learning with pre-trained models: Generalizability and adaptivity are all you need (2023)","DOI":"10.1007\/s11263-024-02218-0"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-73209-6_18","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,4,8]],"date-time":"2025-04-08T12:26:17Z","timestamp":1744115177000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-73209-6_18"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,1]]},"ISBN":["9783031732089","9783031732096"],"references-count":78,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-73209-6_18","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,1]]},"assertion":[{"value":"1 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}