{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T16:34:04Z","timestamp":1775579644005,"version":"3.50.1"},"publisher-location":"Cham","reference-count":37,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783032046239","type":"print"},{"value":"9783032046246","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,9,17]],"date-time":"2025-09-17T00:00:00Z","timestamp":1758067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,9,17]],"date-time":"2025-09-17T00:00:00Z","timestamp":1758067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-3-032-04624-6_25","type":"book-chapter","created":{"date-parts":[[2025,9,16]],"date-time":"2025-09-16T05:33:10Z","timestamp":1758000790000},"page":"425-442","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Few-Shot Segmentation of\u00a0Historical Maps via\u00a0Linear Probing of\u00a0Vision Foundation Models"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-0029-8463","authenticated-orcid":false,"given":"Rafael","family":"Sterzinger","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6843-0830","authenticated-orcid":false,"given":"Marco","family":"Peer","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4195-1593","authenticated-orcid":false,"given":"Robert","family":"Sablatnig","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,9,17]]},"reference":[{"key":"25_CR1","doi-asserted-by":"publisher","first-page":"693","DOI":"10.1007\/978-3-030-86337-1_46","volume-title":"Document Analysis and Recognition \u2013 ICDAR 2021","author":"J Chazalon","year":"2021","unstructured":"Chazalon, J., et al.: ICDAR 2021 competition on historical map segmentation. In: Llad\u00f3s, J., Lopresti, D., Uchida, S. (eds.) Document Analysis and Recognition \u2013 ICDAR 2021, pp. 693\u2013707. Springer International Publishing, Cham (2021)"},{"issue":"4","key":"25_CR2","doi-asserted-by":"publisher","first-page":"834","DOI":"10.1109\/tpami.2017.2699184","volume":"40","author":"LC Chen","year":"2018","unstructured":"Chen, L.C., Papandreou, G., Kokkinos, I., Murphy, K., Yuille, A.L.: DeepLab: semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 40(4), 834\u2013848 (2018). https:\/\/doi.org\/10.1109\/tpami.2017.2699184","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"25_CR3","doi-asserted-by":"crossref","unstructured":"Chen, L.C., Zhu, Y., Papandreou, G., Schroff, F., Adam, H.: Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation, pp. 833\u2013851. Springer International Publishing (2018)","DOI":"10.1007\/978-3-030-01234-2_49"},{"key":"25_CR4","unstructured":"Cheng, B., Schwing, A., Kirillov, A.: Per-pixel classification is not all you need for semantic segmentation. In: Ranzato, M., Beygelzimer, A., Dauphin, Y., Liang, P., Vaughan, J.W. (eds.) Advances in Neural Information Processing Systems. vol.\u00a034, pp. 17864\u201317875. Curran Associates, Inc. (2021)"},{"key":"25_CR5","doi-asserted-by":"crossref","unstructured":"De\u00a0Nardin, A., Zottin, S., Paier, M., Foresti, G.L., Colombi, E., Piciarelli, C.: Efficient few-shot learning for pixel-precise handwritten document layout analysis. In: 2023 IEEE\/CVF Winter Conference on Applications of Computer Vision (WACV). IEEE (Jan 2023)","DOI":"10.1109\/WACV56688.2023.00367"},{"key":"25_CR6","doi-asserted-by":"crossref","unstructured":"De\u00a0Nardin, A., Zottin, S., Piciarelli, C., Colombi, E., Foresti, G.L.: A one-shot learning approach to document layout segmentation of ancient arabic manuscripts. In: 2024 IEEE\/CVF Winter Conference on Applications of Computer Vision (WACV). IEEE (Jan 2024)","DOI":"10.1109\/WACV57701.2024.00794"},{"key":"25_CR7","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2022.109018","volume":"133","author":"H Ding","year":"2023","unstructured":"Ding, H., Zhang, H., Jiang, X.: Self-regularized prototypical network for few-shot semantic segmentation. Pattern Recogn. 133, 109018 (2023)","journal-title":"Pattern Recogn."},{"key":"25_CR8","unstructured":"Dosovitskiy, A., et al.: An image is worth 16x16 words: Transformers for image recognition at scale. In: International Conference on Learning Representations (2021)"},{"key":"25_CR9","unstructured":"Hardt, M., Sun, Y.: Test-time training on nearest neighbors for large language models. In: The Twelfth International Conference on Learning Representations (2024)"},{"key":"25_CR10","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770\u2013778. IEEE (Jun 2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"25_CR11","doi-asserted-by":"crossref","unstructured":"He, W., et al.: APSeg: Auto-prompt network for cross-domain few-shot semantic segmentation. In: 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 23762\u201323772. IEEE (Jun 2024)","DOI":"10.1109\/CVPR52733.2024.02243"},{"key":"25_CR12","unstructured":"Hu, E.J., et al.: LoRA: Low-rank adaptation of large language models. In: International Conference on Learning Representations (2022)"},{"key":"25_CR13","unstructured":"Hyeon-Woo, N., Ye-Bin, M., Oh, T.H.: FedPara: low-rank Hadamard product for communication-efficient federated learning. In: International Conference on Learning Representations (2022)"},{"key":"25_CR14","unstructured":"Kim, Y., Oh, J., Kim, S., Yun, S.Y.: How to Fine-tune Models with Few Samples: Update, Data Augmentation, and Test-time Augmentation (Aug 2022)"},{"key":"25_CR15","doi-asserted-by":"crossref","unstructured":"Kirillov, A., et al.: Segment anything. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 4015\u20134026 (October 2023)","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"25_CR16","doi-asserted-by":"crossref","unstructured":"Li, Y., Mao, H., Girshick, R., He, K.: Exploring Plain Vision Transformer Backbones for Object Detection, pp. 280\u2013296. Springer Nature Switzerland (2022)","DOI":"10.1007\/978-3-031-20077-9_17"},{"issue":"2","key":"25_CR17","doi-asserted-by":"publisher","first-page":"318","DOI":"10.1109\/TPAMI.2018.2858826","volume":"42","author":"TY Lin","year":"2020","unstructured":"Lin, T.Y., Goyal, P., Girshick, R., He, K., Dollar, P.: Focal loss for dense object detection. IEEE Trans. Pattern Anal. Mach. Intell. 42(2), 318\u2013327 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"25_CR18","unstructured":"Liu, Y., Zhu, M., Li, H., Chen, H., Wang, X., Shen, C.: Matcher: Segment anything with one shot using all-purpose feature matching. In: The Twelfth International Conference on Learning Representations (2024)"},{"key":"25_CR19","doi-asserted-by":"crossref","unstructured":"Mao, Y., Huang, K., Guan, C., Bao, G., Mo, F., Xu, J.: DoRA: enhancing parameter-efficient fine-tuning with dynamic rank distribution. In: Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 11662\u201311675. Association for Computational Linguistics (2024)","DOI":"10.18653\/v1\/2024.acl-long.626"},{"key":"25_CR20","unstructured":"Oquab, M., et al.: DINOv2: Learning robust visual features without supervision. Transactions on Machine Learning Research (2024)"},{"key":"25_CR21","doi-asserted-by":"crossref","unstructured":"Perera, R., Halgamuge, S.: Discriminative sample-guided and parameter-efficient feature space adaptation for cross-domain few-shot learning. In: 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 23794\u201323804. IEEE (Jun 2024)","DOI":"10.1109\/CVPR52733.2024.02246"},{"key":"25_CR22","unstructured":"Petitpierre, R., Kaplan, F., di\u00a0Lenardo, I.: Generic semantic segmentation of historical maps. In: Proceedings of the Workshop on Computational Humanities Research (CHR 2021). CEUR, Amsterdam, NL (2021)"},{"key":"25_CR23","doi-asserted-by":"crossref","unstructured":"Ranzinger, M., Heinrich, G., Kautz, J., Molchanov, P.: AM-RADIO: Agglomerative vision foundation model reduce all domains into one. In: 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12490\u201312500. IEEE (Jun 2024)","DOI":"10.1109\/CVPR52733.2024.01187"},{"key":"25_CR24","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-Net: Convolutional Networks for Biomedical Image Segmentation, pp. 234\u2013241. Springer International Publishing (2015)","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"25_CR25","doi-asserted-by":"crossref","unstructured":"Smith, L.N., Topin, N.: Super-Convergence: Very Fast Training of Neural Networks Using Large Learning Rates (May 2018)","DOI":"10.1117\/12.2520589"},{"key":"25_CR26","unstructured":"Snell, J., Swersky, K., Zemel, R.: Prototypical networks for few-shot learning. In: Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol.\u00a030. Curran Associates, Inc. (2017)"},{"key":"25_CR27","doi-asserted-by":"crossref","unstructured":"Sterzinger, R., Brenner, S., Sablatnig, R.: Drawing the Line: deep segmentation for extracting art from ancient etruscan mirrors. In: 2024 ICDAR International Conference on Document Analysis and Recognition, submitted (2024)","DOI":"10.1007\/978-3-031-70543-4_3"},{"key":"25_CR28","doi-asserted-by":"crossref","unstructured":"Sterzinger, R., Stippel, C., Sablatnig, R.: Fusing forces: deep-human-guided refinement of segmentation masks, pp. 154\u2013169. Springer Nature Switzerland (Dec 2024)","DOI":"10.1007\/978-3-031-78198-8_11"},{"key":"25_CR29","doi-asserted-by":"crossref","unstructured":"Sudre, C.H., Li, W., Vercauteren, T., Ourselin, S., Jorge\u00a0Cardoso, M.: Generalised Dice Overlap as a Deep Learning Loss Function for Highly Unbalanced Segmentations, pp. 240\u2013248. Springer International Publishing (2017)","DOI":"10.1007\/978-3-319-67558-9_28"},{"key":"25_CR30","doi-asserted-by":"crossref","unstructured":"Sun, Y., et al.: VRP-SAM: SAM with visual reference prompt. In: 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 23565\u201323574. IEEE (Jun 2024)","DOI":"10.1109\/CVPR52733.2024.02224"},{"key":"25_CR31","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol.\u00a030. Curran Associates, Inc. (2017)"},{"key":"25_CR32","doi-asserted-by":"crossref","unstructured":"Xia, X., Jiao, C., Hurni, L.: Contrastive pretraining for railway detection: Unveiling historical maps with transformers. In: Proceedings of the 6th ACM SIGSPATIAL International Workshop on AI for Geographic Knowledge Discovery, pp. 30\u201333. SIGSPATIAL \u201923, ACM (Nov 2023)","DOI":"10.1145\/3615886.3627738"},{"key":"25_CR33","doi-asserted-by":"crossref","unstructured":"Xia, X., Zhang, D., Song, W., Huang, W., Hurni, L.: MapSAM: adapting segment anything model for automated feature detection in historical maps. GIScience & Remote Sensing 62(1) (Apr 2025)","DOI":"10.1080\/15481603.2025.2494883"},{"key":"25_CR34","unstructured":"Xie, E., Wang, W., Yu, Z., Anandkumar, A., Alvarez, J.M., Luo, P.: Segformer: simple and efficient design for semantic segmentation with transformers. In: Ranzato, M., Beygelzimer, A., Dauphin, Y., Liang, P., Vaughan, J.W. (eds.) Advances in Neural Information Processing Systems, vol.\u00a034, pp. 12077\u201312090. Curran Associates, Inc. (2021)"},{"key":"25_CR35","doi-asserted-by":"crossref","unstructured":"Yun, S., Han, D., Chun, S., Oh, S.J., Yoo, Y., Choe, J.: CutMix: regularization Strategy to Train Strong Classifiers With Localizable Features. In: 2019 IEEE\/CVF International Conference on Computer Vision (ICCV). IEEE (Oct 2019)","DOI":"10.1109\/ICCV.2019.00612"},{"key":"25_CR36","unstructured":"Zhang, H., Ciss\u00e9, M., Dauphin, Y.N., Lopez-Paz, D.: Mixup: beyond empirical risk minimization. In: 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings (2018)"},{"key":"25_CR37","unstructured":"Zhang, R., et al.: Personalize segment anything model with one shot. In: The Twelfth International Conference on Learning Representations (2024)"}],"container-title":["Lecture Notes in Computer Science","Document Analysis and Recognition \u2013 ICDAR 2025"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-032-04624-6_25","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,16]],"date-time":"2025-09-16T05:33:24Z","timestamp":1758000804000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-032-04624-6_25"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,17]]},"ISBN":["9783032046239","9783032046246"],"references-count":37,"URL":"https:\/\/doi.org\/10.1007\/978-3-032-04624-6_25","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,9,17]]},"assertion":[{"value":"17 September 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICDAR","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Document Analysis and Recognition","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Wuhan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16 September 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21 September 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icdar2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/iapr.org\/icdar2025","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}