{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,2]],"date-time":"2026-01-02T03:29:26Z","timestamp":1767324566392,"version":"3.48.0"},"publisher-location":"Cham","reference-count":35,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783032128393","type":"print"},{"value":"9783032128409","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-3-032-12840-9_22","type":"book-chapter","created":{"date-parts":[[2026,1,2]],"date-time":"2026-01-02T03:25:28Z","timestamp":1767324328000},"page":"337-352","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Assessing Foundation Models for\u00a0Mold Colony Detection with\u00a0Limited Training Data"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0008-6616-3108","authenticated-orcid":false,"given":"Henrik","family":"Pichler","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1327-1243","authenticated-orcid":false,"given":"Janis","family":"Keuper","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0002-0844-4055","authenticated-orcid":false,"given":"Matthew","family":"Copping","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,2]]},"reference":[{"key":"22_CR1","unstructured":"360iResearch: Indoor air quality monitors market size & forecast to 2030. Market Research Report Report ID 5889448, 360iResearch (2025). https:\/\/www.researchandmarkets.com\/report\/indoor-air-quality-monitoring. Accessed 10 June 2025"},{"key":"22_CR2","doi-asserted-by":"publisher","unstructured":"Alcorn, M.A., Li, Q., Gong, Z., Wang, C., Mai, L., Ku, W.S., Nguyen, A.: Strike (with) a pose: Neural networks are easily fooled by strange poses of familiar objects. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR). pp. 4845\u20134854. IEEE, Long Beach, CA, USA (Jun 2019). https:\/\/doi.org\/10.48550\/arXiv.1811.11553, https:\/\/openaccess.thecvf.com\/content_CVPR_2019\/html\/Alcorn_Strike_With_a_Pose_Neural_Networks_Are_Easily_Fooled_by_CVPR_2019_paper.html","DOI":"10.48550\/arXiv.1811.11553"},{"key":"22_CR3","unstructured":"Barbu, A., et al.: ObjectNet: a large-scale bias-controlled dataset for pushing the limits of object recognition models. In: Advances in Neural Information Processing Systems 32 (NeurIPS 2019), pp. 9448\u20139458. Advances in Neural Information Processing Systems, Curran Associates, Inc. (2019). https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2019\/file\/97af07a14cacba681feacf3012730892-Paper.pdf"},{"key":"22_CR4","doi-asserted-by":"publisher","unstructured":"Bommasani, R., Hudson, D.A., et\u00a0al.: On the opportunities and risks of foundation models. arXiv (2021). https:\/\/doi.org\/10.48550\/arXiv.2108.07258, https:\/\/arxiv.org\/abs\/2108.07258","DOI":"10.48550\/arXiv.2108.07258"},{"key":"22_CR5","doi-asserted-by":"publisher","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Li, F.F.: ImageNet: a large-scale hierarchical image database. In: Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 248\u2013255. IEEE (2009). https:\/\/doi.org\/10.1109\/CVPR.2009.5206848","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"22_CR6","doi-asserted-by":"publisher","unstructured":"Dodge, S., Karam, L.: A study and comparison of human and deep learning recognition performance under visual distortions. In: Proceedings of the 26th International Conference on Computer Communications and Networks (ICCCN), pp.\u00a01\u20137. IEEE (2017). https:\/\/doi.org\/10.1109\/ICCCN.2017.8038465","DOI":"10.1109\/ICCCN.2017.8038465"},{"key":"22_CR7","doi-asserted-by":"publisher","unstructured":"Geirhos, R., Rubisch, P., Michaelis, C., Bethge, M., Wichmann, F.A., Brendel, W.: Imagenet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness. In: Proceedings of the 7th International Conference on Learning Representations (ICLR) (2019). https:\/\/doi.org\/10.48550\/arXiv.1811.12231, https:\/\/openreview.net\/forum?id=Bygh9j09KX","DOI":"10.48550\/arXiv.1811.12231"},{"key":"22_CR8","unstructured":"Goldblum, M., et al.: Battle of the backbones: a large-scale comparison of pretrained models across computer vision tasks. In: Advances in Neural Information Processing Systems 36 (NeurIPS 2023), Datasets and Benchmarks Track. Proceedings of the Neural Information Processing Systems, vol.\u00a036, pp. 1\u201329. Neural Information Processing Systems Foundation (2023). https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2023\/file\/5d9571470bb750f0e2325a030016f63f-Paper-Datasets_and_Benchmarks.pdf"},{"key":"22_CR9","doi-asserted-by":"publisher","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P., Girshick, R.: Mask R\u2013CNN. In: Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), pp. 2980\u20132988. IEEE (2017). https:\/\/doi.org\/10.1109\/ICCV.2017.322","DOI":"10.1109\/ICCV.2017.322"},{"key":"22_CR10","doi-asserted-by":"publisher","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770\u2013778. IEEE (2016). https:\/\/doi.org\/10.1109\/CVPR.2016.90","DOI":"10.1109\/CVPR.2016.90"},{"key":"22_CR11","unstructured":"Kelman, B., Stock, A., Robbins, C.: Indoor air quality: health effects of airborne mold & how mold is measured. White paper, J.S. Held LLC, Jericho (2022). https:\/\/www.jsheld.com\/insights\/articles\/indoor-air-quality-health-effects-of-airborne-mold-how-mold-is-measured-indoors. Accessed 10 June 2025"},{"key":"22_CR12","doi-asserted-by":"publisher","unstructured":"Kim, H.E., Cosa-Linan, A., Santhanam, N., Jannesari, M., Maros, M.E., Ganslandt, T.: Transfer learning for medical image classification: a literature review. BMC Med. Imaging 22, 69 (2022). https:\/\/doi.org\/10.1186\/s12880-022-00793-7","DOI":"10.1186\/s12880-022-00793-7"},{"key":"22_CR13","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. In: Pereira, F., Burges, C.J.C., Bottou, L., Weinberger, K.Q. (eds.) Advances in Neural Information Processing Systems 25 (NeurIPS 2012). Advances in Neural Information Processing Systems, vol.\u00a025, pp. 1097\u20131105. Curran Associates, Inc., Lake Tahoe (2012). https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2012\/file\/c399862d3b9d6b76c8436e924a68c45b-Paper.pdf"},{"key":"22_CR14","doi-asserted-by":"publisher","unstructured":"Li, F., Zhang, H., Xu, H., Liu, S., Zhang, L., Ni, L.M., Shum, H.Y.: Mask DINO: towards a unified transformer-based framework for object detection and segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3041\u20133050. IEEE (2023). https:\/\/doi.org\/10.1109\/CVPR52729.2023.00297","DOI":"10.1109\/CVPR52729.2023.00297"},{"key":"22_CR15","doi-asserted-by":"publisher","unstructured":"Liang, F., Li, Y., Marculescu, D.: SupMAE: supervised masked autoencoders are efficient vision learners. arXiv (2022). https:\/\/doi.org\/10.48550\/arXiv.2205.14540, https:\/\/arxiv.org\/abs\/2205.14540. Edge Intelligence Workshop, AAAI 2024; version 3, Accessed 21 Jan 2024","DOI":"10.48550\/arXiv.2205.14540"},{"key":"22_CR16","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1007\/978-3-319-10602-1_48","volume-title":"Computer Vision \u2013 ECCV 2014","author":"T-Y Lin","year":"2014","unstructured":"Lin, T.-Y., et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 740\u2013755. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10602-1_48"},{"key":"22_CR17","doi-asserted-by":"publisher","unstructured":"Liu, Z., Mao, H., Wu, C., Feichtenhofer, C., Darrell, T., Xie, S.: A convnet for the 2020s. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 11976\u201311986. IEEE (2022). https:\/\/doi.org\/10.1109\/CVPR52688.2022.01167","DOI":"10.1109\/CVPR52688.2022.01167"},{"key":"22_CR18","unstructured":"Radford, A., Kim, J.W., Sutskever, I., Krueger, G., Agarwal, S.: CLIP: connecting text and images (2021). https:\/\/openai.com\/index\/clip\/. OpenAI Research blog post"},{"key":"22_CR19","doi-asserted-by":"publisher","unstructured":"Rahman, M.A., et al.: Classification of fungal genera from microscopic images using artificial intelligence. J. Pathol. Inform. 14, 100314 (2023). https:\/\/doi.org\/10.1016\/j.jpi.2023.100314","DOI":"10.1016\/j.jpi.2023.100314"},{"key":"22_CR20","doi-asserted-by":"publisher","unstructured":"Ravi, N., et al.: SAM 2: segment anything in images and videos. arXiv (2024). https:\/\/doi.org\/10.48550\/arXiv.2408.00714, https:\/\/arxiv.org\/abs\/2408.00714","DOI":"10.48550\/arXiv.2408.00714"},{"key":"22_CR21","unstructured":"Robinson, I., Robicheaux, P., Popov, M., et\u00a0al.: RF-DETR: a real-time transformer-based object detection model. https:\/\/github.com\/roboflow\/rf-detr"},{"key":"22_CR22","unstructured":"scikit-learn developers: sklearn.model_selection.stratifiedkfold\u2014scikit-learn 1.7.0 documentation (2025). https:\/\/scikit-learn.org\/stable\/modules\/generated\/sklearn.model_selection.StratifiedKFold.html"},{"key":"22_CR23","doi-asserted-by":"publisher","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: Proceedings of the 3rd International Conference on Learning Representations (ICLR) (2015). https:\/\/doi.org\/10.48550\/arXiv.1409.1556, https:\/\/arxiv.org\/abs\/1409.1556","DOI":"10.48550\/arXiv.1409.1556"},{"key":"22_CR24","doi-asserted-by":"publisher","unstructured":"Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z.: Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2818\u20132826. IEEE (2016). https:\/\/doi.org\/10.1109\/CVPR.2016.308","DOI":"10.1109\/CVPR.2016.308"},{"key":"22_CR25","doi-asserted-by":"publisher","unstructured":"Terven, J., C\u00f3rdova-Esparza, D., Romero-Gonz\u00e1lez, J.: A comprehensive review of YOLO architectures in computer vision: from YOLOv1 to YOLOv8 and YOLO-NAS. Mach. Learn. Knowl. Extract. 5(4), 1680\u20131716 (2023). https:\/\/doi.org\/10.3390\/make5040083","DOI":"10.3390\/make5040083"},{"key":"22_CR26","doi-asserted-by":"publisher","unstructured":"Tsang, C.C., et al.: Automatic identification of clinically important Aspergillus species by artificial intelligence-based image recognition: proof-of-concept study. Emerg. Microbes Infect. 14(1), 2434573 (2025). https:\/\/doi.org\/10.1080\/22221751.2024.2434573","DOI":"10.1080\/22221751.2024.2434573"},{"key":"22_CR27","unstructured":"United States Environmental Protection Agency: Care for your air: A guide to indoor air quality. Technical report, U.S. Environmental Protection Agency (2008). https:\/\/www.epa.gov\/indoor-air-quality-iaq\/care-your-air-guide-indoor-air-quality. Accessed 10 June 2025"},{"key":"22_CR28","unstructured":"Verein Deutscher Ingenieure: VDI 6022 Blatt 1: Raumlufttechnik, Raumluftqualit\u00e4t \u2013 Hygieneanforderungen an raumlufttechnische Anlagen und Ger\u00e4te (VDI-L\u00fcftungsregeln). VDI Guideline VDI 6022-1, Verein Deutscher Ingenieure e.\u00a0V. (2018). https:\/\/www.vdi.de\/en\/home\/vdi-standards\/details\/vdi-6022-blatt-1-ventilation-and-indoor-air-quality-hygiene-requirements-for-ventilation-and-air-conditioning-systems-and-units-vdi-ventilation-code-of-practice-1"},{"key":"22_CR29","doi-asserted-by":"publisher","unstructured":"Wang, C., Yeh, I., Liao, H.M.: YOLOv9: learning what you want to learn using programmable gradient information. In: Computer Vision \u2013 ECCV 2024. Lecture Notes in Computer Science, vol. 15089, pp. 1\u201321. Springer (2024). https:\/\/doi.org\/10.1007\/978-3-031-72751-1_1","DOI":"10.1007\/978-3-031-72751-1_1"},{"key":"22_CR30","doi-asserted-by":"publisher","unstructured":"Wang, H., et al.: Early detection and classification of live bacteria using time-lapse coherent imaging and deep learning. Light: Sci. Appl. 9, 118 (2020). https:\/\/doi.org\/10.1038\/s41377-020-00358-9","DOI":"10.1038\/s41377-020-00358-9"},{"key":"22_CR31","doi-asserted-by":"publisher","unstructured":"Wang, W., Zheng, V.W., Yu, H., Miao, C.: A survey of zero-shot learning: settings, methods, and applications. ACM Trans. Intell. Syst. Technol. 10(2), 13:1\u201313:37 (2019). https:\/\/doi.org\/10.1145\/3293318","DOI":"10.1145\/3293318"},{"key":"22_CR32","doi-asserted-by":"publisher","unstructured":"Wu, M., et al.: Classification of dry and wet macular degeneration based on the ConvNeXT model. Front. Comput. Neurosci. 16, 1079155 (2022). https:\/\/doi.org\/10.3389\/fncom.2022.1079155","DOI":"10.3389\/fncom.2022.1079155"},{"key":"22_CR33","doi-asserted-by":"publisher","unstructured":"Yu, J., Wang, Z., Vasudevan, V., Yeung, L., Seyedhosseini, M., Wu, Y.: CoCa: contrastive captioners are image-text foundation models. arxiv abs\/2205.01917 (2022). https:\/\/doi.org\/10.48550\/arXiv.2205.01917, https:\/\/arxiv.org\/abs\/2205.01917","DOI":"10.48550\/arXiv.2205.01917"},{"key":"22_CR34","doi-asserted-by":"publisher","unstructured":"Zhao, T., et al.: A foundation model for joint segmentation, detection and recognition of biomedical objects across nine modalities. Nature Methods 22(1), 166\u2013176 (2025). https:\/\/doi.org\/10.1038\/s41592-024-02499-w","DOI":"10.1038\/s41592-024-02499-w"},{"key":"22_CR35","doi-asserted-by":"publisher","unstructured":"Zieli\u0144ski, B., Plichta, A., Misztal, K., Spurek, P., Brzychczy-W\u0142och, M., Ocho\u0144ska, D.: Deep learning approach to bacterial colony classification. PLOS ONE 12(9), e0184554 (2017). https:\/\/doi.org\/10.1371\/journal.pone.0184554","DOI":"10.1371\/journal.pone.0184554"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-032-12840-9_22","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,2]],"date-time":"2026-01-02T03:25:31Z","timestamp":1767324331000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-032-12840-9_22"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9783032128393","9783032128409"],"references-count":35,"URL":"https:\/\/doi.org\/10.1007\/978-3-032-12840-9_22","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"2 January 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"DAGM GCPR","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"DAGM German Conference on Pattern Recognition","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Freiburg","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Germany","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 September 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 September 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"47","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"dagm2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.dagm-gcpr.de\/year\/2025","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}