{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,25]],"date-time":"2025-10-25T19:14:26Z","timestamp":1761419666307,"version":"3.44.0"},"reference-count":22,"publisher":"Springer Science and Business Media LLC","issue":"32","license":[{"start":{"date-parts":[[2025,1,29]],"date-time":"2025-01-29T00:00:00Z","timestamp":1738108800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2025,1,29]],"date-time":"2025-01-29T00:00:00Z","timestamp":1738108800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100002809","name":"Generalitat de Catalunya","doi-asserted-by":"publisher","award":["CERCA Programme","ACE034\/21\/000084"],"award-info":[{"award-number":["CERCA Programme","ACE034\/21\/000084"]}],"id":[{"id":"10.13039\/501100002809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100014440","name":"Ministerio de Ciencia, Innovaci\u00f3n y Universidades","doi-asserted-by":"publisher","award":["RYC2020- 030777-I \/ AEI \/ 10.13039\/501100011033"],"award-info":[{"award-number":["RYC2020- 030777-I \/ AEI \/ 10.13039\/501100011033"]}],"id":[{"id":"10.13039\/100014440","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"abstract":"<jats:title>Abstract<\/jats:title>\n          <jats:p>In this paper, we address the challenges of automatic metadata annotation in the domain of Galleries, Libraries, Archives, and Museums (GLAMs) by introducing a novel dataset, EUFCC-340K, collected from the Europeana portal. Comprising over 340,000 images, the EUFCC-340K dataset is organized across multiple facets \u2013 Materials, Object Types, Disciplines, and Subjects \u2013 following a hierarchical structure based on the Art &amp; Architecture Thesaurus (AAT). We developed several baseline models, incorporating multiple heads on a ConvNeXT backbone for multi-label image tagging on these facets, and fine-tuning a CLIP model with our image-text pairs. Our experiments to evaluate model robustness and generalization capabilities in two different test scenarios demonstrate the dataset\u2019s utility in improving multi-label classification tools that have the potential to alleviate cataloging tasks in the cultural heritage sector. The EUFCC-340K dataset is publicly available at <jats:ext-link xmlns:xlink=\"http:\/\/www.w3.org\/1999\/xlink\" xlink:href=\"https:\/\/github.com\/cesc47\/EUFCC-340K\" ext-link-type=\"uri\">https:\/\/github.com\/cesc47\/EUFCC-340K<\/jats:ext-link>.<\/jats:p>","DOI":"10.1007\/s11042-024-20561-9","type":"journal-article","created":{"date-parts":[[2025,1,28]],"date-time":"2025-01-28T23:53:35Z","timestamp":1738108415000},"page":"39259-39282","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["EUFCC-340K: A faceted hierarchical dataset for metadata annotation in GLAM collections"],"prefix":"10.1007","volume":"84","author":[{"given":"Francesc","family":"Net","sequence":"first","affiliation":[]},{"given":"Marc","family":"Folia","sequence":"additional","affiliation":[]},{"given":"Pep","family":"Casals","sequence":"additional","affiliation":[]},{"given":"Andrew D.","family":"Bagdanov","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1408-9803","authenticated-orcid":false,"given":"Lluis","family":"G\u00f3mez","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,1,29]]},"reference":[{"key":"20561_CR1","unstructured":"Van\u00a0Horn G, Perona P (2017) The devil is in the tails: Fine-grained classification in the wild. arXiv:1709.01450"},{"issue":"1","key":"20561_CR2","doi-asserted-by":"publisher","first-page":"122","DOI":"10.1162\/dint_a_00162","volume":"5","author":"M Wu","year":"2023","unstructured":"Wu M, Brandhorst H, Marinescu M-C, Lopez JM, Hlava M, Busch J (2023) Automated metadata annotation: What is and is not possible with machine learning. Data Intell 5(1):122\u2013138","journal-title":"Data Intell"},{"key":"20561_CR3","doi-asserted-by":"crossref","unstructured":"Veit A, Nickel M, Belongie S, Maaten L (2018) Separating self-expression and visual content in hashtag supervision. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5919\u20135927","DOI":"10.1109\/CVPR.2018.00620"},{"key":"20561_CR4","unstructured":"Getty Research Institute (2024) Art & Architecture Thesaurus (AAT). [Online; accessed 16-Nov-2024]. https:\/\/www.getty.edu\/research\/tools\/vocabularies\/aat\/"},{"issue":"11","key":"20561_CR5","doi-asserted-by":"publisher","first-page":"1958","DOI":"10.1109\/TPAMI.2008.128","volume":"30","author":"A Torralba","year":"2008","unstructured":"Torralba A, Fergus R, Freeman WT (2008) 80 million tiny images: A large data set for nonparametric object and scene recognition. IEEE Trans Pattern Anal Mach Intell 30(11):1958\u20131970","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"20561_CR6","doi-asserted-by":"crossref","unstructured":"Fellbaum C (1998) WordNet: An electronic lexical database. MIT Press","DOI":"10.7551\/mitpress\/7287.001.0001"},{"key":"20561_CR7","doi-asserted-by":"crossref","unstructured":"Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L (2009) Imagenet: A large-scale hierarchical image database. In: 2009 IEEE conference on computer vision and pattern recognition, pp 248\u2013255. IEEE","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"20561_CR8","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, Huang Z, Karpathy A, Khosla A et al (2015) Imagenet large scale visual recognition challenge. Int J Comput Vis 115:211\u2013252","journal-title":"Int J Comput Vis"},{"key":"20561_CR9","doi-asserted-by":"publisher","first-page":"328","DOI":"10.1016\/j.eswa.2018.09.022","volume":"116","author":"Y Seo","year":"2019","unstructured":"Seo Y, Shin K-S (2019) Hierarchical convolutional neural networks for fashion image classification. Expert Syst Appl 116:328\u2013339","journal-title":"Expert Syst Appl"},{"key":"20561_CR10","unstructured":"Xiao H, Rasul K, Vollgraf R (2017) Fashion-MNIST: A novel image dataset for benchmarking machine learning algorithms"},{"issue":"7","key":"20561_CR11","doi-asserted-by":"publisher","first-page":"1956","DOI":"10.1007\/s11263-020-01316-z","volume":"128","author":"A Kuznetsova","year":"2020","unstructured":"Kuznetsova A, Rom H, Alldrin N, Uijlings J, Krasin I, Pont-Tuset J, Kamali S, Popov S, Malloci M, Kolesnikov A et al (2020) The open images dataset v4: Unified image classification, object detection, and visual relationship detection at scale. Int J Comput Vis 128(7):1956\u20131981","journal-title":"Int J Comput Vis"},{"key":"20561_CR12","doi-asserted-by":"crossref","unstructured":"Van\u00a0Horn G, Mac\u00a0Aodha O, Song Y, Cui Y, Sun C, Shepard A, Adam H, Perona P, Belongie S (2018) The inaturalist species classification and detection dataset. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8769\u20138778","DOI":"10.1109\/CVPR.2018.00914"},{"key":"20561_CR13","doi-asserted-by":"crossref","unstructured":"Van\u00a0Horn G, Cole E, Beery S, Wilber K, Belongie S, Mac\u00a0Aodha O (2021) Benchmarking representation learning for natural world image collections. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 12884\u201312893","DOI":"10.1109\/CVPR46437.2021.01269"},{"key":"20561_CR14","doi-asserted-by":"publisher","first-page":"3320","DOI":"10.3389\/fpls.2021.787127","volume":"12","author":"R Lutio","year":"2022","unstructured":"Lutio R, Park JY, Watson KA, D\u2019Aronco S, Wegner JD, Wieringa JJ, Tulig M, Pyle RL, Gallaher TJ, Brown G et al (2022) The herbarium 2021 half-earth challenge dataset and machine learning competition. Front Plant Sci 12:3320","journal-title":"Front Plant Sci"},{"key":"20561_CR15","doi-asserted-by":"crossref","unstructured":"Parmar G, Zhang R, Zhu J-Y (2022) On aliased resizing and surprising subtleties in gan evaluation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 11410\u201311420","DOI":"10.1109\/CVPR52688.2022.01112"},{"key":"20561_CR16","doi-asserted-by":"crossref","unstructured":"Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z (2016) Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2818\u20132826","DOI":"10.1109\/CVPR.2016.308"},{"key":"20561_CR17","doi-asserted-by":"crossref","unstructured":"Liu Z, Mao H, Wu C-Y, Feichtenhofer C, Darrell T, Xie S (2022) A convnet for the 2020s. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 11976\u201311986","DOI":"10.1109\/CVPR52688.2022.01167"},{"key":"20561_CR18","doi-asserted-by":"crossref","unstructured":"Sun C, Shrivastava A, Singh S, Gupta A (2017) Revisiting unreasonable effectiveness of data in deep learning era. In: Proceedings of the IEEE international conference on computer vision, pp 843\u2013852","DOI":"10.1109\/ICCV.2017.97"},{"key":"20561_CR19","doi-asserted-by":"crossref","unstructured":"Joulin A, Van Der\u00a0Maaten L, Jabri A, Vasilache N (2016) Learning visual features from large weakly supervised data. In: Computer vision\u2013ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11\u201314, 2016, Proceedings, Part VII 14, pp 67\u201384. Springer","DOI":"10.1007\/978-3-319-46478-7_5"},{"key":"20561_CR20","doi-asserted-by":"crossref","unstructured":"Redmon J, Farhadi A (2017) Yolo9000: better, faster, stronger. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7263\u20137271","DOI":"10.1109\/CVPR.2017.690"},{"key":"20561_CR21","unstructured":"Radford A, Kim JW, Hallacy C, Ramesh A, Goh G, Agarwal S, Sastry G, Askell A, Mishkin P, Clark J et al (2021) Learning transferable visual models from natural language supervision. In: International conference on machine learning, pp 8748\u20138763. PMLR"},{"key":"20561_CR22","unstructured":"Lakshminarayanan B, Pritzel A, Blundell C (2017) Simple and scalable predictive uncertainty estimation using deep ensembles. Advances in neural information processing systems, vol 30"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-20561-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-024-20561-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-20561-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,26]],"date-time":"2025-09-26T15:12:29Z","timestamp":1758899549000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-024-20561-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,1,29]]},"references-count":22,"journal-issue":{"issue":"32","published-online":{"date-parts":[[2025,9]]}},"alternative-id":["20561"],"URL":"https:\/\/doi.org\/10.1007\/s11042-024-20561-9","relation":{},"ISSN":["1573-7721"],"issn-type":[{"type":"electronic","value":"1573-7721"}],"subject":[],"published":{"date-parts":[[2025,1,29]]},"assertion":[{"value":"16 April 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 November 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 December 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 January 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare the following financial interests\/personal relationships which may be considered as potential competing interests: This work was supported by the ACCIO INNOTEC 2021 project Coeli-IA (ACE034\/21\/000084) and the CERCA Programme\/Generalitat de Catalunya. Lluis Gomez is funded by the Ramon y Cajal research fellowship RYC2020-030777-I \/ AEI \/ 10.13039\/501100011033.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of Interest\/Competing Interests"}}]}}