{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,10]],"date-time":"2026-02-10T07:28:26Z","timestamp":1770708506787,"version":"3.49.0"},"reference-count":55,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"funder":[{"name":"Democritus University of Thrace;"},{"name":"Neapolis Academic Enterprises Ltd.; and"},{"name":"European Union and Greek National Funds through the Operational Program Competitiveness, Entrepreneurship and Innovation, by the Call RESEARCH-CREATE-INNOVATE"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2023]]},"DOI":"10.1109\/access.2023.3267804","type":"journal-article","created":{"date-parts":[[2023,4,17]],"date-time":"2023-04-17T18:07:06Z","timestamp":1681754826000},"page":"38208-38217","source":"Crossref","is-referenced-by-count":3,"title":["Universal Image Embedding: Retaining and Expanding Knowledge With Multi-Domain Fine-Tuning"],"prefix":"10.1109","volume":"11","author":[{"given":"Socratis","family":"Gkelios","sequence":"first","affiliation":[{"name":"Department of Electrical and Computer Engineering, Democritus University of Thrace, Xanthi, Kimmeria, Greece"}]},{"given":"Anestis","family":"Kastellos","sequence":"additional","affiliation":[{"name":"Department of Computer Science, Intelligent Systems Laboratory, Neapolis University Pafos, Paphos, Cyprus"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3717-3565","authenticated-orcid":false,"given":"Yiannis S.","family":"Boutalis","sequence":"additional","affiliation":[{"name":"Department of Electrical and Computer Engineering, Democritus University of Thrace, Xanthi, Kimmeria, Greece"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4657-4435","authenticated-orcid":false,"given":"Savvas A.","family":"Chatzichristofis","sequence":"additional","affiliation":[{"name":"Department of Computer Science, Intelligent Systems Laboratory, Neapolis University Pafos, Paphos, Cyprus"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-015-0816-y"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2021.114940"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/DCOSS52077.2021.00065"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2014.131"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58558-7_29"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/BIBM55620.2022.9995049"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1002\/int.22957"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2021.107338"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00780"},{"key":"ref10","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Radford"},{"key":"ref11","first-page":"1","article-title":"LAION-5B: An open large-scale dataset for training next generation image-text models","volume-title":"Proc. 26th Conf. Neural Inf. Process. Syst. Datasets Benchmarks Track","author":"Schuhmann"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00482"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58621-8_43"},{"key":"ref14","article-title":"Google landmark recognition 2020 competition third place solution","author":"Ha","year":"2020","journal-title":"arXiv:2010.05350"},{"key":"ref15","first-page":"4904","article-title":"Scaling up visual and vision-language representation learning with noisy text supervision","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Jia"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01759"},{"key":"ref17","article-title":"Florence: A new foundation model for computer vision","author":"Yuan","year":"2021","journal-title":"arXiv:2111.11432"},{"key":"ref18","first-page":"9694","article-title":"Align before fuse: Vision and language representation learning with momentum distillation","volume-title":"Proc. Adv. Neural Inf. Process. Syst. (NIPS)","volume":"34","author":"Li"},{"key":"ref19","article-title":"CoCa: Contrastive captioners are image-text foundation models","author":"Yu","year":"2022","journal-title":"arXiv:2205.01917"},{"issue":"2","key":"ref20","first-page":"207","article-title":"Distance metric learning for large margin nearest neighbor classification","volume":"10","author":"Weinberger","year":"2009","journal-title":"J. Mach. Learn. Res."},{"key":"ref21","first-page":"1","article-title":"Improved deep metric learning with multi-class N-pair loss objective","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"29","author":"Sohn"},{"key":"ref22","article-title":"Representation learning with contrastive predictive coding","author":"Van Den Oord","year":"2018","journal-title":"arXiv:1807.03748"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.5555\/3524938.3525087"},{"key":"ref24","first-page":"18661","article-title":"Supervised contrastive learning","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"33","author":"Khosla"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00516"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1145\/3123266.3123359"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.47"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00552"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.713"},{"key":"ref30","article-title":"Reproducible scaling laws for contrastive language-image learning","author":"Cherti","year":"2022","journal-title":"arXiv:2212.07143"},{"key":"ref31","article-title":"An image is worth 16\u00d716 words: Transformers for image recognition at scale","author":"Dosovitskiy","year":"2020","journal-title":"arXiv:2010.11929"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.03762"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1145\/2072298.2072034"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-540-88682-2_24"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2007.383172"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2008.4587635"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1145\/2700292"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW.2013.77"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.124"},{"key":"ref40","article-title":"Caltech-UCSD birds 200","author":"Welinder","year":"2010"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00265"},{"key":"ref42","article-title":"RP2K: A large-scale retail product dataset for fine-grained image classification","author":"Peng","year":"2020","journal-title":"arXiv:2006.12634"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-10599-4_29"},{"key":"ref44","article-title":"Products-10K: A large-scale product recognition dataset","author":"Bai","year":"2020","journal-title":"arXiv:2008.10545"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.434"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.163"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-48881-3_2"},{"key":"ref48","first-page":"1","article-title":"The met dataset: Instance-level recognition for artworks","volume-title":"Proc. 34th Conf. Neural Inf. Process. Syst. Datasets Benchmarks Track","author":"Ypsilantis"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1109\/WACV56688.2023.00019"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01156"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-92273-3_18"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00726"},{"key":"ref53","first-page":"1","article-title":"Guided similarity separation for image retrieval","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"32","author":"Liu"},{"key":"ref54","article-title":"Boosting performance of a baseline visual place recognition technique by predicting the maximally complementary technique","author":"Malone","year":"2022","journal-title":"arXiv:2210.07509"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1016\/j.jvcir.2021.103282"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6287639\/10005208\/10103696.pdf?arnumber=10103696","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,14]],"date-time":"2024-03-14T01:41:36Z","timestamp":1710380496000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10103696\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"references-count":55,"URL":"https:\/\/doi.org\/10.1109\/access.2023.3267804","relation":{},"ISSN":["2169-3536"],"issn-type":[{"value":"2169-3536","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023]]}}}