{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,15]],"date-time":"2026-03-15T15:57:22Z","timestamp":1773590242147,"version":"3.50.1"},"reference-count":58,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,7,22]],"date-time":"2025-07-22T00:00:00Z","timestamp":1753142400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,7,22]],"date-time":"2025-07-22T00:00:00Z","timestamp":1753142400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"DOI":"10.13039\/501100007601","name":"Horizon 2020","doi-asserted-by":"publisher","award":["101137074"],"award-info":[{"award-number":["101137074"]}],"id":[{"id":"10.13039\/501100007601","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100007601","name":"Horizon 2020","doi-asserted-by":"publisher","award":["101137074"],"award-info":[{"award-number":["101137074"]}],"id":[{"id":"10.13039\/501100007601","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100007601","name":"Horizon 2020","doi-asserted-by":"publisher","award":["101137074"],"award-info":[{"award-number":["101137074"]}],"id":[{"id":"10.13039\/501100007601","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100007601","name":"Horizon 2020","doi-asserted-by":"publisher","award":["101137074"],"award-info":[{"award-number":["101137074"]}],"id":[{"id":"10.13039\/501100007601","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100006108","name":"National Center for Advancing Translational Sciences","doi-asserted-by":"publisher","award":["TR004399"],"award-info":[{"award-number":["TR004399"]}],"id":[{"id":"10.13039\/100006108","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100006108","name":"National Center for Advancing Translational Sciences","doi-asserted-by":"publisher","award":["TR004399"],"award-info":[{"award-number":["TR004399"]}],"id":[{"id":"10.13039\/100006108","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100006108","name":"National Center for Advancing Translational Sciences","doi-asserted-by":"publisher","award":["TR004399"],"award-info":[{"award-number":["TR004399"]}],"id":[{"id":"10.13039\/100006108","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100006108","name":"National Center for Advancing Translational Sciences","doi-asserted-by":"publisher","award":["TR004399"],"award-info":[{"award-number":["TR004399"]}],"id":[{"id":"10.13039\/100006108","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100006108","name":"National Center for Advancing Translational Sciences","doi-asserted-by":"publisher","award":["TR004399"],"award-info":[{"award-number":["TR004399"]}],"id":[{"id":"10.13039\/100006108","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100001818","name":"Research to Prevent Blindness","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100001818","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["npj Digit. Med."],"DOI":"10.1038\/s41746-025-01801-0","type":"journal-article","created":{"date-parts":[[2025,7,22]],"date-time":"2025-07-22T04:51:24Z","timestamp":1753159884000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Re-identification of patients from imaging features extracted by foundation models"],"prefix":"10.1038","volume":"8","author":[{"given":"Giacomo","family":"Nebbia","sequence":"first","affiliation":[]},{"given":"Sourav","family":"Kumar","sequence":"additional","affiliation":[]},{"given":"Stephen Michael","family":"McNamara","sequence":"additional","affiliation":[]},{"given":"Christopher","family":"Bridge","sequence":"additional","affiliation":[]},{"given":"J. Peter","family":"Campbell","sequence":"additional","affiliation":[]},{"given":"Michael F.","family":"Chiang","sequence":"additional","affiliation":[]},{"given":"Naresh","family":"Mandava","sequence":"additional","affiliation":[]},{"given":"Praveer","family":"Singh","sequence":"additional","affiliation":[]},{"given":"Jayashree","family":"Kalpathy-Cramer","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,7,22]]},"reference":[{"key":"1801_CR1","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1016\/j.ejrad.2019.02.038","volume":"114","author":"L Saba","year":"2019","unstructured":"Saba, L. et al. The present and future of deep learning in radiology. Eur. J. Radiol. 114, 14\u201324 (2019).","journal-title":"Eur. J. Radiol."},{"key":"1801_CR2","doi-asserted-by":"publisher","DOI":"10.1148\/ryai.210064","volume":"4","author":"AC Yu","year":"2022","unstructured":"Yu, A. C., Mohajer, B. & Eng, J. External validation of deep learning algorithms for radiologic diagnosis: a systematic review. Radiol. Artif. Intell. 4, e210064 (2022).","journal-title":"Radiol. Artif. Intell."},{"key":"1801_CR3","doi-asserted-by":"publisher","first-page":"101878","DOI":"10.1016\/j.artmed.2020.101878","volume":"106","author":"MMA Monshi","year":"2020","unstructured":"Monshi, M. M. A., Poon, J. & Chung, V. Deep learning in generating radiology reports: a survey. Artif. Intell. Med. 106, 101878 (2020).","journal-title":"Artif. Intell. Med."},{"key":"1801_CR4","doi-asserted-by":"publisher","first-page":"101813","DOI":"10.1016\/j.media.2020.101813","volume":"67","author":"CL Srinidhi","year":"2021","unstructured":"Srinidhi, C. L., Ciga, O. & Martel, A. L. Deep neural network models for computational histopathology: a survey. Med. Image Anal. 67, 101813 (2021).","journal-title":"Med. Image Anal."},{"key":"1801_CR5","doi-asserted-by":"publisher","first-page":"686","DOI":"10.1038\/s41416-020-01122-x","volume":"124","author":"A Echle","year":"2021","unstructured":"Echle, A. et al. Deep learning in cancer pathology: a new generation of clinical biomarkers. Br. J. Cancer 124, 686\u2013696 (2021).","journal-title":"Br. J. Cancer"},{"key":"1801_CR6","doi-asserted-by":"publisher","first-page":"775","DOI":"10.1038\/s41591-021-01343-4","volume":"27","author":"J Van der Laak","year":"2021","unstructured":"Van der Laak, J., Litjens, G. & Ciompi, F. Deep learning in histopathology: the path to the clinic. Nat. Med. 27, 775\u2013784 (2021).","journal-title":"Nat. Med."},{"key":"1801_CR7","doi-asserted-by":"publisher","first-page":"2518","DOI":"10.1038\/s41433-022-02366-y","volume":"37","author":"A Ramanathan","year":"2023","unstructured":"Ramanathan, A., Athikarisamy, S. E. & Lam, G. C. Artificial intelligence for the diagnosis of retinopathy of prematurity: a systematic review of current algorithms. Eye 37, 2518\u20132526 (2023).","journal-title":"Eye"},{"key":"1801_CR8","doi-asserted-by":"publisher","first-page":"101971","DOI":"10.1016\/j.media.2021.101971","volume":"69","author":"T Li","year":"2021","unstructured":"Li, T. et al. Applications of deep learning in fundus images: a review. Med. Image Anal. 69, 101971 (2021).","journal-title":"Med. Image Anal."},{"key":"1801_CR9","doi-asserted-by":"publisher","first-page":"84","DOI":"10.3390\/jimaging9040084","volume":"9","author":"S Muchuchuti","year":"2023","unstructured":"Muchuchuti, S. & Viriri, S. Retinal disease detection using deep learning techniques: a comprehensive review. J. Imaging 9, 84 (2023).","journal-title":"J. Imaging"},{"key":"1801_CR10","first-page":"5998","volume":"30","author":"A Vaswani","year":"2017","unstructured":"Vaswani, A. et al. Attention is all you need. Adv. Neural Inf. Process. Syst. 30, 5998\u20136008 (2017).","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"1801_CR11","unstructured":"Dosovitskiy, A. et al. An image is worth 16x16 words: transformers for image recognition at scale. In International Conference on Learning Representations (2021)."},{"key":"1801_CR12","first-page":"25278","volume":"35","author":"C Schuhmann","year":"2022","unstructured":"Schuhmann, C. et al. Laion-5b: an open large-scale dataset for training next generation image-text models. Adv. Neural Inf. Process. Syst. 35, 25278\u201325294 (2022).","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"1801_CR13","doi-asserted-by":"publisher","unstructured":"Bommasani, R. et al. On the opportunities and risks of foundation models. Preprint at arXiv https:\/\/doi.org\/10.48550\/arXiv.2108.07258 (2021).","DOI":"10.48550\/arXiv.2108.07258"},{"key":"1801_CR14","doi-asserted-by":"publisher","first-page":"756","DOI":"10.1038\/s41551-023-01049-7","volume":"7","author":"S Azizi","year":"2023","unstructured":"Azizi, S. et al. Robust and data-efficient generalization of self-supervised machine learning for diagnostic imaging. Nat. Biomed. Eng. 7, 756\u2013779 (2023).","journal-title":"Nat. Biomed. Eng."},{"key":"1801_CR15","unstructured":"Kenton, J. D. M.-W. C. & Toutanova, L. K. BERT: pre-training of deep bidirectional transformers for language understanding. In Proc. NAACL-HLT. 4171\u20134186 (Association for Computational Linguistics, 2019)."},{"key":"1801_CR16","unstructured":"Radford, A., Narasimhan, K., Salimans, T., Sutskever, I. & others. Improving language understanding by generative pre-training (2018)"},{"key":"1801_CR17","unstructured":"Zhang, H. et al. DINO: DETR with improved denoising anchor boxes for end-to-end object detection. In The Eleventh International Conference on Learning Representations (2023)."},{"key":"1801_CR18","first-page":"31","volume":"1","author":"M Oquab","year":"2024","unstructured":"Oquab, M. et al. DINOv2: learning robust visual features without supervision. Trans. Mach. Learn. Res. J. 1, 31 (2024).","journal-title":"Trans. Mach. Learn. Res. J."},{"key":"1801_CR19","first-page":"10078","volume":"35","author":"Z Tong","year":"2022","unstructured":"Tong, Z., Song, Y., Wang, J. & Wang, L. Videomae: masked autoencoders are data-efficient learners for self-supervised video pre-training. Adv. Neural Inf. Process. Syst. 35, 10078\u201310093 (2022).","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"1801_CR20","doi-asserted-by":"crossref","unstructured":"Wang, L. et al. Videomae v2: scaling video masked autoencoders with dual masking. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 14549\u201314560 (IEEE, 2023).","DOI":"10.1109\/CVPR52729.2023.01398"},{"key":"1801_CR21","doi-asserted-by":"publisher","unstructured":"Bai, J. et al. Qwen-vl: a versatile vision-language model for understanding, localization, text reading, and beyond. Preprint at arXiv https:\/\/doi.org\/10.48550\/arXiv.2308.12966 (2023).","DOI":"10.48550\/arXiv.2308.12966"},{"key":"1801_CR22","unstructured":"Zhu, D., Chen, J., Shen, X., Li, X. & Elhoseiny, M. MiniGPT-4: enhancing vision-language understanding with advanced large language models. The Twelfth International Conference on Learning Representation (2024)."},{"key":"1801_CR23","doi-asserted-by":"publisher","first-page":"454","DOI":"10.1148\/radiol.212482","volume":"305","author":"AB Sellergren","year":"2022","unstructured":"Sellergren, A. B. et al. Simplified transfer learning for chest radiography models using less data. Radiology 305, 454\u2013465 (2022).","journal-title":"Radiology"},{"key":"1801_CR24","unstructured":"Yao, J. et al. EVA-X: a foundation model for general chest X-ray analysis with self-supervised learning. CoRR abs\/2405.05237 (2024)."},{"key":"1801_CR25","unstructured":"Xu, S. et al. ELIXR: towards a general purpose X-ray artificial intelligence system through alignment of large language models and radiology vision encoders. CoRR abs\/2308.01317 (2023)."},{"key":"1801_CR26","unstructured":"Dippel, J. et al. RudolfV: a foundation model by pathologists for pathologists. CoRR abs\/2401.04079 (2024)."},{"key":"1801_CR27","doi-asserted-by":"publisher","unstructured":"Vorontsov, E. et al. A foundation model for clinical-grade computational pathology and rare cancers detection. Nat. Med. https:\/\/doi.org\/10.1038\/s41591-024-03141-0 (2024).","DOI":"10.1038\/s41591-024-03141-0"},{"key":"1801_CR28","doi-asserted-by":"publisher","first-page":"2307","DOI":"10.1038\/s41591-023-02504-3","volume":"29","author":"Z Huang","year":"2023","unstructured":"Huang, Z., Bianchi, F., Yuksekgonul, M., Montine, T. J. & Zou, J. A visual\u2013language foundation model for pathology image analysis using medical Twitter. Nat. Med. 29, 2307\u20132316 (2023).","journal-title":"Nat. Med."},{"key":"1801_CR29","doi-asserted-by":"publisher","first-page":"156","DOI":"10.1038\/s41586-023-06555-x","volume":"622","author":"Y Zhou","year":"2023","unstructured":"Zhou, Y. et al. A foundation model for generalizable disease detection from retinal images. Nature 622, 156\u2013163 (2023).","journal-title":"Nature"},{"key":"1801_CR30","doi-asserted-by":"crossref","unstructured":"Haghighi, T. et al. EYE-Llama, an in-domain large language model for ophthalmology. iScience 112984 (2025).","DOI":"10.1016\/j.isci.2025.112984"},{"key":"1801_CR31","doi-asserted-by":"publisher","unstructured":"Chia, M. A. et al. Foundation models in ophthalmology. Br. J. Ophthalmol. bjo-2024-325459 https:\/\/doi.org\/10.1136\/bjo-2024-325459 (2024).","DOI":"10.1136\/bjo-2024-325459"},{"key":"1801_CR32","unstructured":"Ma, J. et al. Segment anything in medical images and videos: benchmark and deployment. Preprint at http:\/\/arxiv.org\/abs\/2408.03322 (2024)."},{"key":"1801_CR33","unstructured":"Li, C. et al. LLaVA-Med: training a large language-and-vision assistant for biomedicine in one day. in Advances in Neural Information Processing Systems (eds Oh, A. et al.) 36 28541\u201328564 (Curran Associates, Inc., 2023)."},{"key":"1801_CR34","doi-asserted-by":"publisher","first-page":"136","DOI":"10.1038\/s41591-020-01192-7","volume":"27","author":"E Pierson","year":"2021","unstructured":"Pierson, E., Cutler, D. M., Leskovec, J., Mullainathan, S. & Obermeyer, Z. An algorithmic approach to reducing unexplained pain disparities in underserved populations. Nat. Med. 27, 136\u2013140 (2021).","journal-title":"Nat. Med."},{"key":"1801_CR35","doi-asserted-by":"publisher","first-page":"e406","DOI":"10.1016\/S2589-7500(22)00063-2","volume":"4","author":"JW Gichoya","year":"2022","unstructured":"Gichoya, J. W. et al. AI recognition of patient race in medical imaging: a modelling study. Lancet Digit. Health 4, e406\u2013e414 (2022).","journal-title":"Lancet Digit. Health"},{"key":"1801_CR36","doi-asserted-by":"publisher","unstructured":"Banerjee, I. et al. Reading race: AI recognises patient\u2019s racial identity in medical images. Preprint at arXiv https:\/\/doi.org\/10.48550\/arXiv.2107.10356 (2021).","DOI":"10.48550\/arXiv.2107.10356"},{"key":"1801_CR37","doi-asserted-by":"publisher","DOI":"10.1016\/j.ebiom.2024.105047","volume":"102","author":"R Wang","year":"2024","unstructured":"Wang, R. et al. Drop the shortcuts: image augmentation improves fairness and decreases AI detection of race and other demographics from medical images. eBioMedicine 102, 105047 (2024).","journal-title":"eBioMedicine"},{"key":"1801_CR38","doi-asserted-by":"publisher","first-page":"1714","DOI":"10.3390\/diagnostics12071714","volume":"12","author":"NC Khan","year":"2022","unstructured":"Khan, N. C. et al. Predicting systemic health features from retinal fundus images using transfer-learning-based artificial intelligence models. Diagnostics 12, 1714 (2022).","journal-title":"Diagnostics"},{"key":"1801_CR39","doi-asserted-by":"publisher","first-page":"e526","DOI":"10.1016\/S2589-7500(20)30216-8","volume":"2","author":"TH Rim","year":"2020","unstructured":"Rim, T. H. et al. Prediction of systemic biomarkers from retinal photographs: development and validation of deep-learning algorithms. Lancet Digit. Health 2, e526\u2013e536 (2020).","journal-title":"Lancet Digit. Health"},{"key":"1801_CR40","doi-asserted-by":"publisher","first-page":"543","DOI":"10.1001\/jamaophthalmol.2023.1310","volume":"141","author":"AS Coyner","year":"2023","unstructured":"Coyner, A. S. et al. Association of biomarker-based artificial intelligence with risk of racial bias in retinal images. JAMA Ophthalmol. 141, 543 (2023).","journal-title":"JAMA Ophthalmol."},{"key":"1801_CR41","doi-asserted-by":"crossref","unstructured":"American Academy of Ophthalmology Board of Trustees. Special Commentary: Balancing benefits and risks: the case for retinal images to be considered as nonprotected health information for research purposes. Ophthalmology 132, 115\u2013118 (2025).","DOI":"10.1016\/j.ophtha.2024.07.031"},{"key":"1801_CR42","doi-asserted-by":"publisher","first-page":"1941","DOI":"10.1007\/s10278-023-00850-9","volume":"36","author":"Y Ueda","year":"2023","unstructured":"Ueda, Y. & Morishita, J. Patient identification based on deep metric learning for preventing human errors in follow-up X-ray examinations. J. Digit. Imaging 36, 1941\u20131953 (2023).","journal-title":"J. Digit. Imaging"},{"key":"1801_CR43","doi-asserted-by":"publisher","DOI":"10.1038\/s41598-022-19045-3","volume":"12","author":"K Packh\u00e4user","year":"2022","unstructured":"Packh\u00e4user, K. et al. Deep learning-based patient re-identification is able to exploit the biometric nature of medical chest X-ray data. Sci. Rep. 12, 14851 (2022).","journal-title":"Sci. Rep."},{"key":"1801_CR44","doi-asserted-by":"publisher","first-page":"1124","DOI":"10.1007\/s10278-024-01017-w","volume":"37","author":"Y Ueda","year":"2024","unstructured":"Ueda, Y., Ogawa, D. & Ishida, T. Patient re-identification based on deep metric learning in trunk computed tomography images acquired from devices from different vendors. J. Imaging Inform. Med. 37, 1124\u20131136 (2024).","journal-title":"J. Imaging Inform. Med."},{"key":"1801_CR45","unstructured":"Tan, M. & Le, Q. Efficientnetv2: smaller models and faster training. In International Conference on Machine Learning 10096\u201310106 (PMLR, 2021)."},{"key":"1801_CR46","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S. & Sun, J. Deep residual learning for image recognition. In Proc. IEEE Conference on Computer Vision and Pattern Recognition. 770\u2013778 (IEEE Computer Society, 2016).","DOI":"10.1109\/CVPR.2016.90"},{"key":"1801_CR47","doi-asserted-by":"publisher","DOI":"10.1038\/s41597-023-02424-4","volume":"10","author":"X Huang","year":"2023","unstructured":"Huang, X. et al. GRAPE: a multi-modal dataset of longitudinal follow-up visual field and fundus images for glaucoma management. Sci. Data 10, 520 (2023).","journal-title":"Sci. Data"},{"key":"1801_CR48","doi-asserted-by":"crossref","unstructured":"Wang, X. et al. ChestX-ray8: hospital-scale chest X-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 3462\u20133471 (IEEE Computer Society, 2017).","DOI":"10.1109\/CVPR.2017.369"},{"key":"1801_CR49","first-page":"590","volume":"33","author":"J Irvin","year":"2019","unstructured":"Irvin, J. et al. CheXpert: a large chest radiograph dataset with uncertainty labels and expert comparison. Proc. AAAI Conf. Artif. Intell. 33, 590\u2013597 (2019).","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"1801_CR50","doi-asserted-by":"publisher","first-page":"158","DOI":"10.1038\/s41551-018-0195-0","volume":"2","author":"R Poplin","year":"2018","unstructured":"Poplin, R. et al. Prediction of cardiovascular risk factors from retinal fundus photographs via deep learning. Nat. Biomed. Eng. 2, 158\u2013164 (2018).","journal-title":"Nat. Biomed. Eng."},{"key":"1801_CR51","doi-asserted-by":"publisher","first-page":"e0233166","DOI":"10.1371\/journal.pone.0233166","volume":"15","author":"L Zhang","year":"2020","unstructured":"Zhang, L. et al. Prediction of hypertension, hyperglycemia and dyslipidemia from retinal fundus photographs via deep learning: a cross-sectional study of chronic diseases in central China. PLoS ONE 15, e0233166 (2020).","journal-title":"PLoS ONE"},{"key":"1801_CR52","doi-asserted-by":"publisher","first-page":"1151","DOI":"10.1016\/j.jacr.2022.06.008","volume":"19","author":"J Adleberg","year":"2022","unstructured":"Adleberg, J. et al. Predicting patient demographics from chest radiographs with deep learning. J. Am. Coll. Radiol. 19, 1151\u20131161 (2022).","journal-title":"J. Am. Coll. Radiol."},{"key":"1801_CR53","unstructured":"Jabbour, S., Fouhey, D., Kazerooni, E., Sjoding, M. W. & Wiens, J. Deep learning applied to chest X-rays: exploiting and preventing shortcuts. In Machine Learning for Healthcare Conference 750\u2013782 (PMLR, 2020)."},{"key":"1801_CR54","doi-asserted-by":"crossref","unstructured":"M\u00fcller, S., Koch, L. M., Lensch, H. P. A. & Berens, P. Disentangling representations of retinal images with generative models. Med. Image Anal. 103628 (2025).","DOI":"10.1016\/j.media.2025.103628"},{"key":"1801_CR55","doi-asserted-by":"publisher","first-page":"102516","DOI":"10.1016\/j.media.2022.102516","volume":"80","author":"X Liu","year":"2022","unstructured":"Liu, X., Sanchez, P., Thermos, S., O\u2019Neil, A. Q. & Tsaftaris, S. A. Learning disentangled representations in the imaging domain. Med. Image Anal. 80, 102516 (2022).","journal-title":"Med. Image Anal."},{"key":"1801_CR56","doi-asserted-by":"publisher","first-page":"3778","DOI":"10.1002\/mp.14929","volume":"48","author":"Y Fei","year":"2021","unstructured":"Fei, Y. et al. Deep learning-based multi-modal computing with feature disentanglement for MRI image synthesis. Med. Phys. 48, 3778\u20133789 (2021).","journal-title":"Med. Phys."},{"key":"1801_CR57","doi-asserted-by":"crossref","unstructured":"He, K. et al. Masked autoencoders are scalable vision learners. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 16000\u201316009 (IEEE, 2022).","DOI":"10.1109\/CVPR52688.2022.01553"},{"key":"1801_CR58","unstructured":"Chen, T., Kornblith, S., Norouzi, M. & Hinton, G. A simple framework for contrastive learning of visual representations. In Proc. 37th International Conference on Machine Learning (eds III, H. D. & Singh, A.) 119 1597\u20131607 (PMLR, 2020)."}],"container-title":["npj Digital Medicine"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.nature.com\/articles\/s41746-025-01801-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s41746-025-01801-0","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s41746-025-01801-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,7]],"date-time":"2025-09-07T17:21:55Z","timestamp":1757265715000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.nature.com\/articles\/s41746-025-01801-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,22]]},"references-count":58,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2025,12]]}},"alternative-id":["1801"],"URL":"https:\/\/doi.org\/10.1038\/s41746-025-01801-0","relation":{},"ISSN":["2398-6352"],"issn-type":[{"value":"2398-6352","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,7,22]]},"assertion":[{"value":"24 February 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 June 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 July 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"Author S.M. has previously acted as a paid consultant for Evolution Optiks but declares no non-financial competing interests. Author J.P.C. has received financial support from Genentech and Boston AI Lab and is the owner of Siloam Vision but declares no non-financial competing interests. Author N.M. has previously acted as a paid consultant for Soma Logic and ONL Therapeutics but declares no non-financial competing interests. Author J.K.C. has previously acted as a paid consultant for Siloam Vision and has received financial support from Genentech and Boston AI Lab but declares no non-financial competing interests. All other authors declare no financial or non-financial competing interests.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"469"}}