{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,30]],"date-time":"2026-03-30T22:18:50Z","timestamp":1774909130147,"version":"3.50.1"},"reference-count":164,"publisher":"Springer Science and Business Media LLC","issue":"8092","license":[{"start":{"date-parts":[[2025,11,5]],"date-time":"2025-11-05T00:00:00Z","timestamp":1762300800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2025,11,5]],"date-time":"2025-11-05T00:00:00Z","timestamp":1762300800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Nature"],"published-print":{"date-parts":[[2025,12,4]]},"abstract":"<jats:title>Abstract<\/jats:title>\n                  <jats:p>\n                    Computer vision is central to many artificial intelligence (AI) applications, from autonomous vehicles to consumer devices. However, the data behind such technical innovations are often collected with insufficient consideration of ethical concerns\n                    <jats:sup>1\u20133<\/jats:sup>\n                    . This has led to a reliance on datasets that lack diversity, perpetuate biases and are collected without the consent of data rights holders. These datasets compromise the fairness and accuracy of AI models and disenfranchise stakeholders\n                    <jats:sup>4\u20138<\/jats:sup>\n                    . Although awareness of the problems of bias in computer vision technologies, particularly facial recognition, has become widespread\n                    <jats:sup>9<\/jats:sup>\n                    , the field lacks publicly available, consensually collected datasets for evaluating bias for most tasks\n                    <jats:sup>3,10,11<\/jats:sup>\n                    . In response, we introduce the Fair Human-Centric Image Benchmark (FHIBE, pronounced \u2018Feebee\u2019), a publicly available human image dataset implementing best practices for consent, privacy, compensation, safety, diversity and utility. FHIBE can be used responsibly as a fairness evaluation dataset for many human-centric computer vision tasks, including pose estimation, person segmentation, face detection and verification, and visual question answering. By leveraging comprehensive annotations capturing demographic and physical attributes, environmental factors, instrument and pixel-level annotations, FHIBE can identify a wide variety of biases. The annotations also enable more nuanced and granular bias diagnoses, enabling practitioners to better understand sources of bias and mitigate potential downstream harms. FHIBE therefore represents an important step forward towards trustworthy AI, raising the bar for fairness benchmarks and providing a road map for responsible data curation in AI.\n                  <\/jats:p>","DOI":"10.1038\/s41586-025-09716-2","type":"journal-article","created":{"date-parts":[[2025,11,5]],"date-time":"2025-11-05T16:03:02Z","timestamp":1762358582000},"page":"97-108","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Fair human-centric image dataset for ethical AI benchmarking"],"prefix":"10.1038","volume":"648","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7907-9353","authenticated-orcid":false,"given":"Alice","family":"Xiang","sequence":"first","affiliation":[]},{"given":"Jerone T. A.","family":"Andrews","sequence":"additional","affiliation":[]},{"given":"Rebecca L.","family":"Bourke","sequence":"additional","affiliation":[]},{"given":"William","family":"Thong","sequence":"additional","affiliation":[]},{"given":"Julienne M.","family":"LaChance","sequence":"additional","affiliation":[]},{"given":"Tiffany","family":"Georgievski","sequence":"additional","affiliation":[]},{"given":"Apostolos","family":"Modas","sequence":"additional","affiliation":[]},{"given":"Aida","family":"Rahmattalabbi","sequence":"additional","affiliation":[]},{"given":"Yunhao","family":"Ba","sequence":"additional","affiliation":[]},{"given":"Shruti","family":"Nagpal","sequence":"additional","affiliation":[]},{"given":"Orestis","family":"Papakyriakopoulos","sequence":"additional","affiliation":[]},{"given":"Dora","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Jinru","family":"Xue","sequence":"additional","affiliation":[]},{"given":"Victoria","family":"Matthews","sequence":"additional","affiliation":[]},{"given":"Linxia","family":"Gong","sequence":"additional","affiliation":[]},{"given":"Austin T.","family":"Hoag","sequence":"additional","affiliation":[]},{"given":"Mircea","family":"Cimpoi","sequence":"additional","affiliation":[]},{"given":"Swami","family":"Sankaranarayanan","sequence":"additional","affiliation":[]},{"given":"Wiebke","family":"Hutiri","sequence":"additional","affiliation":[]},{"given":"Morgan K.","family":"Scheuerman","sequence":"additional","affiliation":[]},{"given":"Albert S.","family":"Abedi","sequence":"additional","affiliation":[]},{"given":"Peter","family":"Stone","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9349-0624","authenticated-orcid":false,"given":"Peter R.","family":"Wurman","sequence":"additional","affiliation":[]},{"given":"Hiroaki","family":"Kitano","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9443-7008","authenticated-orcid":false,"given":"Michael","family":"Spranger","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,11,5]]},"reference":[{"key":"9716_CR1","doi-asserted-by":"crossref","unstructured":"Sambasivan, N. et al. \u201cEveryone wants to do the model work, not the data work\u201d: data cascades in high-stakes AI. In Proc. ACM CHI Conference on Human Factors in Computing Systems (ACM, 2021).","DOI":"10.1145\/3411764.3445518"},{"key":"9716_CR2","doi-asserted-by":"crossref","unstructured":"Birhane, A. & Prabhu, V. U. Large image datasets: a pyrrhic win for computer vision? In Proc. IEEE Winter Conference on Applications of Computer Vision (WACV) 1536\u20131546 (IEEE, 2021).","DOI":"10.1109\/WACV48630.2021.00158"},{"key":"9716_CR3","unstructured":"Andrews, J. T. et al. Ethical considerations for collecting human-centric image datasets. In Proc. Advances in Neural Information Processing Systems Datasets and Benchmarks Track (NeurIPS D&B) 55320\u201355360 (Curran Associates, 2023)."},{"key":"9716_CR4","doi-asserted-by":"crossref","unstructured":"Hundt, A., Agnew, W., Zeng, V., Kacianka, S. & Gombolay, M. Robots enact malignant stereotypes. In Proc. ACM Conference on Fairness, Accountability, and Transparency (FAccT) 743\u2013756 (ACM, 2022).","DOI":"10.1145\/3531146.3533138"},{"key":"9716_CR5","unstructured":"Wilson, B., Hoffman, J. & Morgenstern, J. Predictive inequity in object detection. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) (IEEE, 2019)."},{"key":"9716_CR6","unstructured":"Birhane, A. et al. Into the LAION\u2019s den: investigating hate in multimodal datasets. In Proc. Advances in Neural Information Processing Systems Datasets and Benchmarks Track (NeurIPS D&B) 21268\u201321284 (2024)."},{"key":"9716_CR7","doi-asserted-by":"crossref","unstructured":"Xiang, A. Being \u2018seen\u2019 vs. \u2018mis-seen\u2019: tensions between privacy and fairness in computer vision. Harvard J. Law Technol. 36, 1\u201360 (2022).","DOI":"10.2139\/ssrn.4068921"},{"key":"9716_CR8","unstructured":"Peng, K., Mathur, A. & Narayanan, A. Mitigating dataset harms requires stewardship: lessons from 1000 papers. In Proc. Advances in Neural Information Processing Systems Datasets and Benchmarks Track (NeurIPS D&B) (Curran Associates, 2021)."},{"key":"9716_CR9","unstructured":"Buolamwini, J. & Gebru, T. Gender shades: intersectional accuracy disparities in commercial gender classification. In Proc. ACM Conference on Fairness, Accountability, and Transparency (FAccT) 77\u201391 (ACM, 2018)."},{"key":"9716_CR10","doi-asserted-by":"crossref","unstructured":"Bergman, A. S. et al. Representation in AI evaluations. In Proc. ACM Conference on Fairness, Accountability, and Transparency (FAccT) 519\u2013533 (ACM, 2023).","DOI":"10.1145\/3593013.3594019"},{"key":"9716_CR11","doi-asserted-by":"crossref","unstructured":"Holstein, K., Wortman Vaughan, J., Daum\u00e9 III, H., Dudik, M. & Wallach, H. Improving fairness in machine learning systems: what do industry practitioners need? In Proc. Conference on Human Factors in Computing Systems (CHI) 1\u201316 (ACM, 2019).","DOI":"10.1145\/3290605.3300830"},{"key":"9716_CR12","doi-asserted-by":"crossref","unstructured":"Deng, J. et al. Imagenet: a large-scale hierarchical image database. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 248\u2013255 (IEEE, 2009).","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"9716_CR13","doi-asserted-by":"crossref","unstructured":"Li, F.-F. & Krishna, R. Searching for computer vision North Stars. Daedalus 151, 2 (2022).","DOI":"10.1162\/daed_a_01902"},{"key":"9716_CR14","unstructured":"Lee, N. et al. Survey of social bias in vision-language models. Preprint at arxiv.org\/abs\/2309.14381 (2023)."},{"key":"9716_CR15","doi-asserted-by":"crossref","unstructured":"Andrus, M., Spitzer, E., Brown, J. & Xiang, A. What we can\u2019t measure, we can\u2019t understand: challenges to demographic data procurement in the pursuit of fairness. In Proc. ACM Conference on Fairness, Accountability, and Transparency (FAccT)249\u2013260 (ACM, 2021).;","DOI":"10.1145\/3442188.3445888"},{"key":"9716_CR16","unstructured":"Zhao, D., Andrews, J. T. A. & Xiang, A. Men also do laundry: multi-attribute bias amplification. In Proc. 40th International Conference on Machine Learning 42000\u201342017 (ACM, 2023)."},{"key":"9716_CR17","doi-asserted-by":"publisher","unstructured":"Thiel, D. Identifying and Eliminating CSAM in Generative ML Training Data and Models (Stanford Univ., 2023); https:\/\/doi.org\/10.25740\/kh752sm9123.","DOI":"10.25740\/kh752sm9123"},{"key":"9716_CR18","doi-asserted-by":"crossref","unstructured":"Yew, R.-J. & Xiang, A. Regulating facial processing technologies: tensions between legal and technical considerations in the application of Illinois BIPA. In Proc. ACM Conference on Fairness, Accountability, and Transparency (FAccT) 1017\u20131027 (ACM, 2022).","DOI":"10.1145\/3531146.3533163"},{"key":"9716_CR19","doi-asserted-by":"publisher","first-page":"001","DOI":"10.1093\/cybsec\/tyy001","volume":"4","author":"E Politou","year":"2018","unstructured":"Politou, E., Alepis, E. & Patsakis, C. Forgetting personal data and revoking consent under the GDPR: challenges and proposed solutions. J. Cybersec. 4, 001 (2018).","journal-title":"J. Cybersec."},{"key":"9716_CR20","doi-asserted-by":"crossref","unstructured":"Longpre, S. et al. A large-scale audit of dataset licensing and attribution in AI. Nat. Mach. Intell. 6, 975\u2013987 (2024).","DOI":"10.1038\/s42256-024-00878-8"},{"key":"9716_CR21","unstructured":"Gray, M. L. & Suri, S. Ghost Work: How to Stop Silicon Valley from Building a New Global Underclass (Eamon Dolan Books, 2019)."},{"key":"9716_CR22","doi-asserted-by":"crossref","unstructured":"Wang, D., Prabhat, S. & Sambasivan, N. Whose AI dream? In search of the aspiration in data annotation. In Proc. 2022 CHI Conference on Human Factors in Computing Systems 1\u201316 (ACM, 2022).","DOI":"10.1145\/3491102.3502121"},{"key":"9716_CR23","unstructured":"Grother, P., Ngan, M., Hanaoka, K., Yang, J. C. & Hom, A. Face Recognition Technology Evaluation (FRTE) Part 1: Verification Technical Report (NIST, 2025); www.nist.gov\/programs-projects\/face-recognition-vendor-test-frvt-ongoing."},{"key":"9716_CR24","doi-asserted-by":"crossref","unstructured":"Gustafson, L. et al. FACET: fairness in computer vision evaluation benchmark. In Proc. International Conference on Computer Vision (ICCV) 20370\u201320382 (IEEE, 2023).","DOI":"10.1109\/ICCV51070.2023.01863"},{"key":"9716_CR25","doi-asserted-by":"crossref","unstructured":"Hazirbas, C. et al. Casual conversations: a dataset for measuring fairness in AI. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) 2289\u20132293 (IEEE, 2021).","DOI":"10.1109\/CVPRW53098.2021.00258"},{"key":"9716_CR26","doi-asserted-by":"crossref","unstructured":"Porgali, B., Albiero, V., Ryda, J., Ferrer, C. C. & Hazirbas, C. The Casual Conversations v2 dataset. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition 10\u201317 (IEEE, 2023).","DOI":"10.1109\/CVPRW59228.2023.00006"},{"key":"9716_CR27","doi-asserted-by":"publisher","first-page":"1122","DOI":"10.3758\/s13428-014-0532-5","volume":"47","author":"DS Ma","year":"2015","unstructured":"Ma, D. S., Correll, J. & Wittenbrink, B. The Chicago Face Database: a free stimulus set of faces and norming data. Behav. Res. Methods 47, 1122\u20131135 (2015).","journal-title":"Behav. Res. Methods"},{"key":"9716_CR28","doi-asserted-by":"publisher","unstructured":"Dhar, P., Gleason, J., Roy, A., Castillo, C. D. & Chellappa, R. Pass: protected attribute suppression system for mitigating bias in face recognition. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 15087\u201315096 (IEEE, 2023); https:\/\/doi.org\/10.1109\/ICCV48922.2021.01481.","DOI":"10.1109\/ICCV48922.2021.01481"},{"key":"9716_CR29","unstructured":"Serna, I., Morales, A., Alonso-Fernandez, F. & Fierrez, J. Insidebias: measuring bias in deep networks using FairFaceVar. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 11071\u201311081 (IEEE, 2022)."},{"key":"9716_CR30","unstructured":"Yan, Z., Gong, S. & Hospedales, T. M. Mitigating demographic bias in face recognition via MultiFair representation. In Proc. European Conference on Computer Vision (ECCV) 1\u201318 (2022)."},{"key":"9716_CR31","unstructured":"Nagpal, S., Singh, M., Singh, R. & Vatsa, M. Deep learning for face recognition: pride or prejudiced? Preprint at arxiv.org\/abs\/1904.01219 (2019)."},{"key":"9716_CR32","doi-asserted-by":"crossref","unstructured":"Khan, Z. & Fu, Y. One label, one billion faces: usage and consistency of racial categories in computer vision. In Proc. ACM Conference on Fairness, Accountability, and Transparency (FAccT) 587\u2013597 (ACM, 2021).","DOI":"10.1145\/3442188.3445920"},{"key":"9716_CR33","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3359246","volume":"3","author":"MK Scheuerman","year":"2019","unstructured":"Scheuerman, M. K., Paul, J. M. & Brubaker, J. R. How computers see gender: an evaluation of gender classification in commercial facial analysis services. Proc. ACM Hum. Comput. Interact. 3, 1\u201333 (2019).","journal-title":"Proc. ACM Hum. Comput. Interact."},{"key":"9716_CR34","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3274357","volume":"2","author":"O Keyes","year":"2018","unstructured":"Keyes, O. The misgendering machines: trans\/HCI implications of automatic gender recognition. Proc. ACM on Hum. Comput. Interact. 2, 1\u201322 (2018).","journal-title":"Proc. ACM on Hum. Comput. Interact."},{"key":"9716_CR35","doi-asserted-by":"crossref","unstructured":"Hamidi, F., Scheuerman, M. K. & Branham, S. M. Gender recognition or gender reductionism? The social implications of embedded gender recognition systems. In Proc. Conference on Human Factors in Computing Systems (CHI) 1\u201313 (ACM, 2018).","DOI":"10.1145\/3173574.3173582"},{"key":"9716_CR36","doi-asserted-by":"crossref","unstructured":"Wang, J., Liu, Y. & Levy, C. Fair classification with group-dependent label noise. In Proc. ACM Conference on Fairness, Accountability, and Transparency (FAccT) 526\u2013536 (ACM, 2021).","DOI":"10.1145\/3442188.3445915"},{"key":"9716_CR37","doi-asserted-by":"crossref","unstructured":"The rise and fall (and rise) of datasets. Nat. Mach. Intell. 4, 1\u20132 (2022).","DOI":"10.1038\/s42256-022-00442-2"},{"key":"9716_CR38","unstructured":"Standard Country or Area Codes for Statistical Use (United Nations Department of Economic and Social Affairs, Statistics Division, 2024)."},{"key":"9716_CR39","first-page":"33","volume":"2","author":"TB Fitzpatrick","year":"1975","unstructured":"Fitzpatrick, T. B. Soleil et peau. J. Med. Esthet. 2, 33\u201334 (1975).","journal-title":"J. Med. Esthet."},{"key":"9716_CR40","doi-asserted-by":"crossref","unstructured":"Lin, T.-Y. et al. Microsoft COCO: common objects in context. In Proc. European Conference on Computer Vision (ECCV) 740\u2013755 (Springer, 2014).","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"9716_CR41","doi-asserted-by":"crossref","unstructured":"Goyal, Y., Khot, T., Summers-Stay, D., Batra, D. & Parikh, D. Making the V in VQA matter: elevating the role of image understanding in visual question answering. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 16904-106913 (IEEE, 2017).","DOI":"10.1109\/CVPR.2017.670"},{"key":"9716_CR42","doi-asserted-by":"crossref","unstructured":"Schumann, C., Ricco, S., Prabhu, U., Ferrari, V. & Pantofaru, C. R. A step toward more inclusive people annotations for fairness. In Proc. AAAI\/ACM Conference on AI, Ethics, and Society (AIES) 916\u2013925 (AAAI\/ACM, 2021).","DOI":"10.1145\/3461702.3462594"},{"key":"9716_CR43","first-page":"12979","volume":"35","author":"W Gaviria Rojas","year":"2022","unstructured":"Gaviria Rojas, W. et al. The Dollar Street dataset: images representing the geographic and socioeconomic diversity of the world. Adv. Neur. Inform. Process. Syst. 35, 12979\u201312990 (2022).","journal-title":"Adv. Neur. Inform. Process. Syst."},{"key":"9716_CR44","doi-asserted-by":"crossref","unstructured":"Gebru, T. et al. Datasheets for datasets. Commun. ACM 64, 86\u201392 (ACM, 2021).","DOI":"10.1145\/3458723"},{"key":"9716_CR45","unstructured":"The Nuremberg Code. Trials of War Criminals Before the Nuremberg Military Tribunals Under Control Council Law no. 10 181\u2013182 (US Government, 1949)."},{"key":"9716_CR46","unstructured":"General Data Protection Regulation (European Commission, 2016); gdpr-info.eu\/."},{"key":"9716_CR47","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P. & Ommer, B. High-resolution image synthesis with latent diffusion models. In Proc. Computer Vision and Pattern Recognition (CVPR) 10684\u201310695 (IEEE, 2022).","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"9716_CR48","unstructured":"Luccioni, S., Akiki, C., Mitchell, M. & Jernite, Y. Stable bias: evaluating societal representations in diffusion models. In Proc. 37th International Conference on Neural Information Processing Systems (NIPS) 56338\u201356351 (Curran Associates, 2023)."},{"key":"9716_CR49","unstructured":"Yang, K., Yau, J.H., Fei-Fei, L., Deng, J. & Russakovsky, O. A study of face obfuscation in ImageNet. In Proc. International Conference on Machine Learning (ICML) 25313\u201325330 (PMLR, 2022)."},{"key":"9716_CR50","doi-asserted-by":"crossref","unstructured":"Orekondy, T., Fritz, M. & Schiele, B. Connecting pixels to privacy and utility: automatic redaction of private information in images. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 8466\u20138475 (IEEE, 2018).","DOI":"10.1109\/CVPR.2018.00883"},{"key":"9716_CR51","unstructured":"Global Wage Report 2020-21: Wages and Minimum Wages in the Time of COVID-19 5\u2013207 (International Labour Organization, 2020); www.ilo.org\/wcmsp5\/groups\/public\/---dgreports\/---dcomm\/---publ\/documents\/publication\/wcms_762534.pdf."},{"key":"9716_CR52","unstructured":"Thornton, S. & Tractenberg, R. E. Ethical considerations for data involving human gender and sex variables. In Ethics in Statistics: Opportunities & Challenges 260\u2013291 (Ethics International Press, 2024)."},{"key":"9716_CR53","unstructured":"National Institutes of Health\u2014Division of Program Coordination, Planning and Strategic Initiatives. Gender Pronouns & Their Use in Workplace Communications (2022); dpcpsi.nih.gov\/sgmro\/gender-pronouns-resource."},{"key":"9716_CR54","doi-asserted-by":"publisher","first-page":"829","DOI":"10.1111\/1467-9566.00277","volume":"23","author":"PJ Aspinall","year":"2001","unstructured":"Aspinall, P. J. Operationalising the collection of ethnicity data in studies of the sociology of health and illness. Sociol. Health Illness 23, 829\u2013862 (2001).","journal-title":"Sociol. Health Illness"},{"key":"9716_CR55","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Wang, J. & Sang, J. Counterfactually measuring and eliminating social bias in vision-language pre-training models. In Proc. 30th ACM International Conference on Multimedia 4996\u20135004 (ACM, 2022).","DOI":"10.1145\/3503161.3548396"},{"key":"9716_CR56","doi-asserted-by":"crossref","unstructured":"Mitchell, M. et al. Model cards for model reporting. In Proc. ACM Conference on Fairness, Accountability, and Transparency (FAccT) 220\u2013229 (ACM, 2019).","DOI":"10.1145\/3287560.3287596"},{"key":"9716_CR57","doi-asserted-by":"crossref","unstructured":"Thong, W., Joniak, P. & Xiang, A. Beyond skin tone: a multidimensional measure of apparent skin color. In Proc. IEEE\/CVF International Conference on Computer Vision 4903\u20134913 (IEEE, 2023).","DOI":"10.1109\/ICCV51070.2023.00452"},{"key":"9716_CR58","first-page":"3","volume":"8","author":"C Bonferroni","year":"1936","unstructured":"Bonferroni, C. Teoria statistica delle classi e calcolo delle probabilita. Pubbl. R. Istitut. Super. Sci. Econ. Commer. Firenze 8, 3\u201362 (1936).","journal-title":"Pubbl. R. Istitut. Super. Sci. Econ. Commer. Firenze"},{"key":"9716_CR59","doi-asserted-by":"crossref","unstructured":"Mittal, S., Thakral, K., Kartik, Majumdar, P., Vatsa, M. & Singh, R. Are face detection models biased? In Proc. 2023 IEEE 17th International Conference on Automatic Face and Gesture Recognition (FG) 1\u20137 (IEEE, 2023).","DOI":"10.1109\/FG57933.2023.10042564"},{"key":"9716_CR60","doi-asserted-by":"crossref","unstructured":"Deng, J., Guo, J., Xue, N. & Zafeiriou, S. ArcFace: additive angular margin loss for deep face recognition. In Proc. Computer Vision and Pattern Recognition (CVPR) 4690\u20134699 (IEEE, 2019).","DOI":"10.1109\/CVPR.2019.00482"},{"key":"9716_CR61","doi-asserted-by":"crossref","unstructured":"Huang, Y. et al. Curricularface: adaptive curriculum learning loss for deep face recognition. In Proc. Computer Vision and Pattern Recognition (CVPR) 5901\u20135910 (IEEE, 2020).","DOI":"10.1109\/CVPR42600.2020.00594"},{"key":"9716_CR62","doi-asserted-by":"crossref","unstructured":"Schroff, F., Kalenichenko, D. & Philbin, J. FaceNet: a unified embedding for face recognition and clustering. In Proc. Computer Vision and Pattern Recognition 815\u2013823 (IEEE, 2015).","DOI":"10.1109\/CVPR.2015.7298682"},{"key":"9716_CR63","doi-asserted-by":"crossref","unstructured":"Wang, J., Liu, Y. & Wang, X. Are gender-neutral queries really gender-neutral? Mitigating gender bias in image search. In Proc. 2021 Conference on Empirical Methods in Natural Language Processing 1995\u20132008 (ACL, 2021).","DOI":"10.18653\/v1\/2021.emnlp-main.151"},{"key":"9716_CR64","unstructured":"Bhargava, S. & Forsyth, D. Exposing and correcting the gender bias in image captioning datasets and models. Preprint at arxiv.org\/abs\/1912.00578 (2019)."},{"key":"9716_CR65","unstructured":"Agarwal, S. et al. Evaluating CLIP: towards characterization of broader capabilities and downstream implications. Preprint at arxiv.org\/abs\/2108.02818 (2021)."},{"key":"9716_CR66","unstructured":"Bordes, F. et al. Pug: photorealistic and semantically controllable synthetic data for representation learning. In Proc. 37th International Conference on Neural Information Processing Systems (NIPS) 1952, 45020\u201345054 (Curran Associates, 2023)."},{"key":"9716_CR67","doi-asserted-by":"crossref","unstructured":"Li, X. et al. Imagenet-e: benchmarking neural network robustness via attribute editing. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition 20371\u201320381 (IEEE, 2023).","DOI":"10.1109\/CVPR52729.2023.01951"},{"key":"9716_CR68","doi-asserted-by":"crossref","unstructured":"Paiss, R. et al. Teaching clip to count to ten. In Proc. IEEE\/CVF International Conference on Computer Vision 3170\u20133180 (IEEE< 2023).","DOI":"10.1109\/ICCV51070.2023.00294"},{"key":"9716_CR69","doi-asserted-by":"crossref","unstructured":"Lee, T. et al. VHELM: a holistic evaluation of vision language models. In Proc. The Thirty-eight Conference on Neural Information Processing Systems (NIPS) Datasets and Benchmarks Track. 4464, 140632\u2013140666 (Curran Associates, 2025).","DOI":"10.52202\/079017-4464"},{"key":"9716_CR70","doi-asserted-by":"crossref","unstructured":"Schumann, C., Ricco, S., Prabhu, U., Ferrari, V. & Pantofaru, C. A step toward more inclusive people annotations for fairness. In Proc. 2021 AAAI\/ACM Conference on AI, Ethics, and Society 916\u2013925 (AAAI\/ACM, 2021).","DOI":"10.1145\/3461702.3462594"},{"key":"9716_CR71","doi-asserted-by":"publisher","first-page":"32","DOI":"10.1007\/s11263-016-0981-7","volume":"123","author":"R Krishna","year":"2017","unstructured":"Krishna, R. et al. Visual genome: connecting language and vision using crowdsourced dense image annotations. Int. J. Comput. Vis. 123, 32\u201373 (2017).","journal-title":"Int. J. Comput. Vis."},{"key":"9716_CR72","unstructured":"Wu, X., Wang, Y., Wu, H.-T., Tao, Z. & Fang, Y. Evaluating fairness in large vision-language models across diverse demographic attributes and prompts. Preprint at arxiv.org\/abs\/2406.17974 (2024)."},{"key":"9716_CR73","doi-asserted-by":"crossref","unstructured":"Sathe, A., Jain, P. & Sitaram, S. A unified framework and dataset for assessing societal bias in vision-language models. In Proc. Findings of the Association for Computational Linguistics: EMNLP 2024 1208\u20131249 (ACL, 2024).","DOI":"10.18653\/v1\/2024.findings-emnlp.66"},{"key":"9716_CR74","unstructured":"Radford, A. et al. Learning transferable visual models from natural language supervision. In Proc. International Conference on Machine Learning 8748\u20138763 (PMLR, 2021)."},{"key":"9716_CR75","unstructured":"Li, J., Li, D., Savarese, S. & Hoi, S. Blip-2: bootstrapping language-image pre-training with frozen image encoders and large language models. In Proc. International Conference on Machine Learning 19730\u201319742 (PMLR, 2023)."},{"key":"9716_CR76","unstructured":"Ramaswamy, V. V. et al. Geode: a geographically diverse evaluation dataset for object recognition. In Proc. 37th International Conference on Neural Information Processing Systems (NIPS) 2888, 66127\u201366137 (Curran Associates, 2023)."},{"key":"9716_CR77","unstructured":"Hestness, J. et al. Deep learning scaling is predictable, empirically. Preprint at arxiv.org\/abs\/1712.00409 (2017)."},{"key":"9716_CR78","doi-asserted-by":"crossref","unstructured":"Kirillov, A. et al. Segment anything. In Proc. IEEE\/CVF International Conference on Computer Vision 4015\u20134026 (IEEE, 2023).","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"9716_CR79","first-page":"115","volume":"11","author":"E Singer","year":"2017","unstructured":"Singer, E. & Couper, M. P. Some methodological uses of responses to open questions and other verbatim comments in quantitative surveys. mda 11, 115\u2013134 (2017).","journal-title":"mda"},{"key":"9716_CR80","first-page":"3","volume":"15","author":"C Neuert","year":"2021","unstructured":"Neuert, C., Meitinger, K., Behr, D. & Schonlau, M. The use of open-ended questions in surveys. MDA\u00a015, 3\u20136 (2021).","journal-title":"MDA"},{"key":"9716_CR81","doi-asserted-by":"crossref","unstructured":"Yang, S., Luo, P., Loy, C.-C. & Tang, X. Wider Face: a face detection benchmark. In Proc. Conference on Computer Vision & Pattern Recogntion (CVPR) 5525\u20135533 (IEEE, 2016).","DOI":"10.1109\/CVPR.2016.596"},{"key":"9716_CR82","unstructured":"Bazarevsky, V. et al. Blazepose: on-device real-time body pose tracking. Preprint at arxiv.org\/abs\/2006.10204 (2020)."},{"key":"9716_CR83","unstructured":"Zhang, F. et al. MediaPipe Hands: on-device real-time hand tracking. Preprint at arxiv.org\/abs\/2006.10214 (2020)."},{"key":"9716_CR84","unstructured":"Bazarevsky, V., Kartynnik, Y., Vakunov, A., Raveendran, K. & Grundmann, M. BlazeFace: sub-millisecond neural face detection on mobile GPUs. Preprint at arxiv.org\/abs\/1907.05047 (2019)."},{"key":"9716_CR85","doi-asserted-by":"crossref","unstructured":"Zhao, J. et al. Understanding humans in crowded scenes: deep nested adversarial learning and a new benchmark for multi-human parsing. In Proc. 26th ACM International Conference on Multimedia 792\u2013800 (ACM, 2018).","DOI":"10.1145\/3240508.3240509"},{"key":"9716_CR86","doi-asserted-by":"crossref","unstructured":"Lee, C.-H., Liu, Z., Wu, L. & Luo, P. MaskGAN: towards diverse and interactive facial image manipulation. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 5549\u20135558 (IEEE, 2020).","DOI":"10.1109\/CVPR42600.2020.00559"},{"key":"9716_CR87","doi-asserted-by":"crossref","unstructured":"Wood, E. et al. Fake it till you make it: face analysis in the wild using synthetic data alone. In Proc. IEEE\/CVF International Conference on Computer Vision (ICCV) 3681\u20133691 (IEEE, 2021).","DOI":"10.1109\/ICCV48922.2021.00366"},{"key":"9716_CR88","doi-asserted-by":"crossref","unstructured":"Chen, H., Li, X., Wang, Z. & Hu, X. Robust logo detection in e-commerce images by data augmentation. In Proc. 29th ACM International Conference on Multimedia 4789\u20134793 (ACM, 2021).","DOI":"10.1145\/3474085.3479227"},{"key":"9716_CR89","unstructured":"Google Cloud Vision API (Google, accessed 8 June 2023); cloud.google.com\/vision\/"},{"key":"9716_CR90","doi-asserted-by":"crossref","unstructured":"Papadopoulos, D. P., Uijlings, J. R., Keller, F. & Ferrari, V. Extreme clicking for efficient object annotation. In Proc. International Conference on Computer Vision (ICCV) 4930\u20134939 (IEEE, 2017).","DOI":"10.1109\/ICCV.2017.528"},{"key":"9716_CR91","doi-asserted-by":"crossref","unstructured":"Ruggero Ronchi, M. & Perona, P. Benchmarking and error diagnosis in multi-instance pose estimation. In Proc. International Conference on Computer Vision (ICCV) 369\u2013378 (IEEE, 2017).","DOI":"10.1109\/ICCV.2017.48"},{"key":"9716_CR92","doi-asserted-by":"publisher","first-page":"141","DOI":"10.1186\/s13244-023-01487-6","volume":"14","author":"AS Boehringer","year":"2023","unstructured":"Boehringer, A. S., Sanaat, A., Arabi, H. & Zaidi, H. An active learning approach to train a deep learning algorithm for tumor segmentation from brain MR images. Insights Imag. 14, 141 (2023).","journal-title":"Insights Imag."},{"key":"9716_CR93","unstructured":"COCO Keypoints Evaluation (COCO Consortium, 2016); cocodataset.org\/#keypoints-eval."},{"key":"9716_CR94","doi-asserted-by":"publisher","first-page":"297","DOI":"10.2307\/1932409","volume":"26","author":"LR Dice","year":"1945","unstructured":"Dice, L. R. Measures of the amount of ecologic association between species. Ecology 26, 297\u2013302 (1945).","journal-title":"Ecology"},{"key":"9716_CR95","first-page":"1","volume":"5","author":"T Sorensen","year":"1948","unstructured":"Sorensen, T. A method of establishing groups of equal amplitude in plant sociology based on similarity of species content and its application to analyses of the vegetation on danish commons. Biol. Skrift. 5, 1\u201334 (1948).","journal-title":"Biol. Skrift."},{"key":"9716_CR96","doi-asserted-by":"publisher","first-page":"37","DOI":"10.1111\/j.1469-8137.1912.tb05611.x","volume":"11","author":"P Jaccard","year":"1912","unstructured":"Jaccard, P. The distribution of the flora in the Alpine Zone. 1. N. Phytol. 11, 37\u201350 (1912).","journal-title":"N. Phytol."},{"key":"9716_CR97","unstructured":"Platen, P. et al. Diffusers: state-of-the-art diffusion models. (GitHub, 2022)."},{"key":"9716_CR98","unstructured":"Song, J., Meng, C. & Ermon, S. Denoising diffusion implicit models. In Proc. International Conference on Learning Representations (ICLR) (2021); openreview.net\/forum?id=St1giarCHLP."},{"key":"9716_CR99","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S. & Aila, T. A style-based generator architecture for generative adversarial networks. In Proc. Computer Vision and Pattern Recognition (CVPR) 4217\u20134228 (2019).","DOI":"10.1109\/CVPR.2019.00453"},{"key":"9716_CR100","unstructured":"Karras, T., Aila, T., Laine, S. & Lehtinen, J. Progressive growing of GANs for improved quality, stability, and variation. In Proc. International Conference on Learning Representations (ICLR) 1\u201326 (2018); openreview.net\/forum?id=Hk99zCeAb."},{"key":"9716_CR101","doi-asserted-by":"publisher","first-page":"1016","DOI":"10.1175\/1520-0450(1979)018<1016:LFIOAT>2.0.CO;2","volume":"18","author":"CE Duchon","year":"1979","unstructured":"Duchon, C. E. Lanczos filtering in one and two dimensions. J. App. Meteorol. Climatol. 18, 1016\u20131022 (1979).","journal-title":"J. App. Meteorol. Climatol."},{"key":"9716_CR102","unstructured":"Technology Face Challenges (NIST, 2024); www.nist.gov\/programs-projects\/face-challenges."},{"key":"9716_CR103","doi-asserted-by":"crossref","unstructured":"Zhao, D., Wang, A. & Russakovsky, O. Understanding and evaluating racial biases in image captioning. In Proc. International Conference on Computer Vision (ICCV) 14830\u201314840 (IEEE, 2021).","DOI":"10.1109\/ICCV48922.2021.01456"},{"key":"9716_CR104","doi-asserted-by":"crossref","unstructured":"Jin, S. et al. Whole-body human pose estimation in the wild. In Proc. European Conference on Computer Vision (ECCV) 196\u2013214 (Springer, 2020).","DOI":"10.1007\/978-3-030-58545-7_12"},{"key":"9716_CR105","doi-asserted-by":"publisher","first-page":"396","DOI":"10.1086\/682162","volume":"121","author":"EP Monk Jr","year":"2015","unstructured":"Monk Jr, E. P. The cost of color: skin color, discrimination, and health among African-Americans. Am. J. Sociol. 121, 396\u2013444 (2015).","journal-title":"Am. J. Sociol."},{"key":"9716_CR106","doi-asserted-by":"crossref","unstructured":"Hazirbas, C. et al. Towards measuring fairness in AI: the casual conversations dataset. IEEE Trans. Biometr. Behav. Iden. Sci. 4, 324\u2013332 (2021).","DOI":"10.1109\/TBIOM.2021.3132237"},{"key":"9716_CR107","doi-asserted-by":"crossref","unstructured":"Rothe, R., Timofte, R. & Gool, L. V. Dex: deep expectation of apparent age from a single image. In Proc. IEEE International Conference on Computer Vision Workshops (ICCVW) (IEEE, 2015).","DOI":"10.1109\/ICCVW.2015.41"},{"key":"9716_CR108","doi-asserted-by":"crossref","unstructured":"Xiao, B., Wu, H. & Wei, Y. Simple baselines for human pose estimation and tracking. In Proc. European Conference on Computer Vision (ECCV) 466\u2013481 (Springer, 2018).","DOI":"10.1007\/978-3-030-01231-1_29"},{"key":"9716_CR109","doi-asserted-by":"crossref","unstructured":"Sun, K., Xiao, B., Liu, D. & Wang, J. Deep high-resolution representation learning for human pose estimation. In Proc. Computer Vision and Pattern Recognition (CVPR) 5693\u20135703 (IEEE, 2019).","DOI":"10.1109\/CVPR.2019.00584"},{"key":"9716_CR110","unstructured":"Xu, Y., Zhang, J., Zhang, Q. & Tao, D. ViTPose: simple vision transformer baselines for human pose estimation. In Proc. 36th International Conference on Neural Information Processing Systems (NIPS) 2795, 38571\u201338584 (Curran Associates, 2022)."},{"key":"9716_CR111","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P. & Girshick, R. Mask R-CNN. In Proc. International Conference on Computer Vision (ICCV) 2961\u20132969 (IEEE, 2017).","DOI":"10.1109\/ICCV.2017.322"},{"key":"9716_CR112","doi-asserted-by":"publisher","first-page":"1483","DOI":"10.1109\/TPAMI.2019.2956516","volume":"43","author":"Z Cai","year":"2019","unstructured":"Cai, Z. & Vasconcelos, N. Cascade R-CNN: high quality object detection and instance segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 43, 1483\u20131498 (2019).","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"9716_CR113","doi-asserted-by":"crossref","unstructured":"Cheng, B., Misra, I., Schwing, A. G., Kirillov, A. & Girdhar, R. Masked-attention mask transformer for universal image segmentation. In Proc. Computer Vision and Pattern Recognition (CVPR) 1290\u20131299 (IEEE, 2022).","DOI":"10.1109\/CVPR52688.2022.00135"},{"key":"9716_CR114","doi-asserted-by":"crossref","unstructured":"Carion, N. et al. End-to-end object detection with transformers. In Proc. European Conference on Computer Vision (ECCV) 213\u2013229 (Springer, 2020).","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"9716_CR115","unstructured":"Ren, S., He, K., Girshick, R. & Sun, J. Faster R-CNN: towards real-time object detection with region proposal networks. In Proc. 29th International Conference on Neural Information Processing Systems (NIPS) 1, 91\u201399 (MIT Press, 2015)."},{"key":"9716_CR116","unstructured":"Zhu, X. et al. Deformable DETR: deformable transformers for end-to-end object detection. In Proc. International Conference on Learning Learning Representations (ICLR) (2021)."},{"key":"9716_CR117","doi-asserted-by":"crossref","unstructured":"Chen, Z. et al. Disentangle your dense object detector. In Proc. 29th ACM International Conference on Multimedia 4939\u20134948 (ACM, 2021).","DOI":"10.1145\/3474085.3475351"},{"key":"9716_CR118","doi-asserted-by":"crossref","unstructured":"Xiang, J. & Zhu, G. Joint face detection and facial expression recognition with MTCNN. In Proc. 2017 4th International Conference on Information Science and Control Engineering (ICISCE) 424\u2013427 (IEEE, 2017).","DOI":"10.1109\/ICISCE.2017.95"},{"key":"9716_CR119","doi-asserted-by":"crossref","unstructured":"Cao, Q., Shen, L., Xie, W., Parkhi, O. M. & Zisserman, A. VGGFace2: a dataset for recognising faces across pose and age. In Proc. International Conference on Automatic Face and Gesture Recognition 67\u201374 (IEEE, 2018).","DOI":"10.1109\/FG.2018.00020"},{"key":"9716_CR120","doi-asserted-by":"crossref","unstructured":"Deng, J., Guo, J., Ververas, E., Kotsia, I. & Zafeiriou, S. Retinaface: single-shot multi-level face localisation in the wild. In Proc. Computer Vision and Pattern Recognition (CVPR) 5203\u20135212 (IEEE, 2020).","DOI":"10.1109\/CVPR42600.2020.00525"},{"key":"9716_CR121","unstructured":"facenet-pytorch contributors. Face recognition using PyTorch (2019); github.com\/timesler\/facenet-pytorch."},{"key":"9716_CR122","unstructured":"FaceXLib contributors. FaceXLib (2021); github.com\/xinntao\/facexlib."},{"key":"9716_CR123","doi-asserted-by":"crossref","unstructured":"Zheng, Q., Deng, J., Zhu, Z., Li, Y. & Zafeiriou, S. Decoupled multi-task learning with cyclical self-regulation for face parsing. In Proc. Computer Vision and Pattern Recognition (CVPR) 4156\u20134165 (IEEE, 2022).","DOI":"10.1109\/CVPR52688.2022.00412"},{"key":"9716_CR124","doi-asserted-by":"crossref","unstructured":"Guo, Y., Zhang, L., Hu, Y., He, X. & Gao, J. Ms-celeb-1m: a dataset and benchmark for large-scale face recognition. In Proc. European Conference on Computer Vision (ECCV) 87\u2013102 (Springer, 2016).","DOI":"10.1007\/978-3-319-46487-9_6"},{"key":"9716_CR125","doi-asserted-by":"crossref","unstructured":"Wang, Q., Zhang, P., Xiong, H. & Zhao, J. Face.evoLVe: a high-performance face recognition library. Neurocomputing 494, 443\u2013445 (2022).","DOI":"10.1016\/j.neucom.2022.04.118"},{"key":"9716_CR126","doi-asserted-by":"crossref","unstructured":"Alaluf, Y., Patashnik, O. & Cohen-Or, D. Restyle: a residual-based stylegan encoder via iterative refinement. In Proc. International Conference on Computer Vision (ICCV) 6711\u20136720 (IEEE, 2021).","DOI":"10.1109\/ICCV48922.2021.00664"},{"key":"9716_CR127","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459838","volume":"40","author":"O Tov","year":"2021","unstructured":"Tov, O., Alaluf, Y., Nitzan, Y., Patashnik, O. & Cohen-Or, D. Designing an encoder for stylegan image manipulation. ACM Trans. Graph. 40, 1\u201314 (2021).","journal-title":"ACM Trans. Graph."},{"key":"9716_CR128","doi-asserted-by":"crossref","unstructured":"Richardson, E. et al. Encoding in style: a StyleGAN encoder for image-to-image translation. In Proc. Computer Vision and Pattern Recognition (CVPR) 2287\u20132296 (IEEE, 2021).","DOI":"10.1109\/CVPR46437.2021.00232"},{"key":"9716_CR129","doi-asserted-by":"crossref","unstructured":"Wang, X., Li, Y., Zhang, H. & Shan, Y. Towards real-world blind face restoration with generative facial prior. In Proc. Computer Vision and Pattern Recognition (CVPR) 8494\u20138508 (IEEE, 2021).","DOI":"10.1109\/CVPR46437.2021.00905"},{"key":"9716_CR130","doi-asserted-by":"crossref","unstructured":"Yang, T., Ren, P., Xie, X. & Zhang, L. GAN prior embedded network for blind face restoration in the wild. In Proc. Computer Vision and Pattern Recognition (CVPR) 672\u2013681 (IEEE, 2021).","DOI":"10.1109\/CVPR46437.2021.00073"},{"key":"9716_CR131","doi-asserted-by":"crossref","unstructured":"Andriluka, M., Pishchulin, L., Gehler, P. & Schiele, B. 2D human pose estimation: new benchmark and state of the art analysis. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 3686\u20133693 (IEEE, 2014).","DOI":"10.1109\/CVPR.2014.471"},{"key":"9716_CR132","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E. & Wang, O. The unreasonable effectiveness of deep features as a perceptual metric. In Proc. Computer Vision and Pattern Recognition (CVPR) 586\u2013595 (IEEE, 2018).","DOI":"10.1109\/CVPR.2018.00068"},{"key":"9716_CR133","unstructured":"Simonyan, K. & Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proc. 3rd International Conference on Learning Representations (ICLR) (2015)."},{"key":"9716_CR134","unstructured":"Friedman, D. & Dieng, A. B. The Vendi score: a diversity evaluation metric for machine learning. Trans. Mach. Learn. Res. 6 (2023)."},{"key":"9716_CR135","unstructured":"Pasarkar, A. & Dieng, A. B. Cousins of the Vendi score: a family of similarity-based diversity metrics for science and machine learning. In Proc. 27th International Conference on Artificial Intelligence & Statistics (AISTATS) 238 (2024)."},{"key":"9716_CR136","unstructured":"Goyal, P. et al. Vision models are more robust and fair when pretrained on uncurated images without supervision. Preprint at arxiv.org\/abs\/2202.08360 (2022)."},{"key":"9716_CR137","unstructured":"Gareth, J., Daniela, W., Trevor, H. & Robert, T. An Introduction to Statistical Learning: With Applications in R (Springer, 2013)."},{"key":"9716_CR138","doi-asserted-by":"publisher","first-page":"53","DOI":"10.1023\/B:DAMI.0000005258.31418.83","volume":"8","author":"J Han","year":"2004","unstructured":"Han, J., Pei, J., Yin, Y. & Mao, R. Mining frequent patterns without candidate generation: a frequent-pattern tree approach. Data Mining Knowl. Discov. 8, 53\u201387 (2004).","journal-title":"Data Mining Knowl. Discov."},{"key":"9716_CR139","doi-asserted-by":"publisher","unstructured":"Klare, B. F. et al. Pushing the frontiers of unconstrained face detection and recognition: IARPA Janus Benchmark A. In Proc. 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 1931\u20131939 (IEEE, 2015); https:\/\/doi.org\/10.1109\/CVPR.2015.7298803","DOI":"10.1109\/CVPR.2015.7298803"},{"key":"9716_CR140","doi-asserted-by":"crossref","unstructured":"Ponce-L\u00f3pez, V. et al. Chalearn lap 2016: first round challenge on first impressions-dataset and results. In Proc. European Conference on Computer Vision 400\u2013418 (Springer, 2016).","DOI":"10.1007\/978-3-319-49409-8_32"},{"key":"9716_CR141","doi-asserted-by":"publisher","first-page":"106632","DOI":"10.1016\/j.evolhumbehav.2024.106632","volume":"45","author":"AA Zhao","year":"2024","unstructured":"Zhao, A. A. & Zietsch, B. P. Deep neural networks generate facial metrics that overcome limitations of previous methods and predict in-person attraction. Evol. Hum. Behav. 45, 106632 (2024).","journal-title":"Evol. Hum. Behav."},{"key":"9716_CR142","doi-asserted-by":"publisher","first-page":"2115228119","DOI":"10.1073\/pnas.2115228119","volume":"119","author":"JC Peterson","year":"2022","unstructured":"Peterson, J. C., Uddenberg, S., Griffiths, T. L., Todorov, A. & Suchow, J. W. Deep models of superficial face judgments. Proc. Natl Acad. Sci. USA 119, 2115228119 (2022).","journal-title":"Proc. Natl Acad. Sci. USA"},{"key":"9716_CR143","doi-asserted-by":"publisher","first-page":"66","DOI":"10.1016\/j.patrec.2023.10.016","volume":"175","author":"K Biswas","year":"2023","unstructured":"Biswas, K., Shivakumara, P., Pal, U., Liu, C.-L. & Lu, Y. Vqapt: A new visual question answering model for personality traits in social media images. Pattern Recogn. Lett. 175, 66\u201373 (2023).","journal-title":"Pattern Recogn. Lett."},{"key":"9716_CR144","doi-asserted-by":"crossref","unstructured":"Andrews, M., Smart, A. & Birhane, A. The reanimation of pseudoscience in machine learning and its ethical repercussions. Patterns 5, 1\u201314 (2024).","DOI":"10.1016\/j.patter.2024.101027"},{"key":"9716_CR145","doi-asserted-by":"crossref","unstructured":"Ruggeri, G. et al. A multi-dimensional study on bias in vision-language models. In Proc. Findings of the Association for Computational Linguistics 6445\u20136455 (ACL, 2023).","DOI":"10.18653\/v1\/2023.findings-acl.403"},{"key":"9716_CR146","doi-asserted-by":"crossref","unstructured":"Fraser, K. C. & Kiritchenko, S. Examining gender and racial bias in large vision-language models using a novel dataset of parallel images. In Proc. 18th Conference of the European Chapter of the Association for Computational Linguistics 1, 690\u2013713 (ACL, 2024).","DOI":"10.18653\/v1\/2024.eacl-long.41"},{"key":"9716_CR147","doi-asserted-by":"crossref","unstructured":"Huang, J.-t. et al. VisBias: measuring explicit and implicit social biases in vision language models. Preprint at arxiv.org\/abs\/2503.07575 (2025).","DOI":"10.18653\/v1\/2025.emnlp-main.908"},{"key":"9716_CR148","unstructured":"Xiang, A. et al. Code for \u2018Fair human-centric image dataset for ethical AI benchmarking\u2019. Github https:\/\/github.com\/SonyResearch\/fairness-benchmark-public (2025)."},{"key":"9716_CR149","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1145\/2812802","volume":"59","author":"B Thomee","year":"2016","unstructured":"Thomee, B. et al. YFCC100M: the new data in multimedia research. Commun. ACM 59, 64\u201373 (2016).","journal-title":"Commun. ACM"},{"key":"9716_CR150","doi-asserted-by":"crossref","unstructured":"Kemelmacher-Shlizerman, I., Seitz, S. M., Miller, D. & Brossard, E. The MegaFace benchmark: 1 million faces for recognition at scale. In Proc. IEEE Conference on Computer Vision and Pattern Recognition 4873\u20134882 (IEEE, 2016).","DOI":"10.1109\/CVPR.2016.527"},{"key":"9716_CR151","doi-asserted-by":"crossref","unstructured":"Cao, Q., Shen, L., Xie, W., Parkhi, O. M. & Zisserman, A. VGGFace2: a dataset for recognising faces across pose and age. In Proc. 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018) 67\u201374 (IEEE, 2018).","DOI":"10.1109\/FG.2018.00020"},{"key":"9716_CR152","unstructured":"Merler, M., Ratha, N., Feris, R. S. & Smith, J. R. Diversity in Faces. Preprint at https:\/\/arxiv.org\/abs\/1901.10436 (2019)."},{"key":"9716_CR153","doi-asserted-by":"publisher","unstructured":"Phillips, P. J. et al. Overview of the face recognition grand challenge. In Proc. 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR\u201905) 947\u2013954 (IEEE, 2005); https:\/\/doi.org\/10.1109\/CVPR.2005.268.","DOI":"10.1109\/CVPR.2005.268"},{"key":"9716_CR154","doi-asserted-by":"crossref","unstructured":"Wang, M., Deng, W., Hu, J., Tao, X. & Huang, Y. Racial faces in the wild: reducing racial bias by information maximization adaptation network. In Proc. IEEE\/CVF International Conference on Computer Vision (ICCV) 692\u2013702 (IEEE, 2019).","DOI":"10.1109\/ICCV.2019.00078"},{"key":"9716_CR155","doi-asserted-by":"crossref","unstructured":"Ricanek, K. & Tesafaye, T. Morph: a longitudinal image database of normal adult age-progression. In Proc. 7th International Conference on Automatic Face and Gesture Recognition (FGR06) 341\u2013345 (IEEE, 2006).","DOI":"10.1109\/FGR.2006.78"},{"key":"9716_CR156","doi-asserted-by":"publisher","first-page":"2170","DOI":"10.1109\/TIFS.2014.2359646","volume":"9","author":"E Eidinger","year":"2014","unstructured":"Eidinger, E., Enbar, R. & Hassner, T. Age and gender estimation of unfiltered faces. IEEE Trans. Inform. Foren. Secur. 9, 2170\u20132179 (2014).","journal-title":"IEEE Trans. Inform. Foren. Secur."},{"key":"9716_CR157","doi-asserted-by":"crossref","unstructured":"Wang, M. & Deng, W. Mitigating bias in face recognition using skewness-aware reinforcement learning. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 9322\u20139331 (IEEE, 2020).","DOI":"10.1109\/CVPR42600.2020.00934"},{"key":"9716_CR158","doi-asserted-by":"crossref","unstructured":"Yang, Y. et al. Enhancing fairness in face detection in computer vision systems by demographic bias mitigation. In Proc. AAAI\/ACM Conference on AI, Ethics, and Society (AIES) 813\u2013822 (AAAI\/ACM, 2022).","DOI":"10.1145\/3514094.3534153"},{"key":"9716_CR159","doi-asserted-by":"publisher","first-page":"103954","DOI":"10.1016\/j.imavis.2020.103954","volume":"102","author":"M Georgopoulos","year":"2020","unstructured":"Georgopoulos, M., Panagakis, Y. & Pantic, M. Investigating bias in deep face analysis: the KANFace dataset and empirical study. Image Vis. Comput. 102, 103954 (2020).","journal-title":"Image Vis. Comput."},{"key":"9716_CR160","doi-asserted-by":"crossref","unstructured":"Karkkainen, K. & Joo, J. FairFace: face attribute dataset for balanced race, gender, and age for bias measurement and mitigation. In Proc. IEEE Winter Conference on Applications of Computer Vision (WACV) 1548\u20131558 (IEEE, 2021).","DOI":"10.1109\/WACV48630.2021.00159"},{"key":"9716_CR161","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky, O. et al. ImageNet large scale visual recognition challenge. Int. J. Comput. Vis. 115, 211\u2013252 (2015).","journal-title":"Int. J. Comput. Vis."},{"key":"9716_CR162","doi-asserted-by":"crossref","unstructured":"Liu, Z., Luo, P., Wang, X. & Tang, X. Deep learning face attributes in the wild. In Proc. IEEE International Conference on Computer Vision (ICCV) 3730\u20133738 (IEEE, 2015).","DOI":"10.1109\/ICCV.2015.425"},{"key":"9716_CR163","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Luo, P., Loy, C.C. & Tang, X. Facial landmark detection by deep multi-task learning. In Proc. European Conference on Computer Vision (ECCV) 94\u2013108 (Springer, 2014).","DOI":"10.1007\/978-3-319-10599-4_7"},{"key":"9716_CR164","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Song, Y. & Qi, H. Age progression\/regression by conditional adversarial autoencoder. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 5810\u20135818 (IEEE, 2017).","DOI":"10.1109\/CVPR.2017.463"}],"container-title":["Nature"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.nature.com\/articles\/s41586-025-09716-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s41586-025-09716-2","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s41586-025-09716-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,4]],"date-time":"2025-12-04T06:35:34Z","timestamp":1764830134000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.nature.com\/articles\/s41586-025-09716-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,5]]},"references-count":164,"journal-issue":{"issue":"8092","published-print":{"date-parts":[[2025,12,4]]}},"alternative-id":["9716"],"URL":"https:\/\/doi.org\/10.1038\/s41586-025-09716-2","relation":{},"ISSN":["0028-0836","1476-4687"],"issn-type":[{"value":"0028-0836","type":"print"},{"value":"1476-4687","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,11,5]]},"assertion":[{"value":"13 February 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 October 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 November 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"Sony Group Corporation, with inventors J.T.A.A. and A.X., has a pending US patent application US20240078839A1, filed on 14 August 2023, that is currently under examination. It covers aspects of the human-centric image dataset specification and annotation techniques that were used in this paper. The same application has also been filed in Europe (application number 23761605.7, filed on 15 January 2025) and China (application number 202380024486.X, filed on 30 August 2024) and the applications are pending.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}]}}