{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T21:54:57Z","timestamp":1776117297949,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":64,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,6,20]],"date-time":"2022-06-20T00:00:00Z","timestamp":1655683200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"National Science Foundation","award":["1452854"],"award-info":[{"award-number":["1452854"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,6,21]]},"DOI":"10.1145\/3531146.3534626","type":"proceedings-article","created":{"date-parts":[[2022,6,20]],"date-time":"2022-06-20T14:27:10Z","timestamp":1655735230000},"page":"2069-2082","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":24,"title":["Designing Up with Value-Sensitive Design: Building a Field Guide for Ethical ML Development"],"prefix":"10.1145","author":[{"given":"Karen","family":"Boyd","sequence":"first","affiliation":[{"name":"San Diego Workforce Partnership, USA"}]}],"member":"320","published-online":{"date-parts":[[2022,6,20]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10115-017-1116-3"},{"key":"e_1_3_2_1_2_1","unstructured":"Julia Angwin and Jeff Larson. 2016. Machine Bias. https:\/\/www.propublica.org\/article\/machine-bias-risk-assessments-in-criminal-sentencing"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372859"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3375691"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"crossref","unstructured":"Solon Barocas and Andrew\u00a0D. Selbst. 2016. Big data\u2019s disparate impact. (2016). https:\/\/papers.ssrn.com\/sol3\/papers.cfm?abstract_id=2477899","DOI":"10.2139\/ssrn.2477899"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3375684"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/1978942.1979275"},{"key":"e_1_3_2_1_8_1","volume-title":"Learning Fair Classifiers: A Regularization-Inspired Approach. abs\/1707.00044","author":"Bechavod Yahav","year":"2017","unstructured":"Yahav Bechavod and Katrina Ligett. 2017. Learning Fair Classifiers: A Regularization-Inspired Approach. abs\/1707.00044 (2017)."},{"key":"e_1_3_2_1_9_1","unstructured":"Alex Beutel Jilin Chen Zhe Zhao and Ed\u00a0H. Chi. 2017. Data Decisions and Theoretical Implications when Adversarially Learning Fair Representations. (2017). arxiv:1707.00075http:\/\/arxiv.org\/abs\/1707.00075"},{"key":"e_1_3_2_1_10_1","volume-title":"Advances in Neural Information Processing Systems 29, D.\u00a0D. Lee, M.\u00a0Sugiyama, U.\u00a0V. Luxburg, I.\u00a0Guyon, and R.\u00a0Garnett (Eds.). Curran Associates","author":"Bolukbasi Tolga","unstructured":"Tolga Bolukbasi, Kai-Wei Chang, James\u00a0Y Zou, Venkatesh Saligrama, and Adam\u00a0T Kalai. 2016. Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings. In Advances in Neural Information Processing Systems 29, D.\u00a0D. Lee, M.\u00a0Sugiyama, U.\u00a0V. Luxburg, I.\u00a0Guyon, and R.\u00a0Garnett (Eds.). Curran Associates, Inc., 4349\u20134357. http:\/\/papers.nips.cc\/paper\/6228-man-is-to-computer-programmer-as-woman-is-to-homemaker-debiasing-word-embeddings.pdf"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"crossref","unstructured":"Karen Boyd and Katie Shilton. 2022. Adapting Ethical Sensitivity as a Construct to Study Technology Design Teams. (2022).","DOI":"10.1145\/3463929"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/3479582"},{"key":"e_1_3_2_1_13_1","volume-title":"Conference on fairness, accountability and transparency. 77\u201391","author":"Buolamwini Joy","year":"2018","unstructured":"Joy Buolamwini and Timnit Gebru. 2018. Gender shades: Intersectional accuracy disparities in commercial gender classification. In Conference on fairness, accountability and transparency. 77\u201391."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","unstructured":"Aylin Caliskan Joanna\u00a0J. Bryson and Arvind Narayanan. 2017. Semantics derived automatically from language corpora contain human-like biases. 356 6334(2017) 183\u2013186. https:\/\/doi.org\/10.1126\/science.aal4230 arxiv:1608.07187","DOI":"10.1126\/science.aal4230"},{"key":"e_1_3_2_1_15_1","first-page":"I","article-title":"Optimized Pre-Processing for Discrimination Prevention","volume":"30","author":"Calmon Flavio","year":"2017","unstructured":"Flavio Calmon, Dennis Wei, Bhanukiran Vinzamuri, Karthikeyan Natesan\u00a0Ramamurthy, and Kush\u00a0R Varshney. 2017. Optimized Pre-Processing for Discrimination Prevention. In Advances in Neural Information Processing Systems 30, I.\u00a0Guyon, U.\u00a0V. Luxburg, S.\u00a0Bengio, H.\u00a0Wallach, R.\u00a0Fergus, S.\u00a0Vishwanathan, and R.\u00a0Garnett (Eds.). Curran Associates, Inc., 3992\u20134001. http:\/\/papers.nips.cc\/paper\/6988-optimized-pre-processing-for-discrimination-prevention.pdf","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1080\/09638237.2020.1714011"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287601"},{"key":"e_1_3_2_1_18_1","unstructured":"L.\u00a0Elisa Celis Damian Straszak and Nisheeth\u00a0K. Vishnoi. 2018. Ranking with Fairness Constraints. (2018). arxiv:1704.06840http:\/\/arxiv.org\/abs\/1704.06840"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287587"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDE.2019.00139"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290607.3299057"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1038\/d41586-019-02514-7"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287572"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/3278721.3278729"},{"key":"e_1_3_2_1_25_1","unstructured":"Finale Doshi-Velez and Been Kim. 2017. Towards A Rigorous Science of Interpretable Machine Learning. (2017). arxiv:1702.08608http:\/\/arxiv.org\/abs\/1702.08608"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1089\/big.2016.0054"},{"key":"e_1_3_2_1_27_1","unstructured":"Sorelle\u00a0A. Friedler Carlos Scheidegger and Suresh Venkatasubramanian. 2016. On the (im)possibility of fairness. (2016). arxiv:1609.07236http:\/\/arxiv.org\/abs\/1609.07236"},{"key":"e_1_3_2_1_28_1","unstructured":"Sorelle\u00a0A. Friedler Carlos Scheidegger Suresh Venkatasubramanian Sonam Choudhary Evan\u00a0P. Hamilton and Derek Roth. 2018. A comparative study of fairness-enhancing interventions in machine learning. (2018). arxiv:1802.04422http:\/\/arxiv.org\/abs\/1802.04422"},{"key":"e_1_3_2_1_29_1","volume-title":"Value-sensitive design. 3, 6","author":"Friedman Batya","year":"1996","unstructured":"Batya Friedman. 1996. Value-sensitive design. 3, 6 (1996), 16\u201323. http:\/\/dl.acm.org\/citation.cfm?id=242493"},{"key":"e_1_3_2_1_30_1","unstructured":"Timnit Gebru Jamie Morgenstern Briana Vecchione Jennifer\u00a0Wortman Vaughan Hanna Wallach Hal Daume\u00e9\u00a0III and Kate Crawford. 2018. Datasheets for Datasets. (2018). arxiv:1803.09010http:\/\/arxiv.org\/abs\/1803.09010"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287563"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","unstructured":"Daniel Greene Anna\u00a0Lauren Hoffmann and Luke Stark. 2019. Better Nicer Clearer Fairer: A Critical Assessment of the Movement for Ethical Artificial Intelligence and Machine Learning. https:\/\/doi.org\/10.24251\/HICSS.2019.258 Accepted: 2019-01-03T00:00:45Z.","DOI":"10.24251\/HICSS.2019.258"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1525\/pol.1997.20.1.114"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.artint.2015.08.008"},{"key":"e_1_3_2_1_35_1","unstructured":"Moritz Hardt Eric Price and Nati Srebro. 2016. Equality of Opportunity in Supervised Learning. https:\/\/openreview.net\/forum?id=S1btBvZOZr"},{"key":"e_1_3_2_1_36_1","unstructured":"Tatsunori\u00a0B. Hashimoto Megha Srivastava Hongseok Namkoong and Percy Liang. 2018. Fairness Without Demographics in Repeated Loss Minimization. (2018). arxiv:1806.08010http:\/\/arxiv.org\/abs\/1806.08010"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","unstructured":"Kenneth Holstein Jennifer\u00a0Wortman Vaughan Hal Daum\u00e9\u00a0III Miro Dud\u00edk and Hanna Wallach. 2018. Improving fairness in machine learning systems: What do industry practitioners need?(2018). https:\/\/doi.org\/10.1145\/3290605.3300830 arxiv:1812.05239","DOI":"10.1145\/3290605.3300830"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287600"},{"key":"e_1_3_2_1_39_1","volume-title":"Consortium for the Science of Sociotechnical Systems Research Summer Institute","author":"Irani Lily","year":"2019","unstructured":"[39] Lily Irani.2019. (2019). Consortium for the Science of Sociotechnical Systems Research Summer Institute 2019."},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1145\/2470654.2470742"},{"key":"e_1_3_2_1_41_1","volume-title":"Blind Justice: Fairness with Encrypted Sensitive Attributes.","author":"Kilbertus Niki","year":"2018","unstructured":"Niki Kilbertus, Adri\u00e0 Gasc\u00f3n, Matt\u00a0J. Kusner, Michael Veale, Krishna\u00a0P. Gummadi, and Adrian Weller. 2018. Blind Justice: Fairness with Encrypted Sensitive Attributes. (2018). arxiv:1806.03281http:\/\/arxiv.org\/abs\/1806.03281"},{"key":"e_1_3_2_1_42_1","first-page":"I","article-title":"Avoiding Discrimination through Causal Reasoning","volume":"30","author":"Kilbertus Niki","year":"2017","unstructured":"Niki Kilbertus, Mateo Rojas\u00a0Carulla, Giambattista Parascandolo, Moritz Hardt, Dominik Janzing, and Bernhard Sch\u00f6lkopf. 2017. Avoiding Discrimination through Causal Reasoning. In Advances in Neural Information Processing Systems 30, I.\u00a0Guyon, U.\u00a0V. Luxburg, S.\u00a0Bengio, H.\u00a0Wallach, R.\u00a0Fergus, S.\u00a0Vishwanathan, and R.\u00a0Garnett (Eds.). Curran Associates, Inc., 656\u2013666. http:\/\/papers.nips.cc\/paper\/6668-avoiding-discrimination-through-causal-reasoning.pdf","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1145\/3512899"},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376445"},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","unstructured":"Rishabh Mehrotra Ashton Anderson Fernando Diaz Amit Sharma Hanna Wallach and Emine Yilmaz. 2017. Auditing Search Engines for Differential Satisfaction Across Demographics. (2017) 626\u2013633. https:\/\/doi.org\/10.1145\/3041021.3054197 arxiv:1705.10689","DOI":"10.1145\/3041021.3054197"},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","unstructured":"Margaret Mitchell Simone Wu Andrew Zaldivar Parker Barnes Lucy Vasserman Ben Hutchinson Elena Spitzer Inioluwa\u00a0Deborah Raji and Timnit Gebru. 2019. Model Cards for Model Reporting. (2019) 220\u2013229. https:\/\/doi.org\/10.1145\/3287560.3287596 arxiv:1810.03993","DOI":"10.1145\/3287560.3287596"},{"key":"e_1_3_2_1_47_1","unstructured":"Laura Nader. 1972. Up the Anthropologist: Perspectives Gained From Studying Up. (1972). https:\/\/eric.ed.gov\/?id=ED065375"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287593"},{"key":"e_1_3_2_1_49_1","unstructured":"Cathy O\u2019Neil. 2016. Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy. Crown."},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1145\/3449081"},{"key":"e_1_3_2_1_51_1","volume-title":"Postconventional moral thinking: A neo-Kohlbergian approach","author":"Rest James","unstructured":"James Rest, Darcia Narvaez, Muriel\u00a0J. Bebeau, and Stephen\u00a0J. Thoma. 1999. Postconventional moral thinking: A neo-Kohlbergian approach. Lawrence Erlbaum Associates Publishers. Pages: ix, 229."},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.2307\/3560621"},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1145\/3341164"},{"key":"e_1_3_2_1_54_1","volume-title":"Advances in Neural Information Processing Systems 28, C.\u00a0Cortes, N.\u00a0D.","author":"Sculley D.","unstructured":"D. Sculley, Gary Holt, Daniel Golovin, Eugene Davydov, Todd Phillips, Dietmar Ebner, Vinay Chaudhary, Michael Young, Jean-Fran\u00e7ois Crespo, and Dan Dennison. 2015. Hidden Technical Debt in Machine Learning Systems. In Advances in Neural Information Processing Systems 28, C.\u00a0Cortes, N.\u00a0D. Lawrence, D.\u00a0D. Lee, M.\u00a0Sugiyama, and R.\u00a0Garnett (Eds.). Curran Associates, Inc., 2503\u20132511. http:\/\/papers.nips.cc\/paper\/5656-hidden-technical-debt-in-machine-learning-systems.pdf"},{"key":"e_1_3_2_1_55_1","volume-title":"Studying Up: The Ethnography of Technologists","author":"Seaver Nick","year":"2014","unstructured":"Nick Seaver. 2014. Studying Up: The Ethnography of Technologists. http:\/\/ethnographymatters.net\/blog\/2014\/03\/10\/studying-up\/"},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","unstructured":"Andrew\u00a0D. Selbst. 2017. Disparate Impact in Big Data Policing. (2017). https:\/\/doi.org\/10.2139\/ssrn.2819182","DOI":"10.2139\/ssrn.2819182"},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1177\/0162243912436985"},{"key":"e_1_3_2_1_58_1","unstructured":"Latanya Sweeney. 2013. Discrimination in Online Ad Delivery. (2013). arxiv:1301.6822http:\/\/arxiv.org\/abs\/1301.6822"},{"key":"e_1_3_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11023-007-9055-5"},{"key":"e_1_3_2_1_60_1","unstructured":"Indr\u00e8 \u017dliobait\u00e8. 2015. A survey on measuring indirect discrimination in machine learning. (2015). arxiv:1511.00148http:\/\/arxiv.org\/abs\/1511.00148"},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10506-016-9182-5"},{"key":"e_1_3_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1111\/j.1365-2648.2008.04625.x"},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372833"},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.1145\/3278721.3278779"}],"event":{"name":"FAccT '22: 2022 ACM Conference on Fairness, Accountability, and Transparency","location":"Seoul Republic of Korea","acronym":"FAccT '22","sponsor":["ACM Association for Computing Machinery"]},"container-title":["2022 ACM Conference on Fairness Accountability and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3531146.3534626","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3531146.3534626","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:31:30Z","timestamp":1750188690000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3531146.3534626"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,6,20]]},"references-count":64,"alternative-id":["10.1145\/3531146.3534626","10.1145\/3531146"],"URL":"https:\/\/doi.org\/10.1145\/3531146.3534626","relation":{},"subject":[],"published":{"date-parts":[[2022,6,20]]},"assertion":[{"value":"2022-06-20","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}