{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T20:55:22Z","timestamp":1776113722711,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":69,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,5,16]],"date-time":"2022-05-16T00:00:00Z","timestamp":1652659200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,5,16]]},"DOI":"10.1145\/3522664.3528600","type":"proceedings-article","created":{"date-parts":[[2022,10,17]],"date-time":"2022-10-17T16:30:14Z","timestamp":1666024214000},"page":"90-100","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":16,"title":["Method cards for prescriptive machine-learning transparency"],"prefix":"10.1145","author":[{"given":"David","family":"Adkins","sequence":"first","affiliation":[{"name":"Meta AI"}]},{"given":"Bilal","family":"Alsallakh","sequence":"additional","affiliation":[{"name":"Meta AI"}]},{"given":"Adeel","family":"Cheema","sequence":"additional","affiliation":[{"name":"Meta AI"}]},{"given":"Narine","family":"Kokhlikyan","sequence":"additional","affiliation":[{"name":"Meta AI"}]},{"given":"Emily","family":"McReynolds","sequence":"additional","affiliation":[{"name":"Meta AI"}]},{"given":"Pushkar","family":"Mishra","sequence":"additional","affiliation":[{"name":"Meta AI, UK"}]},{"given":"Chavez","family":"Procope","sequence":"additional","affiliation":[{"name":"Meta AI"}]},{"given":"Jeremy","family":"Sawruk","sequence":"additional","affiliation":[{"name":"Meta AI"}]},{"given":"Erin","family":"Wang","sequence":"additional","affiliation":[{"name":"Meta AI"}]},{"given":"Polina","family":"Zvyagina","sequence":"additional","affiliation":[{"name":"Meta AI"}]}],"member":"320","published-online":{"date-parts":[[2022,10,17]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Gradio: Hassle-free sharing and testing of ml models in the wild. arXiv preprint arXiv:1906.02569","author":"Abid Abubakar","year":"2019","unstructured":"Abubakar Abid, Ali Abdalla, Ali Abid, Dawood Khan, Abdulrahman Alfozan, and James Zou. 2019. Gradio: Hassle-free sharing and testing of ml models in the wild. arXiv preprint arXiv:1906.02569 (2019)."},{"key":"e_1_3_2_1_2_1","unstructured":"A. Adcock V. Reis M. Singh Z. Yan L. van der Maaten K. Zhang S. Motwani J. Guerin N. Goyal I. Misra L. Gustafson C. Changhan and P. Goyal. 2019. Classy Vision. https:\/\/github.com\/facebookresearch\/ClassyVision."},{"key":"e_1_3_2_1_3_1","volume-title":"Method Cards for Prescriptive Machine-Learning Transparency. In Extended Abstracts of the CHI 2022 Conference on Human Factors in Computing Systems.","author":"Adkins David","year":"2022","unstructured":"David Adkins, Bilal Alsallakh, Adeel Cheema, Narine Kokhlikyan, Emily McReynolds, Pushkar Mishra, Chavez Procope, Jeremy Sawruk, Erin Wang, and Polina Zvyagina. 2022. Method Cards for Prescriptive Machine-Learning Transparency. In Extended Abstracts of the CHI 2022 Conference on Human Factors in Computing Systems."},{"key":"e_1_3_2_1_4_1","unstructured":"Bilal Alsallakh Narine Kokhlikyan Vivek Miglani Shubham Muttepawar Edward Wang Sara Zhang David Adkins and Orion Reblitz-Richardson. 2021. Debugging the Internals of Convolutional Networks. In explainable AI approaches for debugging and diagnosis - NeurIPS Workshop."},{"key":"e_1_3_2_1_5_1","volume-title":"Mind the Pad - CNNs Can Develop Blind Spots. In Intl. Conference on Learning Representations (ICLR).","author":"Alsallakh Bilal","year":"2021","unstructured":"Bilal Alsallakh, Narine Kokhlikyan, Vivek Miglani, Jun Yuan, and Orion Reblitz-Richardson. 2021. Mind the Pad - CNNs Can Develop Blind Spots. In Intl. Conference on Learning Representations (ICLR)."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2016.2619719"},{"key":"e_1_3_2_1_7_1","unstructured":"Long Ang. 2018. A bunch of tips and tricks for training deep neural networks. (2018). https:\/\/towardsdatascience.com\/a-bunch-of-tips-and-tricks-for-training-deep-neural-networks-3ca24c31ddc8"},{"key":"e_1_3_2_1_8_1","unstructured":"Apple. 2019. Human interface guidelines for machine learning. (2019). https:\/\/developer.apple.com\/design\/human-interface-guidelines\/machine-learning\/"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1147\/JRD.2019.2942288"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2017.7989163"},{"key":"e_1_3_2_1_11_1","volume-title":"ICLR Workshop on Science and Engineering of Deep Learning arXiv:2104","author":"Bell Samuel J","year":"2021","unstructured":"Samuel J Bell and Onno P Kampman. 2021. Perspectives on Machine Learning from Psychology's Reproducibility Crisis. ICLR Workshop on Science and Engineering of Deep Learning arXiv:2104.08878 (2021)."},{"key":"e_1_3_2_1_12_1","volume-title":"NeurIPS Workshop on Dataset Curation and Security.","author":"Chmielinski Kasia S","year":"2020","unstructured":"Kasia S Chmielinski, Sarah Newman, Matt Taylor, Josh Joseph, Kemi Thomas, Jessica Yurkofsky, and Yue Chelsea Qiu. 2020. The dataset nutrition label (2nd Gen): Leveraging context to mitigate harms in artificial intelligence. In NeurIPS Workshop on Dataset Curation and Security."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICSE.2003.1201264"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/WICSA.2015.18"},{"key":"e_1_3_2_1_15_1","volume-title":"A review of research in mechanical engineering design. Part I: Descriptive, prescriptive, and computer-based models of design processes. Research in engineering design 1, 1","author":"Finger Susan","year":"1989","unstructured":"Susan Finger and John R Dixon. 1989. A review of research in mechanical engineering design. Part I: Descriptive, prescriptive, and computer-based models of design processes. Research in engineering design 1, 1 (1989), 51--67."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3458723"},{"key":"e_1_3_2_1_17_1","volume-title":"International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=Bygh9j09KX","author":"Geirhos Robert","year":"2019","unstructured":"Robert Geirhos, Patricia Rubisch, Claudio Michaelis, Matthias Bethge, Felix A. Wichmann, and Wieland Brendel. 2019. ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=Bygh9j09KX"},{"key":"e_1_3_2_1_18_1","volume-title":"Ahmed Hosny, Farnoosh Khodakarami, Levi Waldron, Bo Wang, Chris McIntosh, Anna Goldenberg, Anshul Kundaje, Casey S Greene, et al.","author":"Haibe-Kains Benjamin","year":"2020","unstructured":"Benjamin Haibe-Kains, George Alexandru Adam, Ahmed Hosny, Farnoosh Khodakarami, Levi Waldron, Bo Wang, Chris McIntosh, Anna Goldenberg, Anshul Kundaje, Casey S Greene, et al. 2020. Transparency and reproducibility in artificial intelligence. Nature 586, 7829 (2020), E14--E16."},{"key":"e_1_3_2_1_19_1","volume-title":"Assurance cases and prescriptive software safety certification: A comparative study. Safety science 59","author":"Hawkins Richard","year":"2013","unstructured":"Richard Hawkins, Ibrahim Habli, Tim Kelly, and John McDermid. 2013. Assurance cases and prescriptive software safety certification: A comparative study. Safety science 59 (2013), 55--71."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00065"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/2976767.2976808"},{"key":"e_1_3_2_1_22_1","first-page":"19000","article-title":"The origins and prevalence of texture bias in convolutional neural networks","volume":"33","author":"Hermann Katherine","year":"2020","unstructured":"Katherine Hermann, Ting Chen, and Simon Kornblith. 2020. The origins and prevalence of texture bias in convolutional neural networks. Advances in Neural Information Processing Systems 33 (2020), 19000--19015.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/3334480.3383051"},{"key":"e_1_3_2_1_24_1","volume-title":"The dataset nutrition label: A framework to drive higher data quality standards. arXiv preprint arXiv:1805.03677","author":"Holland Sarah","year":"2018","unstructured":"Sarah Holland, Ahmed Hosny, Sarah Newman, Joshua Joseph, and Kasia Chmielinski. 2018. The dataset nutrition label: A framework to drive higher data quality standards. arXiv preprint arXiv:1805.03677 (2018)."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300830"},{"key":"e_1_3_2_1_26_1","unstructured":"Stephen C Johnson. 1977. Lint a C program checker. Bell Telephone Laboratories Murray Hill."},{"key":"e_1_3_2_1_27_1","volume-title":"International conference on machine learning. PMLR, 2668--2677","author":"Kim Been","year":"2018","unstructured":"Been Kim, Martin Wattenberg, Justin Gilmer, Carrie Cai, James Wexler, Fernanda Viegas, et al. 2018. Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (TCAV). In International conference on machine learning. PMLR, 2668--2677."},{"key":"e_1_3_2_1_28_1","volume-title":"International conference on machine learning. PMLR","author":"Koh Pang Wei","year":"2017","unstructured":"Pang Wei Koh and Percy Liang. 2017. Understanding black-box predictions via influence functions. In International conference on machine learning. PMLR, 1885--1894."},{"key":"e_1_3_2_1_29_1","volume-title":"Captum: A unified and generic model interpretability library for pytorch. arXiv preprint arXiv:2009.07896","author":"Kokhlikyan Narine","year":"2020","unstructured":"Narine Kokhlikyan, Vivek Miglani, Miguel Martin, Edward Wang, Bilal Alsallakh, Jonathan Reynolds, Alexander Melnikov, Natalia Kliushkina, Carlos Araya, Siqi Yan, et al. 2020. Captum: A unified and generic model interpretability library for pytorch. arXiv preprint arXiv:2009.07896 (2020)."},{"key":"e_1_3_2_1_30_1","volume-title":"Machine learning design patterns","author":"Lakshmanan Valliappa","unstructured":"Valliappa Lakshmanan, Sara Robinson, and Michael Munn. 2020. Machine learning design patterns. O'Reilly Media."},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/3236386.3241340"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10270-003-0020-3"},{"key":"e_1_3_2_1_34_1","volume-title":"Proceedings of the Computational Methods in Systems and Software. Springer, 256--274","author":"Meng Wong Hoo","year":"2019","unstructured":"Wong Hoo Meng and Sagaya Sabestinal Amalathas. 2019. A new approach towards developing a prescriptive analytical logic model for software application error analysis. In Proceedings of the Computational Methods in Systems and Software. Springer, 256--274."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445880"},{"key":"e_1_3_2_1_36_1","unstructured":"Microsoft. 2021. The HAX Toolkit. (2021). https:\/\/www.microsoft.com\/en-us\/haxtoolkit\/"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287596"},{"key":"e_1_3_2_1_38_1","volume-title":"Advances in Neural Information Processing Systems","volume":"29","author":"Nguyen Anh","year":"2016","unstructured":"Anh Nguyen, Alexey Dosovitskiy, Jason Yosinski, Thomas Brox, and Jeff Clune. 2016. Synthesizing the preferred inputs for neurons in neural networks via deep generator networks. In Advances in Neural Information Processing Systems, Vol. 29. Curran Associates, Inc. https:\/\/proceedings.neurips.cc\/paper\/2016\/file\/5d79099fcdf499f12b79770834c0164a-Paper.pdf"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1177\/1745691612459058"},{"key":"e_1_3_2_1_40_1","volume-title":"Sinan Kalkan, and Emre Akbas.","author":"Oksuz Kemal","year":"2020","unstructured":"Kemal Oksuz, Baris Can Cam, Sinan Kalkan, and Emre Akbas. 2020. Imbalance problems in object detection: A review. IEEE transactions on pattern analysis and machine intelligence 43, 10 (2020), 3388--3415."},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.23915\/distill.00007"},{"key":"e_1_3_2_1_42_1","volume-title":"https:\/\/pair.withgoogle.com\/guidebook","author":"Guidebook Google PAIR.","year":"2019","unstructured":"Google PAIR. 2019. People + AI Guidebook. (2019). https:\/\/pair.withgoogle.com\/guidebook"},{"key":"e_1_3_2_1_43_1","volume-title":"On Buggy Resizing Libraries and Surprising Subtleties in FID Calculation. arXiv preprint arXiv:2104.11222","author":"Parmar Gaurav","year":"2021","unstructured":"Gaurav Parmar, Richard Zhang, and Jun-Yan Zhu. 2021. On Buggy Resizing Libraries and Surprising Subtleties in FID Calculation. arXiv preprint arXiv:2104.11222 (2021)."},{"key":"e_1_3_2_1_44_1","unstructured":"Adam Paszke Sam Gross Francisco Massa Adam Lerer James Bradbury Gregory Chanan Trevor Killeen Zeming Lin Natalia Gimelshein et al. 2019. PyTorch: An Imperative Style High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems (NeurIPS). 8024--8035."},{"key":"e_1_3_2_1_45_1","volume-title":"Improving reproducibility in machine learning research: a report from the NeurIPS 2019 reproducibility program. Journal of Machine Learning Research 22","author":"Pineau Joelle","year":"2021","unstructured":"Joelle Pineau, Philippe Vincent-Lamarre, Koustuv Sinha, Vincent Larivi\u00e8re, Alina Beygelzimer, Florence d'Alch\u00e9 Buc, Emily Fox, and Hugo Larochelle. 2021. Improving reproducibility in machine learning research: a report from the NeurIPS 2019 reproducibility program. Journal of Machine Learning Research 22 (2021)."},{"key":"e_1_3_2_1_46_1","unstructured":"William H Press Brian P Flannery Saul A Teukolsky William T Vetterling et al. 1989. Numerical recipes."},{"key":"e_1_3_2_1_47_1","unstructured":"Alec Radford Karthik Narasimhan Tim Salimans and Ilya Sutskever. 2018. Improving language understanding by generative pre-training. (2018)."},{"key":"e_1_3_2_1_48_1","first-page":"5485","article-title":"A step toward quantifying independently reproducible machine learning research","volume":"32","author":"Raff Edward","year":"2019","unstructured":"Edward Raff. 2019. A step toward quantifying independently reproducible machine learning research. Advances in Neural Information Processing Systems 32 (2019), 5485--5495.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_49_1","volume-title":"Chand","author":"Raghavan Sridhar A.","year":"1989","unstructured":"Sridhar A. Raghavan and Donald R. Chand. 1989. Diffusing software-engineering methods. IEEE software 6, 4 (1989), 81--90."},{"key":"e_1_3_2_1_50_1","volume-title":"ABOUT ML: Annotation and benchmarking on understanding and transparency of machine learning lifecycles. arXiv preprint arXiv:1912.06166","author":"Raji Inioluwa Deborah","year":"2019","unstructured":"Inioluwa Deborah Raji and Jingying Yang. 2019. ABOUT ML: Annotation and benchmarking on understanding and transparency of machine learning lifecycles. arXiv preprint arXiv:1912.06166 (2019)."},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1145\/2939672.2939778"},{"key":"e_1_3_2_1_52_1","volume-title":"Hidden technical debt in machine learning systems. Advances in neural information processing systems 28","author":"Sculley David","year":"2015","unstructured":"David Sculley, Gary Holt, Daniel Golovin, Eugene Davydov, Todd Phillips, Dietmar Ebner, Vinay Chaudhary, Michael Young, Jean-Francois Crespo, and Dan Dennison. 2015. Hidden technical debt in machine learning systems. Advances in neural information processing systems 28 (2015), 2503--2511."},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.74"},{"key":"e_1_3_2_1_54_1","unstructured":"Y. Shibui. 2020. Machine learning system design patterns. (2020). https:\/\/github.com\/mercari\/ml-system-design-pattern"},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1177\/1745691616658637"},{"key":"e_1_3_2_1_56_1","volume-title":"Gradients of counterfactuals. arXiv preprint arXiv:1611.02639","author":"Sundararajan Mukund","year":"2016","unstructured":"Mukund Sundararajan, Ankur Taly, and Qiqi Yan. 2016. Gradients of counterfactuals. arXiv preprint arXiv:1611.02639 (2016)."},{"key":"e_1_3_2_1_57_1","volume-title":"REVISE: A Tool for Measuring and Mitigating Bias in Visual Datasets. In European Conference on Computer Vision (ECCV).","author":"Wang Angelina","year":"2020","unstructured":"Angelina Wang, Arvind Narayanan, and Olga Russakovsky. 2020. REVISE: A Tool for Measuring and Mitigating Bias in Visual Datasets. In European Conference on Computer Vision (ECCV)."},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1109\/MC.2021.3137227"},{"key":"e_1_3_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1109\/IWESEP49350.2019.00017"},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.23915\/distill.00002"},{"key":"e_1_3_2_1_61_1","volume-title":"Explainable AI: Interpreting, Explaining and Visualizing Deep Learning","author":"Weller Adrian","unstructured":"Adrian Weller. 2019. Transparency: motivations and challenges. In Explainable AI: Interpreting, Explaining and Visualizing Deep Learning. Springer, 23--40."},{"key":"e_1_3_2_1_62_1","volume-title":"The what-if tool: Interactive probing of machine learning models","author":"Wexler James","year":"2019","unstructured":"James Wexler, Mahima Pushkarna, Tolga Bolukbasi, Martin Wattenberg, Fernanda Viegas, and Jimbo Wilson. 2019. The what-if tool: Interactive probing of machine learning models. IEEE transactions on visualization and computer graphics 26, 1 (2019), 56--65."},{"key":"e_1_3_2_1_63_1","volume-title":"NeurIPS 2021 Workshop on ImageNet: Past, Present, and Future.","author":"Wightman Ross","year":"2021","unstructured":"Ross Wightman, Hugo Touvron, and Herve Jegou. 2021. Resnet strikes back: An improved training procedure in timm. arXiv preprint arXiv:2110.00476 (2021). NeurIPS 2021 Workshop on ImageNet: Past, Present, and Future."},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.24018\/ejece.2019.3.5.133"},{"key":"e_1_3_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICSA-C.2019.00055"},{"key":"e_1_3_2_1_66_1","volume-title":"In ICML Workshop on Deep Learning.","author":"Yosinski Jason","year":"2015","unstructured":"Jason Yosinski, Jeff Clune, Thomas Fuchs, and Hod Lipson. 2015. Understanding neural networks through deep visualization. In In ICML Workshop on Deep Learning."},{"key":"e_1_3_2_1_67_1","unstructured":"Jun Yuan Bilal Alsallakh Narine Kokhlikyan Vivek Miglani and Orion Reblitz-Richardson. 2021. Convolution Can Incur Foveation Effects. In Beyond static papers: Rethinking how we share scientific understanding in ML-ICLR 2021 workshop."},{"key":"e_1_3_2_1_68_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-10590-1_53"},{"key":"e_1_3_2_1_69_1","volume-title":"Rethinking pre-training and self-training. arXiv preprint arXiv:2006.06882","author":"Zoph Barret","year":"2020","unstructured":"Barret Zoph, Golnaz Ghiasi, Tsung-Yi Lin, Yin Cui, Hanxiao Liu, Ekin D Cubuk, and Quoc V Le. 2020. Rethinking pre-training and self-training. arXiv preprint arXiv:2006.06882 (2020)."}],"event":{"name":"CAIN '22: 1st Conference on AI Engineering - Software Engineering for AI","location":"Pittsburgh Pennsylvania","acronym":"CAIN '22","sponsor":["SIGSOFT ACM Special Interest Group on Software Engineering","IEEE TCSC IEEE Technical Committee on Scalable Computing"]},"container-title":["Proceedings of the 1st International Conference on AI Engineering: Software Engineering for AI"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3522664.3528600","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3522664.3528600","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T18:09:34Z","timestamp":1750183774000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3522664.3528600"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,5,16]]},"references-count":69,"alternative-id":["10.1145\/3522664.3528600","10.1145\/3522664"],"URL":"https:\/\/doi.org\/10.1145\/3522664.3528600","relation":{},"subject":[],"published":{"date-parts":[[2022,5,16]]},"assertion":[{"value":"2022-10-17","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}