{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,7,4]],"date-time":"2025-07-04T05:38:39Z","timestamp":1751607519355,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":50,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,4,27]],"date-time":"2022-04-27T00:00:00Z","timestamp":1651017600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022,4,27]]},"DOI":"10.1145\/3491101.3519724","type":"proceedings-article","created":{"date-parts":[[2022,4,29]],"date-time":"2022-04-29T16:49:48Z","timestamp":1651250988000},"page":"1-9","source":"Crossref","is-referenced-by-count":9,"title":["Prescriptive and Descriptive Approaches to Machine-Learning Transparency"],"prefix":"10.1145","author":[{"given":"David","family":"Adkins","sequence":"first","affiliation":[{"name":"Responsible AI, Meta AI, United States"}]},{"given":"Bilal","family":"Alsallakh","sequence":"additional","affiliation":[{"name":"Responsible AI, Meta AI, United States"}]},{"given":"Adeel","family":"Cheema","sequence":"additional","affiliation":[{"name":"Responsible AI, Meta AI, United States"}]},{"given":"Narine","family":"Kokhlikyan","sequence":"additional","affiliation":[{"name":"Responsible AI, Meta AI, United States"}]},{"given":"Emily","family":"McReynolds","sequence":"additional","affiliation":[{"name":"Meta AI, United States"}]},{"given":"Pushkar","family":"Mishra","sequence":"additional","affiliation":[{"name":"Responsible AI, Meta AI, United Kingdom"}]},{"given":"Chavez","family":"Procope","sequence":"additional","affiliation":[{"name":"Responsible AI, Meta AI, United States"}]},{"given":"Jeremy","family":"Sawruk","sequence":"additional","affiliation":[{"name":"Responsible AI, Meta AI, United States"}]},{"given":"Erin","family":"Wang","sequence":"additional","affiliation":[{"name":"Responsible AI, Meta AI, United States"}]},{"given":"Polina","family":"Zvyagina","sequence":"additional","affiliation":[{"name":"Meta AI, United States"}]}],"member":"320","published-online":{"date-parts":[[2022,4,28]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"A. Adcock V. Reis M. Singh Z. Yan L. van der Maaten K. Zhang S. Motwani J. Guerin N. Goyal I. Misra L. Gustafson C. Changhan and P. Goyal. 2019. Classy Vision. https:\/\/github.com\/facebookresearch\/ClassyVision.  A. Adcock V. Reis M. Singh Z. Yan L. van der Maaten K. Zhang S. Motwani J. Guerin N. Goyal I. Misra L. Gustafson C. Changhan and P. Goyal. 2019. Classy Vision. https:\/\/github.com\/facebookresearch\/ClassyVision."},{"key":"e_1_3_2_1_2_1","unstructured":"B. Alsallakh N. Kokhlikyan V. Miglani S. Muttepawar E. Wang S. Zhang D. Adkins and O. Reblitz-Richardson. 2021. Debugging the Internals of Convolutional Networks. In eXplainable AI approaches for debugging and diagnosis - NeurIPS Workshop.  B. Alsallakh N. Kokhlikyan V. Miglani S. Muttepawar E. Wang S. Zhang D. Adkins and O. Reblitz-Richardson. 2021. Debugging the Internals of Convolutional Networks. In eXplainable AI approaches for debugging and diagnosis - NeurIPS Workshop."},{"volume-title":"Intl. Conference on Learning Representations (ICLR).","author":"Alsallakh B.","key":"e_1_3_2_1_3_1","unstructured":"B. Alsallakh , N. Kokhlikyan , V. Miglani , J. Yuan , and O. Reblitz-Richardson . 2021. Mind the Pad \u2013 CNNs Can Develop Blind Spots . In Intl. Conference on Learning Representations (ICLR). B. Alsallakh, N. Kokhlikyan, V. Miglani, J. Yuan, and O. Reblitz-Richardson. 2021. Mind the Pad \u2013 CNNs Can Develop Blind Spots. In Intl. Conference on Learning Representations (ICLR)."},{"key":"e_1_3_2_1_4_1","unstructured":"Long Ang. 2018. A bunch of tips and tricks for training deep neural networks. (2018). https:\/\/towardsdatascience.com\/a-bunch-of-tips-and-tricks-for-training-deep-neural-networks-3ca24c31ddc8  Long Ang. 2018. A bunch of tips and tricks for training deep neural networks. (2018). https:\/\/towardsdatascience.com\/a-bunch-of-tips-and-tricks-for-training-deep-neural-networks-3ca24c31ddc8"},{"key":"e_1_3_2_1_5_1","unstructured":"Apple. 2019. Human interface guidelines for machine learning. (2019). https:\/\/developer.apple.com\/design\/human-interface-guidelines\/machine-learning\/  Apple. 2019. Human interface guidelines for machine learning. (2019). https:\/\/developer.apple.com\/design\/human-interface-guidelines\/machine-learning\/"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1147\/JRD.2019.2942288"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2017.7989163"},{"key":"e_1_3_2_1_8_1","volume-title":"Reproducibility Crisis. ICLR Workshop on Science and Engineering of Deep Learning arXiv:2104","author":"Bell J","year":"2021","unstructured":"Samuel\u00a0 J Bell and Onno\u00a0 P Kampman . 2021 . Perspectives on Machine Learning from Psychology\u2019s Reproducibility Crisis. ICLR Workshop on Science and Engineering of Deep Learning arXiv:2104 .08878 (2021). Samuel\u00a0J Bell and Onno\u00a0P Kampman. 2021. Perspectives on Machine Learning from Psychology\u2019s Reproducibility Crisis. ICLR Workshop on Science and Engineering of Deep Learning arXiv:2104.08878 (2021)."},{"key":"e_1_3_2_1_9_1","volume-title":"NeurIPS Workshop on Dataset Curation and Security.","author":"Chmielinski S","year":"2020","unstructured":"Kasia\u00a0 S Chmielinski , Sarah Newman , Matt Taylor , Josh Joseph , Kemi Thomas , Jessica Yurkofsky , and Yue\u00a0Chelsea Qiu . 2020 . The dataset nutrition label (2nd Gen): Leveraging context to mitigate harms in artificial intelligence . In NeurIPS Workshop on Dataset Curation and Security. Kasia\u00a0S Chmielinski, Sarah Newman, Matt Taylor, Josh Joseph, Kemi Thomas, Jessica Yurkofsky, and Yue\u00a0Chelsea Qiu. 2020. The dataset nutrition label (2nd Gen): Leveraging context to mitigate harms in artificial intelligence. In NeurIPS Workshop on Dataset Curation and Security."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICSE.2003.1201264"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1109\/WICSA.2015.18"},{"key":"e_1_3_2_1_12_1","volume-title":"A review of research in mechanical engineering design. Part I: Descriptive, prescriptive, and computer-based models of design processes. Research in engineering design 1, 1","author":"Finger Susan","year":"1989","unstructured":"Susan Finger and John\u00a0 R Dixon . 1989. A review of research in mechanical engineering design. Part I: Descriptive, prescriptive, and computer-based models of design processes. Research in engineering design 1, 1 ( 1989 ), 51\u201367. Susan Finger and John\u00a0R Dixon. 1989. A review of research in mechanical engineering design. Part I: Descriptive, prescriptive, and computer-based models of design processes. Research in engineering design 1, 1 (1989), 51\u201367."},{"key":"e_1_3_2_1_13_1","first-page":"12","article-title":"Datasheets for","volume":"64","author":"Gebru T.","year":"2021","unstructured":"T. Gebru , J. Morgenstern , B. Vecchione , J.\u00a0 Wortman Vaughan , H. Wallach , Hal\u00a0Daum\u00e9 III, and K. Crawford . 2021 . Datasheets for Datasets. Commun. ACM 64 , 12 (nov 2021), 86\u201392. https:\/\/doi.org\/10.1145\/3458723 T. Gebru, J. Morgenstern, B. Vecchione, J.\u00a0Wortman Vaughan, H. Wallach, Hal\u00a0Daum\u00e9 III, and K. Crawford. 2021. Datasheets for Datasets. Commun. ACM 64, 12 (nov 2021), 86\u201392. https:\/\/doi.org\/10.1145\/3458723","journal-title":"Datasets. Commun. ACM"},{"key":"e_1_3_2_1_14_1","volume-title":"International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=Bygh9j09KX","author":"Geirhos Robert","year":"2019","unstructured":"Robert Geirhos , Patricia Rubisch , Claudio Michaelis , Matthias Bethge , Felix\u00a0 A. Wichmann , and Wieland Brendel . 2019 . ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness .. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=Bygh9j09KX Robert Geirhos, Patricia Rubisch, Claudio Michaelis, Matthias Bethge, Felix\u00a0A. Wichmann, and Wieland Brendel. 2019. ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness.. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=Bygh9j09KX"},{"key":"e_1_3_2_1_15_1","volume-title":"Transparency and reproducibility in artificial intelligence. Nature 586, 7829","author":"Haibe-Kains Benjamin","year":"2020","unstructured":"Benjamin Haibe-Kains , George\u00a0Alexandru Adam , Ahmed Hosny , Farnoosh Khodakarami , Levi Waldron , Bo Wang , Chris McIntosh , Anna Goldenberg , Anshul Kundaje , Casey\u00a0 S Greene , 2020. Transparency and reproducibility in artificial intelligence. Nature 586, 7829 ( 2020 ), E14\u2013E16. Benjamin Haibe-Kains, George\u00a0Alexandru Adam, Ahmed Hosny, Farnoosh Khodakarami, Levi Waldron, Bo Wang, Chris McIntosh, Anna Goldenberg, Anshul Kundaje, Casey\u00a0S Greene, 2020. Transparency and reproducibility in artificial intelligence. Nature 586, 7829 (2020), E14\u2013E16."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"crossref","unstructured":"Richard Hawkins Ibrahim Habli Tim Kelly and John McDermid. 2013. Assurance cases and prescriptive software safety certification: A comparative study. Safety science 59(2013) 55\u201371.  Richard Hawkins Ibrahim Habli Tim Kelly and John McDermid. 2013. Assurance cases and prescriptive software safety certification: A comparative study. Safety science 59(2013) 55\u201371.","DOI":"10.1016\/j.ssci.2013.04.007"},{"volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 558\u2013567","author":"He T.","key":"e_1_3_2_1_17_1","unstructured":"T. He , Z. Zhang , H. Zhang , Z. Zhang , J. Xie , and M. Li . 2019. Bag of tricks for image classification with convolutional neural networks . In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 558\u2013567 . T. He, Z. Zhang, H. Zhang, Z. Zhang, J. Xie, and M. Li. 2019. Bag of tricks for image classification with convolutional neural networks. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 558\u2013567."},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/2976767.2976808"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3334480.3383051"},{"key":"e_1_3_2_1_20_1","unstructured":"Sarah Holland Ahmed Hosny Sarah Newman Joshua Joseph and Kasia Chmielinski. 2018. The dataset nutrition label: A framework to drive higher data quality standards. arXiv preprint arXiv:1805.03677(2018).  Sarah Holland Ahmed Hosny Sarah Newman Joshua Joseph and Kasia Chmielinski. 2018. The dataset nutrition label: A framework to drive higher data quality standards. arXiv preprint arXiv:1805.03677(2018)."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300830"},{"key":"e_1_3_2_1_22_1","unstructured":"Valliappa Lakshmanan Sara Robinson and Michael Munn. 2020. Machine learning design patterns. O\u2019Reilly Media.  Valliappa Lakshmanan Sara Robinson and Michael Munn. 2020. Machine learning design patterns. O\u2019Reilly Media."},{"key":"e_1_3_2_1_23_1","volume-title":"The Mythos of Model Interpretability: In machine learning, the concept of interpretability is both important and slippery.Queue 16, 3","author":"Lipton C","year":"2018","unstructured":"Zachary\u00a0 C Lipton . 2018. The Mythos of Model Interpretability: In machine learning, the concept of interpretability is both important and slippery.Queue 16, 3 ( 2018 ), 31\u201357. Zachary\u00a0C Lipton. 2018. The Mythos of Model Interpretability: In machine learning, the concept of interpretability is both important and slippery.Queue 16, 3 (2018), 31\u201357."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10270-003-0020-3"},{"key":"e_1_3_2_1_26_1","volume-title":"Proceedings of the Computational Methods in Systems and Software. Springer, 256\u2013274","author":"Meng Wong\u00a0Hoo","year":"2019","unstructured":"Wong\u00a0Hoo Meng and Sagaya\u00a0Sabestinal Amalathas . 2019 . A new approach towards developing a prescriptive analytical logic model for software application error analysis . In Proceedings of the Computational Methods in Systems and Software. Springer, 256\u2013274 . Wong\u00a0Hoo Meng and Sagaya\u00a0Sabestinal Amalathas. 2019. A new approach towards developing a prescriptive analytical logic model for software application error analysis. In Proceedings of the Computational Methods in Systems and Software. Springer, 256\u2013274."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445880"},{"key":"e_1_3_2_1_28_1","unstructured":"Microsoft. 2021. The HAX Toolkit. (2021). https:\/\/www.microsoft.com\/en-us\/haxtoolkit\/  Microsoft. 2021. The HAX Toolkit. (2021). https:\/\/www.microsoft.com\/en-us\/haxtoolkit\/"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287596"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1177\/1745691612459058"},{"key":"e_1_3_2_1_31_1","volume-title":"https:\/\/pair.withgoogle.com\/guidebook","author":"Guidebook Google PAIR.","year":"2019","unstructured":"Google PAIR. 2019. People + AI Guidebook . ( 2019 ). https:\/\/pair.withgoogle.com\/guidebook Google PAIR. 2019. People + AI Guidebook. (2019). https:\/\/pair.withgoogle.com\/guidebook"},{"key":"e_1_3_2_1_32_1","unstructured":"Gaurav Parmar Richard Zhang and Jun-Yan Zhu. 2021. On Buggy Resizing Libraries and Surprising Subtleties in FID Calculation. arXiv preprint arXiv:2104.11222(2021).  Gaurav Parmar Richard Zhang and Jun-Yan Zhu. 2021. On Buggy Resizing Libraries and Surprising Subtleties in FID Calculation. arXiv preprint arXiv:2104.11222(2021)."},{"volume-title":"PyTorch: An Imperative Style","author":"Paszke Adam","key":"e_1_3_2_1_33_1","unstructured":"Adam Paszke , Sam Gross , Francisco Massa , Adam Lerer , James Bradbury , Gregory Chanan , Trevor Killeen , Zeming Lin , Natalia Gimelshein , 2019. PyTorch: An Imperative Style , High-Performance Deep Learning Library . In Advances in Neural Information Processing Systems (NeurIPS). 8024\u20138035. Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, 2019. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems (NeurIPS). 8024\u20138035."},{"key":"e_1_3_2_1_34_1","volume-title":"Improving reproducibility in machine learning research: a report from the NeurIPS 2019 reproducibility program. Journal of Machine Learning Research 22","author":"Pineau Joelle","year":"2021","unstructured":"Joelle Pineau , Philippe Vincent-Lamarre , Koustuv Sinha , Vincent Larivi\u00e8re , Alina Beygelzimer , Florence d\u2019 Alch\u00e9 Buc , Emily Fox , and Hugo Larochelle . 2021. Improving reproducibility in machine learning research: a report from the NeurIPS 2019 reproducibility program. Journal of Machine Learning Research 22 ( 2021 ). Joelle Pineau, Philippe Vincent-Lamarre, Koustuv Sinha, Vincent Larivi\u00e8re, Alina Beygelzimer, Florence d\u2019Alch\u00e9 Buc, Emily Fox, and Hugo Larochelle. 2021. Improving reproducibility in machine learning research: a report from the NeurIPS 2019 reproducibility program. Journal of Machine Learning Research 22 (2021)."},{"key":"e_1_3_2_1_35_1","unstructured":"William\u00a0H Press Brian\u00a0P Flannery Saul\u00a0A Teukolsky William\u00a0T Vetterling 1989. Numerical recipes.  William\u00a0H Press Brian\u00a0P Flannery Saul\u00a0A Teukolsky William\u00a0T Vetterling 1989. Numerical recipes."},{"key":"e_1_3_2_1_36_1","unstructured":"Alec Radford Jeff Wu Rewon Child David Luan Dario Amodei and Ilya Sutskever. 2019. Language Models are Unsupervised Multitask Learners. (2019).  Alec Radford Jeff Wu Rewon Child David Luan Dario Amodei and Ilya Sutskever. 2019. Language Models are Unsupervised Multitask Learners. (2019)."},{"key":"e_1_3_2_1_37_1","first-page":"5485","article-title":"A step toward quantifying independently reproducible machine learning research","volume":"32","author":"Raff Edward","year":"2019","unstructured":"Edward Raff . 2019 . A step toward quantifying independently reproducible machine learning research . Advances in Neural Information Processing Systems 32 (2019), 5485 \u2013 5495 . Edward Raff. 2019. A step toward quantifying independently reproducible machine learning research. Advances in Neural Information Processing Systems 32 (2019), 5485\u20135495.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1109\/52.31655"},{"key":"e_1_3_2_1_39_1","volume-title":"ABOUT ML: Annotation and benchmarking on understanding and transparency of machine learning lifecycles. arXiv preprint arXiv:1912.06166(2019).","author":"Raji D.","year":"2019","unstructured":"I.\u00a0 D. Raji and J. Yang . 2019 . ABOUT ML: Annotation and benchmarking on understanding and transparency of machine learning lifecycles. arXiv preprint arXiv:1912.06166(2019). I.\u00a0D. Raji and J. Yang. 2019. ABOUT ML: Annotation and benchmarking on understanding and transparency of machine learning lifecycles. arXiv preprint arXiv:1912.06166(2019)."},{"key":"e_1_3_2_1_40_1","volume-title":"Hidden technical debt in machine learning systems. Advances in neural information processing systems 28","author":"Sculley David","year":"2015","unstructured":"David Sculley , Gary Holt , Daniel Golovin , Eugene Davydov , Todd Phillips , Dietmar Ebner , Vinay Chaudhary , Michael Young , Jean-Francois Crespo , and Dan Dennison . 2015. Hidden technical debt in machine learning systems. Advances in neural information processing systems 28 ( 2015 ), 2503\u20132511. David Sculley, Gary Holt, Daniel Golovin, Eugene Davydov, Todd Phillips, Dietmar Ebner, Vinay Chaudhary, Michael Young, Jean-Francois Crespo, and Dan Dennison. 2015. Hidden technical debt in machine learning systems. Advances in neural information processing systems 28 (2015), 2503\u20132511."},{"key":"e_1_3_2_1_41_1","unstructured":"Y. Shibui. 2020. Machine learning system design patterns. (2020). https:\/\/github.com\/mercari\/ml-system-design-pattern  Y. Shibui. 2020. Machine learning system design patterns. (2020). https:\/\/github.com\/mercari\/ml-system-design-pattern"},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1177\/1745691616658637"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1109\/MC.2021.3137227"},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1109\/IWESEP49350.2019.00017"},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.23915\/distill.00002"},{"volume-title":"Explainable AI: Interpreting, Explaining and Visualizing Deep Learning","author":"Weller Adrian","key":"e_1_3_2_1_46_1","unstructured":"Adrian Weller . 2019. Transparency: motivations and challenges . In Explainable AI: Interpreting, Explaining and Visualizing Deep Learning . Springer , 23\u201340. Adrian Weller. 2019. Transparency: motivations and challenges. In Explainable AI: Interpreting, Explaining and Visualizing Deep Learning. Springer, 23\u201340."},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.24018\/ejece.2019.3.5.133"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICSA-C.2019.00055"},{"key":"e_1_3_2_1_49_1","unstructured":"J. Yuan B. Alsallakh N. Kokhlikyan V. Miglani and O. Reblitz-Richardson. 2021. Convolution Can Incur Foveation Effects. In Beyond static papers: Rethinking how we share scientific understanding in ML-ICLR 2021 workshop.  J. Yuan B. Alsallakh N. Kokhlikyan V. Miglani and O. Reblitz-Richardson. 2021. Convolution Can Incur Foveation Effects. In Beyond static papers: Rethinking how we share scientific understanding in ML-ICLR 2021 workshop."},{"key":"e_1_3_2_1_50_1","unstructured":"Barret Zoph Golnaz Ghiasi Tsung-Yi Lin Yin Cui Hanxiao Liu Ekin\u00a0D Cubuk and Quoc\u00a0V Le. 2020. Rethinking pre-training and self-training. arXiv preprint arXiv:2006.06882(2020).  Barret Zoph Golnaz Ghiasi Tsung-Yi Lin Yin Cui Hanxiao Liu Ekin\u00a0D Cubuk and Quoc\u00a0V Le. 2020. Rethinking pre-training and self-training. arXiv preprint arXiv:2006.06882(2020)."}],"event":{"name":"CHI '22: CHI Conference on Human Factors in Computing Systems","sponsor":["SIGCHI ACM Special Interest Group on Computer-Human Interaction"],"location":"New Orleans LA USA","acronym":"CHI '22"},"container-title":["CHI Conference on Human Factors in Computing Systems Extended Abstracts"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3491101.3519724","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3491101.3519724","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:30:59Z","timestamp":1750188659000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3491101.3519724"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,27]]},"references-count":50,"alternative-id":["10.1145\/3491101.3519724","10.1145\/3491101"],"URL":"https:\/\/doi.org\/10.1145\/3491101.3519724","relation":{},"subject":[],"published":{"date-parts":[[2022,4,27]]}}}