{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,12]],"date-time":"2025-12-12T13:07:05Z","timestamp":1765544825008,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":23,"publisher":"ACM","license":[{"start":{"date-parts":[[2021,8,14]],"date-time":"2021-08-14T00:00:00Z","timestamp":1628899200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"National Science Foundation","award":["CNS-1943016"],"award-info":[{"award-number":["CNS-1943016"]}]},{"name":"DARPA and Air Force Research Laboratory","award":["DARPA\/AFRL FA8750-15- 2-0277"],"award-info":[{"award-number":["DARPA\/AFRL FA8750-15- 2-0277"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2021,8,14]]},"DOI":"10.1145\/3447548.3470806","type":"proceedings-article","created":{"date-parts":[[2021,8,12]],"date-time":"2021-08-12T06:12:03Z","timestamp":1628748723000},"page":"4035-4036","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":14,"title":["Machine Learning Explainability and Robustness"],"prefix":"10.1145","author":[{"given":"Anupam","family":"Datta","sequence":"first","affiliation":[{"name":"Carnegie Mellon University, Mountain View, CA, USA"}]},{"given":"Matt","family":"Fredrikson","sequence":"additional","affiliation":[{"name":"Carnegie Mellon University, Pittsburgh, PA, USA"}]},{"given":"Klas","family":"Leino","sequence":"additional","affiliation":[{"name":"Carnegie Mellon University, Pittsburgh, PA, USA"}]},{"given":"Kaiji","family":"Lu","sequence":"additional","affiliation":[{"name":"Carnegie Mellon University, Mountain View, CA, USA"}]},{"given":"Shayak","family":"Sen","sequence":"additional","affiliation":[{"name":"Truera, Redwood City, CA, USA"}]},{"given":"Zifan","family":"Wang","sequence":"additional","affiliation":[{"name":"Carnegie Mellon University, Pittsburgh, PA, USA"}]}],"member":"320","published-online":{"date-parts":[[2021,8,14]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"Julius Adebayo Justin Gilmer Michael Muelly Ian Goodfellow Moritz Hardt and Been Kim. 2018. Sanity Checks for Saliency Maps. In Advances in Neural Information Processing Systems .  Julius Adebayo Justin Gilmer Michael Muelly Ian Goodfellow Moritz Hardt and Been Kim. 2018. Sanity Checks for Saliency Maps. In Advances in Neural Information Processing Systems ."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"crossref","unstructured":"Marco Ancona Enea Ceolini Cengiz \u00d6ztireli and Markus Gross. 2018. Towards better understanding of gradient-based attribution methods for Deep Neural Networks. In ICLR .  Marco Ancona Enea Ceolini Cengiz \u00d6ztireli and Markus Gross. 2018. Towards better understanding of gradient-based attribution methods for Deep Neural Networks. In ICLR .","DOI":"10.1007\/978-3-030-28954-6_9"},{"key":"e_1_3_2_1_3_1","unstructured":"Jeremy Cohen Elan Rosenfeld and Zico Kolter. 2019. Certified Adversarial Robustness via Randomized Smoothing. In ICML .  Jeremy Cohen Elan Rosenfeld and Zico Kolter. 2019. Certified Adversarial Robustness via Randomized Smoothing. In ICML ."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1109\/SP.2016.42"},{"key":"e_1_3_2_1_5_1","unstructured":"Ann-Kathrin Dombrowski Maximillian Alber Christopher Anders Marcel Ackermann Klaus-Robert M\u00fcller and Pan Kessel. 2019. Explanations can be manipulated and geometry is to blame. In Advances in Neural Information Processing Systems .  Ann-Kathrin Dombrowski Maximillian Alber Christopher Anders Marcel Ackermann Klaus-Robert M\u00fcller and Pan Kessel. 2019. Explanations can be manipulated and geometry is to blame. In Advances in Neural Information Processing Systems ."},{"key":"e_1_3_2_1_6_1","volume-title":"Zou","author":"Ghorbani A.","year":"2019","unstructured":"A. Ghorbani , Abubakar Abid , and James Y . Zou . 2019 . Interpretation of Neural Networks is Fragile. In AAAI . A. Ghorbani, Abubakar Abid, and James Y. Zou. 2019. Interpretation of Neural Networks is Fragile. In AAAI ."},{"key":"e_1_3_2_1_7_1","unstructured":"Andrew Ilyas Shibani Santurkar Dimitris Tsipras Logan Engstrom Brandon Tran and Aleksander Madry. 2019. Adversarial Examples Are Not Bugs They Are Features. In Advances in Neural Information Processing Systems .  Andrew Ilyas Shibani Santurkar Dimitris Tsipras Logan Engstrom Brandon Tran and Aleksander Madry. 2019. Adversarial Examples Are Not Bugs They Are Features. In Advances in Neural Information Processing Systems ."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/TEST.2018.8624792"},{"key":"e_1_3_2_1_9_1","unstructured":"Klas Leino Zifan Wang and Matt Fredrikson. 2021. Globally-Robust Neural Networks. In ICML .  Klas Leino Zifan Wang and Matt Fredrikson. 2021. Globally-Robust Neural Networks. In ICML ."},{"key":"e_1_3_2_1_10_1","unstructured":"Scott M Lundberg and Su-In Lee. 2017. A Unified Approach to Interpreting Model Predictions. In Advances in Neural Information Processing Systems .  Scott M Lundberg and Su-In Lee. 2017. A Unified Approach to Interpreting Model Predictions. In Advances in Neural Information Processing Systems ."},{"key":"e_1_3_2_1_11_1","unstructured":"Aleksander Madry Aleksandar Makelov Ludwig Schmidt Dimitris Tsipras and Adrian Vladu. 2018. Towards Deep Learning Models Resistant to Adversarial Attacks. In ICLR .  Aleksander Madry Aleksandar Makelov Ludwig Schmidt Dimitris Tsipras and Adrian Vladu. 2018. Towards Deep Learning Models Resistant to Adversarial Attacks. In ICLR ."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"crossref","unstructured":"Marco Tulio Ribeiro Sameer Singh and Carlos Guestrin. 2016. \"Why Should I Trust You?\": Explaining the Predictions of Any Classifier. In KDD .  Marco Tulio Ribeiro Sameer Singh and Carlos Guestrin. 2016. \"Why Should I Trust You?\": Explaining the Predictions of Any Classifier. In KDD .","DOI":"10.18653\/v1\/N16-3020"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"crossref","unstructured":"R. R. Selvaraju Abhishek Das Ramakrishna Vedantam Michael Cogswell D. Parikh and Dhruv Batra. 2019. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. In International Journal of Computer Vision .  R. R. Selvaraju Abhishek Das Ramakrishna Vedantam Michael Cogswell D. Parikh and Dhruv Batra. 2019. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. In International Journal of Computer Vision .","DOI":"10.1007\/s11263-019-01228-7"},{"key":"e_1_3_2_1_14_1","unstructured":"Mukund Sundararajan Ankur Taly and Qiqi Yan. 2017. Axiomatic attribution for deep networks. In ICML .  Mukund Sundararajan Ankur Taly and Qiqi Yan. 2017. Axiomatic attribution for deep networks. In ICML ."},{"key":"e_1_3_2_1_15_1","unstructured":"Christian Szegedy Wojciech Zaremba Ilya Sutskever Joan Bruna Dumitru Erhan Ian Goodfellow and Rob Fergus. 2014. Intriguing properties of neural networks. In ICLR .  Christian Szegedy Wojciech Zaremba Ilya Sutskever Joan Bruna Dumitru Erhan Ian Goodfellow and Rob Fergus. 2014. Intriguing properties of neural networks. In ICLR ."},{"key":"e_1_3_2_1_16_1","unstructured":"Dimitris Tsipras Shibani Santurkar Logan Engstrom Alexander Turner and Aleksander Madry. 2019. Robustness May Be at Odds with Accuracy. In ICLR .  Dimitris Tsipras Shibani Santurkar Logan Engstrom Alexander Turner and Aleksander Madry. 2019. Robustness May Be at Odds with Accuracy. In ICLR ."},{"key":"e_1_3_2_1_17_1","unstructured":"Zifan Wang Matt Fredrikson and Anupam Datta. 2021. Boundary Attributions Provide Normal (Vector) Explanations. arxiv: 2103.11257 [cs.LG]  Zifan Wang Matt Fredrikson and Anupam Datta. 2021. Boundary Attributions Provide Normal (Vector) Explanations. arxiv: 2103.11257 [cs.LG]"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW50498.2020.00013"},{"key":"e_1_3_2_1_19_1","unstructured":"Zifan Wang Haofan Wang Shakul Ramkumar Matt Fredrikson Piotr Mardziel and Anupam Datta. 2020 b. Smoothed Geometry for Robust Attribution. In Advances in Neural Information Processing Systems .  Zifan Wang Haofan Wang Shakul Ramkumar Matt Fredrikson Piotr Mardziel and Anupam Datta. 2020 b. Smoothed Geometry for Robust Attribution. In Advances in Neural Information Processing Systems ."},{"key":"e_1_3_2_1_20_1","volume":"201","author":"Wong Eric","unstructured":"Eric Wong , Frank Schmidt , Jan Hendrik Metzen , and J. Zico Kolter. 201 8. Scaling provable adversarial defenses. In Advances in Neural Information Processing Systems . Eric Wong, Frank Schmidt, Jan Hendrik Metzen, and J. Zico Kolter. 2018. Scaling provable adversarial defenses. In Advances in Neural Information Processing Systems .","journal-title":"J. Zico Kolter."},{"key":"e_1_3_2_1_21_1","unstructured":"Chih-Kuan Yeh Cheng-Yu Hsieh Arun Suggala David I Inouye and Pradeep K Ravikumar. 2019. On the (In)fidelity and Sensitivity of Explanations . In NIPS .  Chih-Kuan Yeh Cheng-Yu Hsieh Arun Suggala David I Inouye and Pradeep K Ravikumar. 2019. On the (In)fidelity and Sensitivity of Explanations . In NIPS ."},{"key":"e_1_3_2_1_22_1","unstructured":"Huan Zhang Hongge Chen Chaowei Xiao Sven Gowal Robert Stanforth Bo Li Duane Boning and Cho-Jui Hsieh. 2020. Towards Stable and Efficient Training of Verifiably Robust Neural Networks. In ICLR .  Huan Zhang Hongge Chen Chaowei Xiao Sven Gowal Robert Stanforth Bo Li Duane Boning and Cho-Jui Hsieh. 2020. Towards Stable and Efficient Training of Verifiably Robust Neural Networks. In ICLR ."},{"key":"e_1_3_2_1_23_1","volume-title":"Laurent El Ghaoui, and Michael Jordan","author":"Zhang Hongyang","year":"2019","unstructured":"Hongyang Zhang , Yaodong Yu , Jiantao Jiao , Eric Xing , Laurent El Ghaoui, and Michael Jordan . 2019 . Theoretically Principled Trade-off between Robustness and Accuracy. In ICML . Hongyang Zhang, Yaodong Yu, Jiantao Jiao, Eric Xing, Laurent El Ghaoui, and Michael Jordan. 2019. Theoretically Principled Trade-off between Robustness and Accuracy. In ICML ."}],"event":{"name":"KDD '21: The 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining","sponsor":["SIGMOD ACM Special Interest Group on Management of Data","SIGKDD ACM Special Interest Group on Knowledge Discovery in Data"],"location":"Virtual Event Singapore","acronym":"KDD '21"},"container-title":["Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery &amp; Data Mining"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3447548.3470806","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3447548.3470806","content-type":"application\/pdf","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3447548.3470806","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:18:32Z","timestamp":1750191512000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3447548.3470806"}},"subtitle":["Connected at the Hip"],"short-title":[],"issued":{"date-parts":[[2021,8,14]]},"references-count":23,"alternative-id":["10.1145\/3447548.3470806","10.1145\/3447548"],"URL":"https:\/\/doi.org\/10.1145\/3447548.3470806","relation":{},"subject":[],"published":{"date-parts":[[2021,8,14]]},"assertion":[{"value":"2021-08-14","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}