{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,11]],"date-time":"2026-03-11T07:44:41Z","timestamp":1773215081964,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":44,"publisher":"ACM","license":[{"start":{"date-parts":[[2021,8,14]],"date-time":"2021-08-14T00:00:00Z","timestamp":1628899200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2021,8,14]]},"DOI":"10.1145\/3447548.3467213","type":"proceedings-article","created":{"date-parts":[[2021,8,12]],"date-time":"2021-08-12T06:12:08Z","timestamp":1628748728000},"page":"1027-1035","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":62,"title":["What Do You See?"],"prefix":"10.1145","author":[{"given":"Yi-Shan","family":"Lin","sequence":"first","affiliation":[{"name":"Purdue University, West Lafayette, IN, USA"}]},{"given":"Wen-Chuan","family":"Lee","sequence":"additional","affiliation":[{"name":"Apple Inc., Cupertino, CA, USA"}]},{"given":"Z. Berkay","family":"Celik","sequence":"additional","affiliation":[{"name":"Purdue University, West Lafayette, IN, USA"}]}],"member":"320","published-online":{"date-parts":[[2021,8,14]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Peeking inside the black-box: A survey on Explainable Artificial Intelligence (XAI)","author":"Adadi Amina","year":"2018","unstructured":"Amina Adadi and Mohammed Berrada . 2018. Peeking inside the black-box: A survey on Explainable Artificial Intelligence (XAI) . IEEE Access ( 2018 ). Amina Adadi and Mohammed Berrada. 2018. Peeking inside the black-box: A survey on Explainable Artificial Intelligence (XAI). IEEE Access (2018)."},{"key":"e_1_3_2_1_2_1","volume-title":"Towards better understanding of gradient-based attribution methods for deep neural networks. arXiv:1711.06104","author":"Ancona Marco","year":"2017","unstructured":"Marco Ancona , Enea Ceolini , Cengiz \u00d6ztireli , and Markus Gross . 2017. Towards better understanding of gradient-based attribution methods for deep neural networks. arXiv:1711.06104 ( 2017 ). Marco Ancona, Enea Ceolini, Cengiz \u00d6ztireli, and Markus Gross. 2017. Towards better understanding of gradient-based attribution methods for deep neural networks. arXiv:1711.06104 (2017)."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2019.12.012"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/3377325.3377498"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.1986.4767851"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/WACV.2018.00097"},{"key":"e_1_3_2_1_7_1","volume-title":"Targeted backdoor attacks on deep learning systems using data poisoning. arXiv:1712.05526","author":"Chen Xinyun","year":"2017","unstructured":"Xinyun Chen , Chang Liu , Bo Li , Kimberly Lu , and Dawn Song . 2017. Targeted backdoor attacks on deep learning systems using data poisoning. arXiv:1712.05526 ( 2017 ). Xinyun Chen, Chang Liu, Bo Li, Kimberly Lu, and Dawn Song. 2017. Targeted backdoor attacks on deep learning systems using data poisoning. arXiv:1712.05526 (2017)."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/SPW50608.2020.00025"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/3427228.3427264"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN48605.2020.9207637"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.371"},{"key":"e_1_3_2_1_12_1","volume-title":"Badnets: Identifying vulnerabilities in the machine learning model supply chain. arXiv:1708.06733","author":"Gu Tianyu","year":"2017","unstructured":"Tianyu Gu , Brendan Dolan-Gavitt , and Siddharth Garg . 2017 . Badnets: Identifying vulnerabilities in the machine learning model supply chain. arXiv:1708.06733 (2017). Tianyu Gu, Brendan Dolan-Gavitt, and Siddharth Garg. 2017. Badnets: Identifying vulnerabilities in the machine learning model supply chain. arXiv:1708.06733 (2017)."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_1_14_1","unstructured":"Sara Hooker Dumitru Erhan Pieter-Jan Kindermans and Been Kim. 2019. A benchmark for interpretability methods in deep neural networks. In Advances in Neural Information Processing Systems. 9737--9748.  Sara Hooker Dumitru Erhan Pieter-Jan Kindermans and Been Kim. 2019. A benchmark for interpretability methods in deep neural networks. In Advances in Neural Information Processing Systems. 9737--9748."},{"key":"e_1_3_2_1_15_1","volume-title":"Evaluating XAI: A comparison of rule-based and example-based explanations. Artificial Intelligence 291, 291","author":"Kaptein F","year":"2021","unstructured":"F Kaptein , J Broekens , K Hindriks , and MA Neerincx . 2021. Evaluating XAI: A comparison of rule-based and example-based explanations. Artificial Intelligence 291, 291 ( 2021 ). F Kaptein, J Broekens, K Hindriks, and MA Neerincx. 2021. Evaluating XAI: A comparison of rule-based and example-based explanations. Artificial Intelligence 291, 291 (2021)."},{"key":"e_1_3_2_1_16_1","volume-title":"International conference on machine learning. PMLR, 2668--2677","author":"Kim Been","year":"2018","unstructured":"Been Kim , MartinWattenberg, Justin Gilmer , Carrie Cai , JamesWexler, Fernanda Viegas , 2018 . Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (tcav) . In International conference on machine learning. PMLR, 2668--2677 . Been Kim, MartinWattenberg, Justin Gilmer, Carrie Cai, JamesWexler, Fernanda Viegas, et al. 2018. Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (tcav). In International conference on machine learning. PMLR, 2668--2677."},{"key":"e_1_3_2_1_17_1","first-page":"1097","article-title":"Imagenet classification with deep convolutional neural networks","volume":"25","author":"Krizhevsky Alex","year":"2012","unstructured":"Alex Krizhevsky , Ilya Sutskever , and Geoffrey E Hinton . 2012 . Imagenet classification with deep convolutional neural networks . Advances in Neural Information Processing Systems 25 (2012), 1097 -- 1105 . Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. 2012. Imagenet classification with deep convolutional neural networks. Advances in Neural Information Processing Systems 25 (2012), 1097--1105.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_18_1","volume-title":"Jiahao Yu, Minhui Xue, Dali Kaafar, and Haojin Zhu.","author":"Li Shaofeng","year":"2019","unstructured":"Shaofeng Li , Benjamin Zi Hao Zhao , Jiahao Yu, Minhui Xue, Dali Kaafar, and Haojin Zhu. 2019 . Invisible backdoor attacks against deep neural networks. arXiv:1909.02742 (2019). Shaofeng Li, Benjamin Zi Hao Zhao, Jiahao Yu, Minhui Xue, Dali Kaafar, and Haojin Zhu. 2019. Invisible backdoor attacks against deep neural networks. arXiv:1909.02742 (2019)."},{"key":"e_1_3_2_1_19_1","volume-title":"Backdoor embedding in convolutional neural network models via invisible perturbation. arXiv:1808.10307","author":"Liao Cong","year":"2018","unstructured":"Cong Liao , Haoti Zhong , Anna Squicciarini , Sencun Zhu , and David Miller . 2018. Backdoor embedding in convolutional neural network models via invisible perturbation. arXiv:1808.10307 ( 2018 ). Cong Liao, Haoti Zhong, Anna Squicciarini, Sencun Zhu, and David Miller. 2018. Backdoor embedding in convolutional neural network models via invisible perturbation. arXiv:1808.10307 (2018)."},{"key":"e_1_3_2_1_20_1","volume-title":"Trojaning Attack on Neural Networks. In Network and Distributed System Security Symposium (NDSS).","author":"Liu Yingqi","year":"2018","unstructured":"Yingqi Liu , Shiqing Ma , Yousra Aafer , Wen-Chuan Lee , Juan Zhai ,WeihangWang, and Xiangyu Zhang . 2018 . Trojaning Attack on Neural Networks. In Network and Distributed System Security Symposium (NDSS). Yingqi Liu, Shiqing Ma, Yousra Aafer,Wen-Chuan Lee, Juan Zhai,WeihangWang, and Xiangyu Zhang. 2018. Trojaning Attack on Neural Networks. In Network and Distributed System Security Symposium (NDSS)."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISQED48828.2020.9137011"},{"key":"e_1_3_2_1_22_1","unstructured":"David Alvarez Melis and Tommi Jaakkola. 2018. Towards robust interpretability with self-explaining neural networks. In Advances in Neural Information Processing Systems. 7775--7784.  David Alvarez Melis and Tommi Jaakkola. 2018. Towards robust interpretability with self-explaining neural networks. In Advances in Neural Information Processing Systems. 7775--7784."},{"key":"e_1_3_2_1_23_1","volume-title":"Constantin Waubert de Puiseau, and Tobias Meisen","author":"Meyes Richard","year":"2019","unstructured":"Richard Meyes , Melanie Lu , Constantin Waubert de Puiseau, and Tobias Meisen . 2019 . Ablation studies in artificial neural networks. arXiv:1901.08644 (2019). Richard Meyes, Melanie Lu, Constantin Waubert de Puiseau, and Tobias Meisen. 2019. Ablation studies in artificial neural networks. arXiv:1901.08644 (2019)."},{"key":"e_1_3_2_1_24_1","volume-title":"Jeremy E Block, and Eric D Ragan","author":"Mohseni Sina","year":"2018","unstructured":"Sina Mohseni , Jeremy E Block, and Eric D Ragan . 2018 . A human-grounded evaluation benchmark for local explanations of machine learning. arXiv:1801.05075 (2018). Sina Mohseni, Jeremy E Block, and Eric D Ragan. 2018. A human-grounded evaluation benchmark for local explanations of machine learning. arXiv:1801.05075 (2018)."},{"key":"e_1_3_2_1_25_1","volume-title":"A multidisciplinary survey and framework for design and evaluation of explainable AI systems. arXiv:1811.11839","author":"Mohseni Sina","year":"2018","unstructured":"Sina Mohseni , Niloofar Zarei , and Eric D Ragan . 2018. A multidisciplinary survey and framework for design and evaluation of explainable AI systems. arXiv:1811.11839 ( 2018 ). Sina Mohseni, Niloofar Zarei, and Eric D Ragan. 2018. A multidisciplinary survey and framework for design and evaluation of explainable AI systems. arXiv:1811.11839 (2018)."},{"key":"e_1_3_2_1_26_1","volume-title":"ICPR-2020 Workshop Explainable Deep Learning-AI. Springer Publishing Company.","author":"Muddamsetty Satya Mahesh","year":"2020","unstructured":"Satya Mahesh Muddamsetty , Mohammad Naser Sabet Jahromi , and Thomas B Moeslund . 2020 . Expert level evaluations for explainable AI (XAI) methods in the medical domain . In ICPR-2020 Workshop Explainable Deep Learning-AI. Springer Publishing Company. Satya Mahesh Muddamsetty, Mohammad Naser Sabet Jahromi, and Thomas B Moeslund. 2020. Expert level evaluations for explainable AI (XAI) methods in the medical domain. In ICPR-2020 Workshop Explainable Deep Learning-AI. Springer Publishing Company."},{"key":"e_1_3_2_1_27_1","volume-title":"Pytorch: An imperative style, high-performance deep learning library. arXiv:1912.01703","author":"Paszke Adam","year":"2019","unstructured":"Adam Paszke , Sam Gross , Francisco Massa , Adam Lerer , James Bradbury , Gregory Chanan , Trevor Killeen , Zeming Lin , Natalia Gimelshein , Luca Antiga , 2019 . Pytorch: An imperative style, high-performance deep learning library. arXiv:1912.01703 (2019). Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. 2019. Pytorch: An imperative style, high-performance deep learning library. arXiv:1912.01703 (2019)."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.91"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/2939672.2939778"},{"key":"e_1_3_2_1_30_1","volume-title":"Learning representations by back-propagating errors. nature 323, 6088","author":"Rumelhart David E","year":"1986","unstructured":"David E Rumelhart , Geoffrey E Hinton , and Ronald J Williams . 1986. Learning representations by back-propagating errors. nature 323, 6088 ( 1986 ), 533--536. David E Rumelhart, Geoffrey E Hinton, and Ronald J Williams. 1986. Learning representations by back-propagating errors. nature 323, 6088 (1986), 533--536."},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"crossref","unstructured":"Olga Russakovsky Jia Deng Hao Su Jonathan Krause Sanjeev Satheesh Sean Ma Zhiheng Huang Andrej Karpathy Aditya Khosla Michael Bernstein etal 2015. Imagenet large scale visual recognition challenge. International journal of computer vision 115 3 (2015) 211--252.  Olga Russakovsky Jia Deng Hao Su Jonathan Krause Sanjeev Satheesh Sean Ma Zhiheng Huang Andrej Karpathy Aditya Khosla Michael Bernstein et al. 2015. Imagenet large scale visual recognition challenge. International journal of computer vision 115 3 (2015) 211--252.","DOI":"10.1007\/s11263-015-0816-y"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i07.6871"},{"key":"e_1_3_2_1_33_1","volume-title":"Evaluating the visualization of what a deep neural network has learned","author":"Samek Wojciech","year":"2016","unstructured":"Wojciech Samek , Alexander Binder , Gr\u00e9goire Montavon , Sebastian Lapuschkin , and Klaus-Robert M\u00fcller . 2016. Evaluating the visualization of what a deep neural network has learned . IEEE transactions on neural networks and learning systems 28, 11 ( 2016 ), 2660--2673. Wojciech Samek, Alexander Binder, Gr\u00e9goire Montavon, Sebastian Lapuschkin, and Klaus-Robert M\u00fcller. 2016. Evaluating the visualization of what a deep neural network has learned. IEEE transactions on neural networks and learning systems 28, 11 (2016), 2660--2673."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.74"},{"key":"e_1_3_2_1_35_1","volume-title":"Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps. arXiv:1312.6034","author":"Simonyan Karen","year":"2014","unstructured":"Karen Simonyan , Andrea Vedaldi , and Andrew Zisserman . 2014. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps. arXiv:1312.6034 ( 2014 ). Karen Simonyan, Andrea Vedaldi, and Andrew Zisserman. 2014. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps. arXiv:1312.6034 (2014)."},{"key":"e_1_3_2_1_36_1","volume-title":"Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556","author":"Simonyan Karen","year":"2014","unstructured":"Karen Simonyan and AndrewZisserman. 2014. Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556 ( 2014 ). Karen Simonyan and AndrewZisserman. 2014. Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556 (2014)."},{"key":"e_1_3_2_1_37_1","volume-title":"Striving for simplicity: The all convolutional net. arXiv:1412.6806","author":"Springenberg Jost Tobias","year":"2014","unstructured":"Jost Tobias Springenberg , Alexey Dosovitskiy , Thomas Brox , and Martin Riedmiller . 2014. Striving for simplicity: The all convolutional net. arXiv:1412.6806 ( 2014 ). Jost Tobias Springenberg, Alexey Dosovitskiy, Thomas Brox, and Martin Riedmiller. 2014. Striving for simplicity: The all convolutional net. arXiv:1412.6806 (2014)."},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1145\/3394486.3403064"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1109\/EuroSP48549.2020.00018"},{"key":"e_1_3_2_1_40_1","volume-title":"Evaluating explanation without ground truth in interpretable machine learning. arXiv:1907.06831","author":"Yang Fan","year":"2019","unstructured":"Fan Yang , Mengnan Du , and Xia Hu. 2019. Evaluating explanation without ground truth in interpretable machine learning. arXiv:1907.06831 ( 2019 ). Fan Yang, Mengnan Du, and Xia Hu. 2019. Evaluating explanation without ground truth in interpretable machine learning. arXiv:1907.06831 (2019)."},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1145\/3319535.3354209"},{"key":"e_1_3_2_1_42_1","unstructured":"Chih-Kuan Yeh Cheng-Yu Hsieh Arun Suggala David I Inouye and Pradeep K Ravikumar. 2019. On the (in) fidelity and sensitivity of explanations. In Advances in Neural Information Processing Systems. 10967--10978.  Chih-Kuan Yeh Cheng-Yu Hsieh Arun Suggala David I Inouye and Pradeep K Ravikumar. 2019. On the (in) fidelity and sensitivity of explanations. In Advances in Neural Information Processing Systems. 10967--10978."},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-10590-1_53"},{"key":"e_1_3_2_1_44_1","volume-title":"Learning Deep Features for Discriminative Localization. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 2921--2929","author":"Zhou Bolei","year":"2016","unstructured":"Bolei Zhou , Aditya Khosla , Agata Lapedriza , Aude Oliva , and Antonio Torralba . 2016 . Learning Deep Features for Discriminative Localization. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 2921--2929 . Bolei Zhou, Aditya Khosla, Agata Lapedriza, Aude Oliva, and Antonio Torralba. 2016. Learning Deep Features for Discriminative Localization. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 2921--2929."}],"event":{"name":"KDD '21: The 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining","location":"Virtual Event Singapore","acronym":"KDD '21","sponsor":["SIGMOD ACM Special Interest Group on Management of Data","SIGKDD ACM Special Interest Group on Knowledge Discovery in Data"]},"container-title":["Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery &amp; Data Mining"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3447548.3467213","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3447548.3467213","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:18:28Z","timestamp":1750191508000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3447548.3467213"}},"subtitle":["Evaluation of Explainable Artificial Intelligence (XAI) Interpretability through Neural Backdoors"],"short-title":[],"issued":{"date-parts":[[2021,8,14]]},"references-count":44,"alternative-id":["10.1145\/3447548.3467213","10.1145\/3447548"],"URL":"https:\/\/doi.org\/10.1145\/3447548.3467213","relation":{},"subject":[],"published":{"date-parts":[[2021,8,14]]},"assertion":[{"value":"2021-08-14","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}