{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:23:28Z","timestamp":1750220608398,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":40,"publisher":"ACM","license":[{"start":{"date-parts":[[2020,6,8]],"date-time":"2020-06-08T00:00:00Z","timestamp":1591574400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2020,6,8]]},"DOI":"10.1145\/3372278.3390672","type":"proceedings-article","created":{"date-parts":[[2020,6,2]],"date-time":"2020-06-02T04:35:27Z","timestamp":1591072527000},"page":"35-43","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":9,"title":["Explaining with Counter Visual Attributes and Examples"],"prefix":"10.1145","author":[{"given":"Sadaf","family":"Gulshad","sequence":"first","affiliation":[{"name":"University of Amsterdam, Amsterdam, Netherlands"}]},{"given":"Arnold","family":"Smeulders","sequence":"additional","affiliation":[{"name":"University of Amsterdam, Amsterdam, Netherlands"}]}],"member":"320","published-online":{"date-parts":[[2020,6,8]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"crossref","unstructured":"Zeynep Akata Scott Reed Daniel Walter Honglak Lee and Bernt Schiele. 2015. Evaluation of output embeddings for fine-grained image classification. In CVPR.  Zeynep Akata Scott Reed Daniel Walter Honglak Lee and Bernt Schiele. 2015. Evaluation of output embeddings for fine-grained image classification. In CVPR.","DOI":"10.1109\/CVPR.2015.7298911"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"crossref","unstructured":"Peter Anderson Xiaodong He Chris Buehler Damien Teney Mark Johnson Stephen Gould and Lei Zhang. 2018. Bottom-up and top-down attention for image captioning and visual question answering. In CVPR.  Peter Anderson Xiaodong He Chris Buehler Damien Teney Mark Johnson Stephen Gould and Lei Zhang. 2018. Bottom-up and top-down attention for image captioning and visual question answering. In CVPR.","DOI":"10.1109\/CVPR.2018.00636"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"crossref","unstructured":"Lisa Anne Hendricks Ronghang Hu Trevor Darrell and Zeynep Akata. 2018. Grounding visual explanations. In ECCV.  Lisa Anne Hendricks Ronghang Hu Trevor Darrell and Zeynep Akata. 2018. Grounding visual explanations. In ECCV.","DOI":"10.1007\/978-3-030-01216-8_17"},{"volume-title":"Towards evaluating the robustness of neural networks","author":"Carlini Nicholas","key":"e_1_3_2_1_4_1","unstructured":"Nicholas Carlini and David Wagner . 2017. Towards evaluating the robustness of neural networks . In SP. IEEE. Nicholas Carlini and David Wagner. 2017. Towards evaluating the robustness of neural networks. In SP. IEEE."},{"volume-title":"[n.d.]. What Makes Paris Look like Paris? ACM Transactions on Graphics (SIGGRAPH)","author":"Doersch Carl","key":"e_1_3_2_1_5_1","unstructured":"Carl Doersch , Saurabh Singh , Abhinav Gupta , Josef Sivic , and Alexei A. Efros . [n.d.]. What Makes Paris Look like Paris? ACM Transactions on Graphics (SIGGRAPH) , Vol. 31 , 4 ( [n.,d.]), 101: 1--101:9. Carl Doersch, Saurabh Singh, Abhinav Gupta, Josef Sivic, and Alexei A. Efros. [n.d.]. What Makes Paris Look like Paris? ACM Transactions on Graphics (SIGGRAPH), Vol. 31, 4 ( [n.,d.]), 101:1--101:9."},{"key":"e_1_3_2_1_6_1","volume-title":"Towards interpretable deep neural networks by leveraging adversarial examples. arXiv","author":"Dong Yinpeng","year":"2017","unstructured":"Yinpeng Dong , Hang Su , Jun Zhu , and Fan Bao . 2017a. Towards interpretable deep neural networks by leveraging adversarial examples. arXiv ( 2017 ). Yinpeng Dong, Hang Su, Jun Zhu, and Fan Bao. 2017a. Towards interpretable deep neural networks by leveraging adversarial examples. arXiv (2017)."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"crossref","unstructured":"Yinpeng Dong Hang Su Jun Zhu and Bo Zhang. 2017b. Improving interpretability of deep neural networks with semantic information. In CVPR.  Yinpeng Dong Hang Su Jun Zhu and Bo Zhang. 2017b. Improving interpretability of deep neural networks with semantic information. In CVPR.","DOI":"10.1109\/CVPR.2017.110"},{"key":"e_1_3_2_1_8_1","volume-title":"Techniques for Interpretable Machine Learning. arXiv","author":"Du Mengnan","year":"2018","unstructured":"Mengnan Du , Ninghao Liu , and Xia Hu. 2018. Techniques for Interpretable Machine Learning. arXiv ( 2018 ). Mengnan Du, Ninghao Liu, and Xia Hu. 2018. Techniques for Interpretable Machine Learning. arXiv (2018)."},{"key":"e_1_3_2_1_9_1","volume-title":"Interpretable explanations of black boxes by meaningful perturbation. arXiv","author":"Fong Ruth C","year":"2017","unstructured":"Ruth C Fong and Andrea Vedaldi . 2017. Interpretable explanations of black boxes by meaningful perturbation. arXiv ( 2017 ). Ruth C Fong and Andrea Vedaldi. 2017. Interpretable explanations of black boxes by meaningful perturbation. arXiv (2017)."},{"key":"e_1_3_2_1_10_1","unstructured":"Ian Goodfellow Jonathon Shlens and Christian Szegedy. 2015. Explaining and Harnessing Adversarial Examples. In ICLR.  Ian Goodfellow Jonathon Shlens and Christian Szegedy. 2015. Explaining and Harnessing Adversarial Examples. In ICLR."},{"key":"e_1_3_2_1_11_1","unstructured":"Yash Goyal Ziyan Wu Jan Ernst Dhruv Batra Devi Parikh and Stefan Lee. 2019. Counterfactual Visual Explanations. (2019).  Yash Goyal Ziyan Wu Jan Ernst Dhruv Batra Devi Parikh and Stefan Lee. 2019. Counterfactual Visual Explanations. (2019)."},{"volume-title":"Generating visual explanations","author":"Hendricks Lisa Anne","key":"e_1_3_2_1_12_1","unstructured":"Lisa Anne Hendricks , Zeynep Akata , Marcus Rohrbach , Jeff Donahue , Bernt Schiele , and Trevor Darrell . 2016. Generating visual explanations . In ECCV. Springer . Lisa Anne Hendricks, Zeynep Akata, Marcus Rohrbach, Jeff Donahue, Bernt Schiele, and Trevor Darrell. 2016. Generating visual explanations. In ECCV. Springer."},{"key":"e_1_3_2_1_13_1","volume-title":"Generating Counterfactual Explanations with Natural Language. In ICML Workshop on Human Interpretability in Machine Learning. 95--98","author":"Hendricks Lisa Anne","year":"2018","unstructured":"Lisa Anne Hendricks , Ronghang Hu , Trevor Darrell , and Zeynep Akata . 2018 . Generating Counterfactual Explanations with Natural Language. In ICML Workshop on Human Interpretability in Machine Learning. 95--98 . Lisa Anne Hendricks, Ronghang Hu, Trevor Darrell, and Zeynep Akata. 2018. Generating Counterfactual Explanations with Natural Language. In ICML Workshop on Human Interpretability in Machine Learning. 95--98."},{"key":"e_1_3_2_1_14_1","unstructured":"Cheng-Yu Hsieh Chih-Kuan Yeh Xuanqing Liu Pradeep Ravikumar Seungyeon Kim Sanjiv Kumar and Cho-Jui Hsieh. 2020. Evaluations and Methods for Explanation through Robustness Analysis.  Cheng-Yu Hsieh Chih-Kuan Yeh Xuanqing Liu Pradeep Ravikumar Seungyeon Kim Sanjiv Kumar and Cho-Jui Hsieh. 2020. Evaluations and Methods for Explanation through Robustness Analysis."},{"key":"e_1_3_2_1_15_1","volume-title":"Recent research advances on interactive machine learning. Journal of Visualization","author":"Jiang Liu","year":"2018","unstructured":"Liu Jiang , Shixia Liu , and Changjian Chen . 2018. Recent research advances on interactive machine learning. Journal of Visualization ( 2018 ). Liu Jiang, Shixia Liu, and Changjian Chen. 2018. Recent research advances on interactive machine learning. Journal of Visualization (2018)."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00880"},{"key":"e_1_3_2_1_17_1","unstructured":"Jinkyu Kim Anna Rohrbach Trevor Darrell John Canny and Zeynep Akata. 2018. Textual explanations for self-driving vehicles. In ECCV.  Jinkyu Kim Anna Rohrbach Trevor Darrell John Canny and Zeynep Akata. 2018. Textual explanations for self-driving vehicles. In ECCV."},{"key":"e_1_3_2_1_18_1","volume-title":"et almbox","author":"Krishna Ranjay","year":"2017","unstructured":"Ranjay Krishna , Yuke Zhu , Oliver Groth , Justin Johnson , Kenji Hata , Joshua Kravitz , Stephanie Chen , Yannis Kalantidis , Li-Jia Li , David A Shamma , et almbox . 2017 . Visual genome: Connecting language and vision using crowdsourced dense image annotations. IJCV ( 2017). Ranjay Krishna, Yuke Zhu, Oliver Groth, Justin Johnson, Kenji Hata, Joshua Kravitz, Stephanie Chen, Yannis Kalantidis, Li-Jia Li, David A Shamma, et almbox. 2017. Visual genome: Connecting language and vision using crowdsourced dense image annotations. IJCV (2017)."},{"key":"e_1_3_2_1_19_1","volume-title":"ICLR workshop","author":"Kurakin Alexey","year":"2017","unstructured":"Alexey Kurakin , Ian Goodfellow , and Samy Bengio . 2017 . Adversarial examples in the physical world . ICLR workshop (2017). Alexey Kurakin, Ian Goodfellow, and Samy Bengio. 2017. Adversarial examples in the physical world. ICLR workshop (2017)."},{"volume-title":"Learning to detect unseen object classes by between-class attribute transfer","author":"Lampert Christoph H","key":"e_1_3_2_1_20_1","unstructured":"Christoph H Lampert , Hannes Nickisch , and Stefan Harmeling . 2009. Learning to detect unseen object classes by between-class attribute transfer . In CVPR. IEEE. Christoph H Lampert, Hannes Nickisch, and Stefan Harmeling. 2009. Learning to detect unseen object classes by between-class attribute transfer. In CVPR. IEEE."},{"key":"e_1_3_2_1_21_1","volume-title":"Towards deep learning models resistant to adversarial attacks. ICLR","author":"Madry Aleksander","year":"2018","unstructured":"Aleksander Madry , Aleksandar Makelov , Ludwig Schmidt , Dimitris Tsipras , and Adrian Vladu . 2018. Towards deep learning models resistant to adversarial attacks. ICLR ( 2018 ). Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu. 2018. Towards deep learning models resistant to adversarial attacks. ICLR (2018)."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"crossref","unstructured":"Christoph Molnar. 2019. Interpretable Machine Learning. https:\/\/christophm.github.io\/interpretable-ml-book\/.  Christoph Molnar. 2019. Interpretable Machine Learning. https:\/\/christophm.github.io\/interpretable-ml-book\/.","DOI":"10.21105\/joss.00786"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"crossref","unstructured":"Seyed-Mohsen Moosavi-Dezfooli Alhussein Fawzi and Pascal Frossard. 2016. Deepfool: a simple and accurate method to fool deep neural networks. In CVPR.  Seyed-Mohsen Moosavi-Dezfooli Alhussein Fawzi and Pascal Frossard. 2016. Deepfool: a simple and accurate method to fool deep neural networks. In CVPR.","DOI":"10.1109\/CVPR.2016.282"},{"volume-title":"EuroS&P","author":"Papernot Nicolas","key":"e_1_3_2_1_24_1","unstructured":"Nicolas Papernot , Patrick McDaniel , Somesh Jha , Matt Fredrikson , Z Berkay Celik , and Ananthram Swami . 2016. The limitations of deep learning in adversarial settings . In EuroS&P . IEEE. Nicolas Papernot, Patrick McDaniel, Somesh Jha, Matt Fredrikson, Z Berkay Celik, and Ananthram Swami. 2016. The limitations of deep learning in adversarial settings. In EuroS&P. IEEE."},{"key":"e_1_3_2_1_25_1","volume-title":"Lisa Anne Hendricks","author":"Park Dong Huk","year":"2018","unstructured":"Dong Huk Park , Lisa Anne Hendricks , Zeynep Akata, Bernt Schiele , Trevor Darrell, and Marcus Rohrbach. 2018 . Multimodal explanations: Justifying decisions and pointing to the evidence. In CVPR. Dong Huk Park, Lisa Anne Hendricks, Zeynep Akata, Bernt Schiele, Trevor Darrell, and Marcus Rohrbach. 2018. Multimodal explanations: Justifying decisions and pointing to the evidence. In CVPR."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"crossref","unstructured":"Marco Tulio Ribeiro Sameer Singh and Carlos Guestrin. 2016. Why should i trust you?: Explaining the predictions of any classifier. In ACM SIGKDD.  Marco Tulio Ribeiro Sameer Singh and Carlos Guestrin. 2016. Why should i trust you?: Explaining the predictions of any classifier. In ACM SIGKDD.","DOI":"10.18653\/v1\/N16-3020"},{"key":"e_1_3_2_1_27_1","volume-title":"Grad-cam: Visual explanations from deep networks via gradient-based localization. In ICCV.","author":"Selvaraju Ramprasaath R","year":"2017","unstructured":"Ramprasaath R Selvaraju , Michael Cogswell , Abhishek Das , Ramakrishna Vedantam , Devi Parikh , and Dhruv Batra . 2017 . Grad-cam: Visual explanations from deep networks via gradient-based localization. In ICCV. Ramprasaath R Selvaraju, Michael Cogswell, Abhishek Das, Ramakrishna Vedantam, Devi Parikh, and Dhruv Batra. 2017. Grad-cam: Visual explanations from deep networks via gradient-based localization. In ICCV."},{"key":"e_1_3_2_1_28_1","unstructured":"Avanti Shrikumar Peyton Greenside and Anshul Kundaje. 2017. Learning important features through propagating activation differences. In ICML.  Avanti Shrikumar Peyton Greenside and Anshul Kundaje. 2017. Learning important features through propagating activation differences. In ICML."},{"key":"e_1_3_2_1_29_1","volume-title":"Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv","author":"Simonyan Karen","year":"2013","unstructured":"Karen Simonyan , Andrea Vedaldi , and Andrew Zisserman . 2013. Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv ( 2013 ). Karen Simonyan, Andrea Vedaldi, and Andrew Zisserman. 2013. Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv (2013)."},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-33709-3_6"},{"key":"e_1_3_2_1_31_1","volume-title":"Danilo Vasconcellos Vargas, and Kouichi Sakurai","author":"Su Jiawei","year":"2019","unstructured":"Jiawei Su , Danilo Vasconcellos Vargas, and Kouichi Sakurai . 2019 . One pixel attack for fooling deep neural networks. TEVC ( 2019). Jiawei Su, Danilo Vasconcellos Vargas, and Kouichi Sakurai. 2019. One pixel attack for fooling deep neural networks. TEVC (2019)."},{"key":"e_1_3_2_1_32_1","unstructured":"Mukund Sundararajan Ankur Taly and Qiqi Yan. 2017. Axiomatic attribution for deep networks. In ICML.  Mukund Sundararajan Ankur Taly and Qiqi Yan. 2017. Axiomatic attribution for deep networks. In ICML."},{"key":"e_1_3_2_1_33_1","volume-title":"Intriguing properties of neural networks. ICLR","author":"Szegedy Christian","year":"2013","unstructured":"Christian Szegedy , Wojciech Zaremba , Ilya Sutskever , Joan Bruna , Dumitru Erhan , Ian Goodfellow , and Rob Fergus . 2013. Intriguing properties of neural networks. ICLR ( 2013 ). Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, and Rob Fergus. 2013. Intriguing properties of neural networks. ICLR (2013)."},{"key":"e_1_3_2_1_34_1","volume-title":"Robustness may be at odds with accuracy. stat","author":"Tsipras Dimitris","year":"2018","unstructured":"Dimitris Tsipras , Shibani Santurkar , Logan Engstrom , Alexander Turner , and Aleksander Madry . 2018. Robustness may be at odds with accuracy. stat , Vol. 1050 ( 2018 ). Dimitris Tsipras, Shibani Santurkar, Logan Engstrom, Alexander Turner, and Aleksander Madry. 2018. Robustness may be at odds with accuracy. stat, Vol. 1050 (2018)."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"crossref","unstructured":"Hristina Uzunova Jan Ehrhardt Timo Kepp and Heinz Handels. 2019. Interpretable explanations of black box classifiers applied on medical images by meaningful perturbations using variational autoencoders. In Medical Imaging 2019: Image Processing.  Hristina Uzunova Jan Ehrhardt Timo Kepp and Heinz Handels. 2019. Interpretable explanations of black box classifiers applied on medical images by meaningful perturbations using variational autoencoders. In Medical Imaging 2019: Image Processing.","DOI":"10.1117\/12.2511964"},{"key":"e_1_3_2_1_36_1","unstructured":"Catherine Wah Steve Branson Peter Welinder Pietro Perona and Serge Belongie. 2011. The caltech-ucsd birds-200--2011 dataset. (2011).  Catherine Wah Steve Branson Peter Welinder Pietro Perona and Serge Belongie. 2011. The caltech-ucsd birds-200--2011 dataset. (2011)."},{"volume-title":"Visualizing and understanding convolutional networks","author":"Zeiler Matthew D","key":"e_1_3_2_1_37_1","unstructured":"Matthew D Zeiler and Rob Fergus . 2014. Visualizing and understanding convolutional networks . In ECCV. Springer . Matthew D Zeiler and Rob Fergus. 2014. Visualizing and understanding convolutional networks. In ECCV. Springer."},{"key":"e_1_3_2_1_38_1","volume-title":"Interpreting Adversarially Trained Convolutional Neural Networks. arXiv","author":"Zhang Tianyuan","year":"2019","unstructured":"Tianyuan Zhang and Zhanxing Zhu . 2019. Interpreting Adversarially Trained Convolutional Neural Networks. arXiv ( 2019 ). Tianyuan Zhang and Zhanxing Zhu. 2019. Interpreting Adversarially Trained Convolutional Neural Networks. arXiv (2019)."},{"key":"e_1_3_2_1_39_1","volume-title":"A Large-scale Attribute Dataset for Zero-shot Learning. arXiv","author":"Zhao Bo","year":"2018","unstructured":"Bo Zhao , Yanwei Fu , Rui Liang , Jiahong Wu , Yonggang Wang , and Yizhou Wang . 2018. A Large-scale Attribute Dataset for Zero-shot Learning. arXiv ( 2018 ). Bo Zhao, Yanwei Fu, Rui Liang, Jiahong Wu, Yonggang Wang, and Yizhou Wang. 2018. A Large-scale Attribute Dataset for Zero-shot Learning. arXiv (2018)."},{"key":"e_1_3_2_1_40_1","volume-title":"Visualizing deep neural network decisions: Prediction difference analysis. ICLR","author":"Zintgraf Luisa M","year":"2017","unstructured":"Luisa M Zintgraf , Taco S Cohen , Tameem Adel , and Max Welling . 2017. Visualizing deep neural network decisions: Prediction difference analysis. ICLR ( 2017 ). Luisa M Zintgraf, Taco S Cohen, Tameem Adel, and Max Welling. 2017. Visualizing deep neural network decisions: Prediction difference analysis. ICLR (2017)."}],"event":{"name":"ICMR '20: International Conference on Multimedia Retrieval","sponsor":["SIGMM ACM Special Interest Group on Multimedia"],"location":"Dublin Ireland","acronym":"ICMR '20"},"container-title":["Proceedings of the 2020 International Conference on Multimedia Retrieval"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3372278.3390672","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3372278.3390672","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T21:32:10Z","timestamp":1750195930000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3372278.3390672"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,6,8]]},"references-count":40,"alternative-id":["10.1145\/3372278.3390672","10.1145\/3372278"],"URL":"https:\/\/doi.org\/10.1145\/3372278.3390672","relation":{},"subject":[],"published":{"date-parts":[[2020,6,8]]},"assertion":[{"value":"2020-06-08","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}