{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T05:09:35Z","timestamp":1755839375863,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":42,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,6,12]],"date-time":"2022-06-12T00:00:00Z","timestamp":1654992000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,6,12]]},"DOI":"10.1145\/3546930.3547502","type":"proceedings-article","created":{"date-parts":[[2022,8,17]],"date-time":"2022-08-17T23:14:49Z","timestamp":1660778089000},"page":"1-7","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":7,"title":["Context sight"],"prefix":"10.1145","author":[{"given":"Jun","family":"Yuan","sequence":"first","affiliation":[{"name":"New York University"}]},{"given":"Enrico","family":"Bertini","sequence":"additional","affiliation":[{"name":"Northeastern University"}]}],"member":"320","published-online":{"date-parts":[[2022,8,17]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2014.2346660"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"crossref","first-page":"175","DOI":"10.1080\/00031305.1992.10475879","article-title":"An introduction to kernel and nearest-neighbor nonparametric regression","volume":"46","author":"Altman Naomi S","year":"1992","unstructured":"Naomi S Altman . 1992 . An introduction to kernel and nearest-neighbor nonparametric regression . The American Statistician 46 , 3 (1992), 175 -- 185 . Naomi S Altman. 1992. An introduction to kernel and nearest-neighbor nonparametric regression. The American Statistician 46, 3 (1992), 175--185.","journal-title":"The American Statistician"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1111\/rssb.12377"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1207\/S15327051HCI16234_05"},{"key":"e_1_3_2_1_5_1","volume-title":"Do convolutional neural networks learn class hierarchy? IEEE transactions on visualization and computer graphics 24, 1","author":"Bilal Alsallakh","year":"2017","unstructured":"Alsallakh Bilal , Amin Jourabloo , Mao Ye , Xiaoming Liu , and Liu Ren . 2017. Do convolutional neural networks learn class hierarchy? IEEE transactions on visualization and computer graphics 24, 1 ( 2017 ), 152--162. Alsallakh Bilal, Amin Jourabloo, Mao Ye, Xiaoming Liu, and Liu Ren. 2017. Do convolutional neural networks learn class hierarchy? IEEE transactions on visualization and computer graphics 24, 1 (2017), 152--162."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/3490099.3511139"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/VAST47406.2019.8986948"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/3301275.3302289"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2020.3030342"},{"key":"e_1_3_2_1_11_1","volume-title":"Explanations based on the missing: Towards contrastive explanations with pertinent negatives. Advances in neural information processing systems 31","author":"Dhurandhar Amit","year":"2018","unstructured":"Amit Dhurandhar , Pin-Yu Chen , Ronny Luss , Chun-Chen Tu , Paishun Ting , Karthikeyan Shanmugam , and Payel Das . 2018. Explanations based on the missing: Towards contrastive explanations with pertinent negatives. Advances in neural information processing systems 31 ( 2018 ). Amit Dhurandhar, Pin-Yu Chen, Ronny Luss, Chun-Chen Tu, Paishun Ting, Karthikeyan Shanmugam, and Payel Das. 2018. Explanations based on the missing: Towards contrastive explanations with pertinent negatives. Advances in neural information processing systems 31 (2018)."},{"key":"e_1_3_2_1_12_1","unstructured":"FICO. 2018. Explainable Machine Learning Challenge. https:\/\/community.fico.com\/s\/explainable-machine-learning-challenge?tabset-3158a=2. FICO. 2018. Explainable Machine Learning Challenge. https:\/\/community.fico.com\/s\/explainable-machine-learning-challenge?tabset-3158a=2."},{"key":"e_1_3_2_1_13_1","volume-title":"Greedy function approximation: a gradient boosting machine. Annals of statistics","author":"Friedman Jerome H","year":"2001","unstructured":"Jerome H Friedman . 2001. Greedy function approximation: a gradient boosting machine. Annals of statistics ( 2001 ), 1189--1232. Jerome H Friedman. 2001. Greedy function approximation: a gradient boosting machine. Annals of statistics (2001), 1189--1232."},{"key":"e_1_3_2_1_14_1","volume-title":"Artificial neural networks (the multilayer perceptron)---a review of applications in the atmospheric sciences. Atmospheric environment 32, 14--15","author":"Gardner Matt W","year":"1998","unstructured":"Matt W Gardner and SR Dorling . 1998. Artificial neural networks (the multilayer perceptron)---a review of applications in the atmospheric sciences. Atmospheric environment 32, 14--15 ( 1998 ), 2627--2636. Matt W Gardner and SR Dorling. 1998. Artificial neural networks (the multilayer perceptron)---a review of applications in the atmospheric sciences. Atmospheric environment 32, 14--15 (1998), 2627--2636."},{"key":"e_1_3_2_1_15_1","volume-title":"Robustness gym: Unifying the nlp evaluation landscape. arXiv preprint arXiv:2101.04840","author":"Goel Karan","year":"2021","unstructured":"Karan Goel , Nazneen Rajani , Jesse Vig , Samson Tan , Jason Wu , Stephan Zheng , Caiming Xiong , Mohit Bansal , and Christopher R\u00e9. 2021. Robustness gym: Unifying the nlp evaluation landscape. arXiv preprint arXiv:2101.04840 ( 2021 ). Karan Goel, Nazneen Rajani, Jesse Vig, Samson Tan, Jason Wu, Stephan Zheng, Caiming Xiong, Mohit Bansal, and Christopher R\u00e9. 2021. Robustness gym: Unifying the nlp evaluation landscape. arXiv preprint arXiv:2101.04840 (2021)."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3377325.3377536"},{"key":"e_1_3_2_1_17_1","volume-title":"Axel Wendt, and Liu Ren.","author":"Gou Liang","year":"2020","unstructured":"Liang Gou , Lincan Zou , Nanxiang Li , Michael Hofmann , Arvind Kumar Shekar , Axel Wendt, and Liu Ren. 2020 . VATLD: a visual analytics system to assess, understand and improve traffic light detection. IEEE transactions on visualization and computer graphics 27, 2 (2020), 261--271. Liang Gou, Lincan Zou, Nanxiang Li, Michael Hofmann, Arvind Kumar Shekar, Axel Wendt, and Liu Ren. 2020. VATLD: a visual analytics system to assess, understand and improve traffic light detection. IEEE transactions on visualization and computer graphics 27, 2 (2020), 261--271."},{"key":"e_1_3_2_1_18_1","volume-title":"Local rule-based explanations of black box decision systems. arXiv preprint arXiv:1805.10820","author":"Guidotti Riccardo","year":"2018","unstructured":"Riccardo Guidotti , Anna Monreale , Salvatore Ruggieri , Dino Pedreschi , Franco Turini , and Fosca Giannotti . 2018. Local rule-based explanations of black box decision systems. arXiv preprint arXiv:1805.10820 ( 2018 ). Riccardo Guidotti, Anna Monreale, Salvatore Ruggieri, Dino Pedreschi, Franco Turini, and Fosca Giannotti. 2018. Local rule-based explanations of black box decision systems. arXiv preprint arXiv:1805.10820 (2018)."},{"key":"e_1_3_2_1_19_1","volume-title":"GNNLens: A Visual Analytics Approach for Prediction Error Diagnosis of Graph Neural Networks. arXiv preprint arXiv:2011.11048","author":"Jin Zhihua","year":"2020","unstructured":"Zhihua Jin , Yong Wang , Qianwen Wang , Yao Ming , Tengfei Ma , and Huamin Qu. 2020. GNNLens: A Visual Analytics Approach for Prediction Error Diagnosis of Graph Neural Networks. arXiv preprint arXiv:2011.11048 ( 2020 ). Zhihua Jin, Yong Wang, Qianwen Wang, Yao Ming, Tengfei Ma, and Huamin Qu. 2020. GNNLens: A Visual Analytics Approach for Prediction Error Diagnosis of Graph Neural Networks. arXiv preprint arXiv:2011.11048 (2020)."},{"key":"e_1_3_2_1_20_1","volume-title":"The bayesian case model: A generative approach for case-based reasoning and prototype classification. Advances in neural information processing systems 27","author":"Kim Been","year":"2014","unstructured":"Been Kim , Cynthia Rudin , and Julie A Shah . 2014. The bayesian case model: A generative approach for case-based reasoning and prototype classification. Advances in neural information processing systems 27 ( 2014 ). Been Kim, Cynthia Rudin, and Julie A Shah. 2014. The bayesian case model: A generative approach for case-based reasoning and prototype classification. Advances in neural information processing systems 27 (2014)."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1109\/VAST.2017.8585720"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376590"},{"key":"e_1_3_2_1_23_1","volume-title":"A unified approach to interpreting model predictions. Advances in neural information processing systems 30","author":"Lundberg Scott M","year":"2017","unstructured":"Scott M Lundberg and Su-In Lee . 2017. A unified approach to interpreting model predictions. Advances in neural information processing systems 30 ( 2017 ). Scott M Lundberg and Su-In Lee. 2017. A unified approach to interpreting model predictions. Advances in neural information processing systems 30 (2017)."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-34885-4_24"},{"key":"e_1_3_2_1_25_1","unstructured":"Microsoft. 2021. microsoft\/responsible-ai-widgets. https:\/\/github.com\/microsoft\/responsible-ai-widgets. Accessed: 2021-10-08. Microsoft. 2021. microsoft\/responsible-ai-widgets. https:\/\/github.com\/microsoft\/responsible-ai-widgets. Accessed: 2021-10-08."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1145\/3292500.3330908"},{"key":"e_1_3_2_1_27_1","unstructured":"Christoph Molnar. 2022. Interpretable Machine Learning (2 ed.). https:\/\/christophm.github.io\/interpretable-ml-book Christoph Molnar. 2022. Interpretable Machine Learning (2 ed.). https:\/\/christophm.github.io\/interpretable-ml-book"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372850"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.dib.2019.104344"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1145\/2939672.2939778"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11491"},{"key":"e_1_3_2_1_32_1","volume-title":"Beyond accuracy: Behavioral testing of NLP models with CheckList. arXiv preprint arXiv:2005.04118","author":"Ribeiro Marco Tulio","year":"2020","unstructured":"Marco Tulio Ribeiro , Tongshuang Wu , Carlos Guestrin , and Sameer Singh . 2020. Beyond accuracy: Behavioral testing of NLP models with CheckList. arXiv preprint arXiv:2005.04118 ( 2020 ). Marco Tulio Ribeiro, Tongshuang Wu, Carlos Guestrin, and Sameer Singh. 2020. Beyond accuracy: Behavioral testing of NLP models with CheckList. arXiv preprint arXiv:2005.04118 (2020)."},{"key":"e_1_3_2_1_33_1","volume-title":"Not just a black box: Learning important features through propagating activation differences. arXiv preprint arXiv:1605.01713","author":"Shrikumar Avanti","year":"2016","unstructured":"Avanti Shrikumar , Peyton Greenside , Anna Shcherbina , and Anshul Kundaje . 2016. Not just a black box: Learning important features through propagating activation differences. arXiv preprint arXiv:1605.01713 ( 2016 ). Avanti Shrikumar, Peyton Greenside, Anna Shcherbina, and Anshul Kundaje. 2016. Not just a black box: Learning important features through propagating activation differences. arXiv preprint arXiv:1605.01713 (2016)."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"crossref","unstructured":"Ian Tenney James Wexler Jasmijn Bastings Tolga Bolukbasi Andy Coenen Sebastian Gehrmann Ellen Jiang Mahima Pushkarna Carey Radebaugh Emily Reif etal 2020. The language interpretability tool: Extensible interactive visualizations and analysis for NLP models. arXiv preprint arXiv:2008.05122 (2020). Ian Tenney James Wexler Jasmijn Bastings Tolga Bolukbasi Andy Coenen Sebastian Gehrmann Ellen Jiang Mahima Pushkarna Carey Radebaugh Emily Reif et al. 2020. The language interpretability tool: Extensible interactive visualizations and analysis for NLP models. arXiv preprint arXiv:2008.05122 (2020).","DOI":"10.18653\/v1\/2020.emnlp-demos.15"},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.chb.2018.09.009"},{"key":"e_1_3_2_1_36_1","volume-title":"Counterfactual explanations for machine learning: A review. arXiv preprint arXiv:2010.10596","author":"Verma Sahil","year":"2020","unstructured":"Sahil Verma , John Dickerson , and Keegan Hines . 2020. Counterfactual explanations for machine learning: A review. arXiv preprint arXiv:2010.10596 ( 2020 ). Sahil Verma, John Dickerson, and Keegan Hines. 2020. Counterfactual explanations for machine learning: A review. arXiv preprint arXiv:2010.10596 (2020)."},{"key":"e_1_3_2_1_37_1","first-page":"841","article-title":"Counterfactual explanations without opening the black box: Automated decisions and the GDPR","volume":"31","author":"Wachter Sandra","year":"2017","unstructured":"Sandra Wachter , Brent Mittelstadt , and Chris Russell . 2017 . Counterfactual explanations without opening the black box: Automated decisions and the GDPR . Harv. JL & Tech. 31 (2017), 841 . Sandra Wachter, Brent Mittelstadt, and Chris Russell. 2017. Counterfactual explanations without opening the black box: Automated decisions and the GDPR. Harv. JL & Tech. 31 (2017), 841.","journal-title":"Harv. JL & Tech."},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300831"},{"key":"e_1_3_2_1_39_1","volume-title":"Deep-vid: Deep visual interpretation and diagnosis for image classifiers via knowledge distillation","author":"Wang Junpeng","year":"2019","unstructured":"Junpeng Wang , Liang Gou , Wei Zhang , Hao Yang , and Han-Wei Shen . 2019 . Deep-vid: Deep visual interpretation and diagnosis for image classifiers via knowledge distillation . IEEE transactions on visualization and computer graphics 25, 6 (2019), 2168--2180. Junpeng Wang, Liang Gou, Wei Zhang, Hao Yang, and Han-Wei Shen. 2019. Deep-vid: Deep visual interpretation and diagnosis for image classifiers via knowledge distillation. IEEE transactions on visualization and computer graphics 25, 6 (2019), 2168--2180."},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1073"},{"key":"e_1_3_2_1_41_1","volume-title":"Jeffrey Heer, and Daniel S Weld.","author":"Wu Tongshuang","year":"2021","unstructured":"Tongshuang Wu , Marco Tulio Ribeiro , Jeffrey Heer, and Daniel S Weld. 2021 . Polyjuice : Generating counterfactuals for explaining, evaluating, and improving models. arXiv preprint arXiv:2101.00288 (2021). Tongshuang Wu, Marco Tulio Ribeiro, Jeffrey Heer, and Daniel S Weld. 2021. Polyjuice: Generating counterfactuals for explaining, evaluating, and improving models. arXiv preprint arXiv:2101.00288 (2021)."},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/3490099.3511146"},{"key":"e_1_3_2_1_43_1","volume-title":"Manifold: A model-agnostic framework for interpretation and diagnosis of machine learning models","author":"Zhang Jiawei","year":"2018","unstructured":"Jiawei Zhang , Yang Wang , Piero Molino , Lezhi Li , and David S Ebert . 2018 . Manifold: A model-agnostic framework for interpretation and diagnosis of machine learning models . IEEE transactions on visualization and computer graphics 25, 1 (2018), 364--373. Jiawei Zhang, Yang Wang, Piero Molino, Lezhi Li, and David S Ebert. 2018. Manifold: A model-agnostic framework for interpretation and diagnosis of machine learning models. IEEE transactions on visualization and computer graphics 25, 1 (2018), 364--373."}],"event":{"name":"SIGMOD\/PODS '22: International Conference on Management of Data","sponsor":["SIGMOD ACM Special Interest Group on Management of Data"],"location":"Philadelphia Pennsylvania","acronym":"SIGMOD\/PODS '22"},"container-title":["Proceedings of the Workshop on Human-In-the-Loop Data Analytics"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3546930.3547502","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3546930.3547502","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:00:41Z","timestamp":1750186841000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3546930.3547502"}},"subtitle":["model understanding and debugging via interpretable context"],"short-title":[],"issued":{"date-parts":[[2022,6,12]]},"references-count":42,"alternative-id":["10.1145\/3546930.3547502","10.1145\/3546930"],"URL":"https:\/\/doi.org\/10.1145\/3546930.3547502","relation":{},"subject":[],"published":{"date-parts":[[2022,6,12]]},"assertion":[{"value":"2022-08-17","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}