{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T00:10:03Z","timestamp":1755821403232,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":45,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,4,30]],"date-time":"2023-04-30T00:00:00Z","timestamp":1682812800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,4,30]]},"DOI":"10.1145\/3543873.3587652","type":"proceedings-article","created":{"date-parts":[[2023,4,28]],"date-time":"2023-04-28T11:36:14Z","timestamp":1682681774000},"page":"950-958","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["CLIME: Completeness-Constrained LIME"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-3649-7747","authenticated-orcid":false,"given":"Claudia","family":"Roberts","sequence":"first","affiliation":[{"name":"Princeton University, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-3651-6197","authenticated-orcid":false,"given":"Ehtsham","family":"Elahi","sequence":"additional","affiliation":[{"name":"Netflix, Inc., USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-1914-8354","authenticated-orcid":false,"given":"Ashok","family":"Chandrashekar","sequence":"additional","affiliation":[{"name":"WarnerMedia, USA"}]}],"member":"320","published-online":{"date-parts":[[2023,4,30]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2021.115849"},{"key":"e_1_3_2_1_2_1","volume-title":"Advances in Neural Information Processing Systems, S.\u00a0Bengio, H.\u00a0Wallach, H.\u00a0Larochelle, K.\u00a0Grauman, N.\u00a0Cesa-Bianchi, and R.\u00a0Garnett (Eds.). Vol.\u00a031. Curran Associates","author":"Alvarez\u00a0Melis David","year":"2018","unstructured":"David Alvarez\u00a0Melis and Tommi Jaakkola. 2018. Towards Robust Interpretability with Self-Explaining Neural Networks. In Advances in Neural Information Processing Systems, S.\u00a0Bengio, H.\u00a0Wallach, H.\u00a0Larochelle, K.\u00a0Grauman, N.\u00a0Cesa-Bianchi, and R.\u00a0Garnett (Eds.). Vol.\u00a031. Curran Associates, Inc.https:\/\/proceedings.neurips.cc\/paper\/2018\/file\/3e9f0fc9b2f89e043bc6233994dfcf76-Paper.pdf"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-981-10-0557-2_87"},{"volume-title":"Pattern Recognition and Machine Learning","author":"Bishop M.","key":"e_1_3_2_1_4_1","unstructured":"Christophper\u00a0M. Bishop. 2006. Pattern Recognition and Machine Learning. Chapter 3, Section 2."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.4304\/jcp.6.9.1896-1902"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3220063"},{"key":"e_1_3_2_1_7_1","volume-title":"Towards A Rigorous Science of Interpretable Machine Learning. arXiv: Machine Learning","author":"Doshi-Velez Finale","year":"2017","unstructured":"Finale Doshi-Velez and Been Kim. 2017. Towards A Rigorous Science of Interpretable Machine Learning. arXiv: Machine Learning (2017)."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3308558.3313545"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/BHI50953.2021.9508618"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/CBMS.2019.00065"},{"key":"e_1_3_2_1_11_1","volume-title":"TrustyAI Explainability Toolkit. CoRR abs\/2104.12717","author":"Geada Rob","year":"2021","unstructured":"Rob Geada, Tommaso Teofili, Rui Vieira, Rebecca Whitworth, and Daniele Zonca. 2021. TrustyAI Explainability Toolkit. CoRR abs\/2104.12717 (2021). arXiv:2104.12717https:\/\/arxiv.org\/abs\/2104.12717"},{"key":"e_1_3_2_1_12_1","volume-title":"CEUR Workshop Proceedings 1153 (01","author":"Gkika Sofia","year":"2014","unstructured":"Sofia Gkika and George Lekakos. 2014. The persuasive role of Explanations in Recommender Systems. CEUR Workshop Proceedings 1153 (01 2014), 59\u201368."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/2827872"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/358916.358995"},{"key":"e_1_3_2_1_15_1","volume-title":"The Promise and Peril of Human Evaluation for Model Interpretability. ArXiv abs\/1711.07414","author":"Herman Bernease","year":"2017","unstructured":"Bernease Herman. 2017. The Promise and Peril of Human Evaluation for Model Interpretability. ArXiv abs\/1711.07414 (2017)."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1007\/s13278-020-0626-2"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445941"},{"volume-title":"Proceedings of the International Conference on Frontiers of Intelligent Computing: Theory and Applications (FICTA), Suresh\u00a0Chandra Satapathy, Siba\u00a0K","author":"Kumar Akshi","key":"e_1_3_2_1_18_1","unstructured":"Akshi Kumar and Abhilasha Sharma. 2013. Alleviating Sparsity and Scalability Issues in Collaborative Filtering Based Recommender Systems. In Proceedings of the International Conference on Frontiers of Intelligent Computing: Theory and Applications (FICTA), Suresh\u00a0Chandra Satapathy, Siba\u00a0K. Udgata, and Bhabendra\u00a0Narayan Biswal (Eds.). Springer Berlin Heidelberg, Berlin, Heidelberg, 103\u2013112."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1046"},{"key":"e_1_3_2_1_20_1","volume-title":"Mining of Massive Datasets","author":"Leskovec Jure","unstructured":"Jure Leskovec, Anand Rajaraman, and Jeffrey\u00a0David Ullman. 2014. Mining of Massive Datasets (2nd ed.). Cambridge University Press, USA.","edition":"2"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/3178876.3186150"},{"key":"e_1_3_2_1_22_1","volume-title":"Do Explanations Reflect Decisions? A Machine-centric Strategy to Quantify the Performance of Explainability Algorithms. ArXiv abs\/1910.07387","author":"Lin Zhong\u00a0Qiu","year":"2019","unstructured":"Zhong\u00a0Qiu Lin, Mohammad\u00a0Javad Shafiee, Stanislav Bochkarev, Michael\u00a0St. Jules, Xiao\u00a0Yu Wang, and Alexander Wong. 2019. Do Explanations Reflect Decisions? A Machine-centric Strategy to Quantify the Performance of Explainability Algorithms. ArXiv abs\/1910.07387 (2019)."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/3233231"},{"volume-title":"Proceedings of the 31st International Conference on Neural Information Processing Systems","author":"M.","key":"e_1_3_2_1_24_1","unstructured":"Scott\u00a0M. Lundberg and Su-In Lee. 2017. A Unified Approach to Interpreting Model Predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems (Long Beach, California, USA) (NIPS\u201917). Curran Associates Inc., Red Hook, NY, USA, 4768\u20134777."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"crossref","unstructured":"Xin Man and Ernest\u00a0P. Chan. 2020. The Best Way to Select Features? Comparing MDA LIME and SHAP.","DOI":"10.3905\/jfds.2020.1.047"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","unstructured":"Andreas Messalas Christos Makris and Yannis Kanellopoulos. 2019. Model-Agnostic Interpretability with Shapley Values. https:\/\/doi.org\/10.1109\/IISA.2019.8900669","DOI":"10.1109\/IISA.2019.8900669"},{"key":"e_1_3_2_1_27_1","unstructured":"Christoph Molnar. 2019. Interpretable Machine Learning. https:\/\/christophm.github.io\/interpretable-ml-book\/."},{"key":"e_1_3_2_1_28_1","unstructured":"Christoph Molnar. 2020. Limitations of Interpretable Machine Learning Methods. https:\/\/github.com\/compstat-lmu\/iml_methods_limitations."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"crossref","unstructured":"Ramaravind\u00a0Kommiya Mothilal Divyat Mahajan Chenhao Tan and Amit Sharma. 2021. Towards Unifying Feature Attribution and Counterfactual Explanations: Different Means to the Same End. In AAAI\/ACM Conference on AI Ethics and Society (AIES). https:\/\/www.microsoft.com\/en-us\/research\/publication\/towards-unifying-feature-attribution-and-counterfactual-explanations-different-means-to-the-same-end\/","DOI":"10.1145\/3461702.3462597"},{"key":"e_1_3_2_1_30_1","volume-title":"Counterfactual Explanation Algorithms for Behavioral and Textual Data. CoRR abs\/1912.01819","author":"Ramon Yanou","year":"2019","unstructured":"Yanou Ramon, David Martens, Foster\u00a0J. Provost, and Theodoros Evgeniou. 2019. Counterfactual Explanation Algorithms for Behavioral and Textual Data. CoRR abs\/1912.01819 (2019). arXiv:1912.01819http:\/\/arxiv.org\/abs\/1912.01819"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/2939672.2939778"},{"key":"e_1_3_2_1_32_1","volume-title":"User profile as a bridge in cross-domain recommender systems. Applied Intelligence (01","author":"Sahu Ashish","year":"2019","unstructured":"Ashish Sahu and Pragya Dwivedi. 2019. User profile as a bridge in cross-domain recommender systems. Applied Intelligence (01 2019)."},{"key":"e_1_3_2_1_33_1","volume-title":"Advances in Neural Information Processing Systems, H.\u00a0Wallach, H.\u00a0Larochelle, A.\u00a0Beygelzimer, F.\u00a0d'Alch\u00e9-Buc, E.\u00a0Fox, and R.\u00a0Garnett (Eds.). Vol.\u00a032. Curran Associates","author":"Schwab Patrick","year":"2019","unstructured":"Patrick Schwab and Walter Karlen. 2019. CXPlain: Causal Explanations for Model Interpretation under Uncertainty. In Advances in Neural Information Processing Systems, H.\u00a0Wallach, H.\u00a0Larochelle, A.\u00a0Beygelzimer, F.\u00a0d'Alch\u00e9-Buc, E.\u00a0Fox, and R.\u00a0Garnett (Eds.). Vol.\u00a032. Curran Associates, Inc.https:\/\/proceedings.neurips.cc\/paper\/2019\/file\/3ab6be46e1d6b21d59a3c3a0b9d0f6ef-Paper.pdf"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.7249\/P0295"},{"key":"e_1_3_2_1_35_1","volume-title":"Proceedings of the 34th International Conference on Machine Learning -","volume":"70","author":"Shrikumar Avanti","year":"2017","unstructured":"Avanti Shrikumar, Peyton Greenside, and Anshul Kundaje. 2017. Learning Important Features through Propagating Activation Differences. In Proceedings of the 34th International Conference on Machine Learning - Volume 70 (Sydney, NSW, Australia) (ICML\u201917). JMLR.org, 3145\u20133153."},{"key":"e_1_3_2_1_36_1","volume-title":"Data Staining: A Method for Comparing Faithfulness of Explainers.","author":"Sippy Jacob","year":"2020","unstructured":"Jacob Sippy, Gagan Bansal, and Daniel\u00a0S. Weld. 2020. Data Staining: A Method for Comparing Faithfulness of Explainers."},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3375627.3375830"},{"key":"e_1_3_2_1_38_1","volume-title":"Proceedings of the 34th International Conference on Machine Learning -","volume":"70","author":"Sundararajan Mukund","year":"2017","unstructured":"Mukund Sundararajan, Ankur Taly, and Qiqi Yan. 2017. Axiomatic Attribution for Deep Networks. In Proceedings of the 34th International Conference on Machine Learning - Volume 70 (Sydney, NSW, Australia) (ICML\u201917). JMLR.org, 3319\u20133328."},{"key":"e_1_3_2_1_39_1","volume-title":"Faithfully Explaining Rankings in a News Recommender System. ArXiv abs\/1805.05447","author":"Hoeve Maartje","year":"2018","unstructured":"Maartje ter Hoeve, Anne Schuth, Daan Odijk, and M. de Rijke. 2018. Faithfully Explaining Rankings in a News Recommender System. ArXiv abs\/1805.05447 (2018)."},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDEW.2007.4401070"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11257-011-9117-5"},{"key":"e_1_3_2_1_42_1","volume-title":"Article arXiv:2012.04218 (Dec.","author":"Velmurugan Mythreyi","year":"2020","unstructured":"Mythreyi Velmurugan, Chun Ouyang, Catarina Moreira, and Renuka Sindhgatta. 2020. Evaluating Explainable Methods for Predictive Process Analytics: A Functionally-Grounded Approach. arXiv e-prints, Article arXiv:2012.04218 (Dec. 2020), arXiv:2012.04218\u00a0pages. arxiv:2012.04218\u00a0[cs.AI]"},{"key":"e_1_3_2_1_43_1","volume-title":"Statistical stability indices for LIME: obtaining reliable explanations for Machine Learning models. CoRR abs\/2001.11757","author":"Visani Giorgio","year":"2020","unstructured":"Giorgio Visani, Enrico Bagli, Federico Chesani, Alessandro Poluzzi, and Davide Capuzzo. 2020. Statistical stability indices for LIME: obtaining reliable explanations for Machine Learning models. CoRR abs\/2001.11757 (2020). arXiv:2001.11757https:\/\/arxiv.org\/abs\/2001.11757"},{"key":"e_1_3_2_1_44_1","volume-title":"Better sampling in explanation methods can prevent dieselgate-like deception. CoRR abs\/2101.11702","author":"Vres Domen","year":"2021","unstructured":"Domen Vres and Marko Robnik-Sikonja. 2021. Better sampling in explanation methods can prevent dieselgate-like deception. CoRR abs\/2101.11702 (2021). arXiv:2101.11702https:\/\/arxiv.org\/abs\/2101.11702"},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1145\/3447548.3467274"}],"event":{"name":"WWW '23: The ACM Web Conference 2023","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"],"location":"Austin TX USA","acronym":"WWW '23"},"container-title":["Companion Proceedings of the ACM Web Conference 2023"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3543873.3587652","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3543873.3587652","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T23:48:23Z","timestamp":1755820103000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3543873.3587652"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,4,30]]},"references-count":45,"alternative-id":["10.1145\/3543873.3587652","10.1145\/3543873"],"URL":"https:\/\/doi.org\/10.1145\/3543873.3587652","relation":{},"subject":[],"published":{"date-parts":[[2023,4,30]]},"assertion":[{"value":"2023-04-30","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}