{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T04:10:11Z","timestamp":1760155811928,"version":"build-2065373602"},"publisher-location":"Cham","reference-count":25,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783032083166","type":"print"},{"value":"9783032083173","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,10,12]],"date-time":"2025-10-12T00:00:00Z","timestamp":1760227200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2025,10,12]],"date-time":"2025-10-12T00:00:00Z","timestamp":1760227200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"abstract":"<jats:title>Abstract<\/jats:title>\n          <jats:p>Concept-based eXplainable AI (C-XAI) is a rapidly growing research field that enhances AI model interpretability by leveraging intermediate, human-understandable concepts. This approach not only enhances model transparency but also enables human intervention, allowing users to interact with these concepts to refine and improve the model\u2019s performance. Concept Bottleneck Models (CBMs) explicitly predict concepts before making final decisions, enabling interventions to correct misclassified concepts. While CBMs remain effective in Out-Of-Distribution (OOD) settings with intervention, they struggle to match the performance of black-box models. Concept Embedding Models (CEMs) address this by learning concept embeddings from both concept predictions and input data, enhancing In-Distribution (ID) accuracy but reducing the effectiveness of interventions, especially in OOD scenarios. In this work, we propose the Variational Concept Embedding Model (V-CEM), which leverages variational inference to improve intervention responsiveness in CEMs. We evaluated our model on various textual and visual datasets in terms of ID performance, intervention responsiveness in both ID and OOD settings, and Concept Representation Cohesiveness (CRC), a metric we propose to assess the quality of the concept embedding representations. The results demonstrate that V-CEM retains CEM-level ID performance while achieving intervention effectiveness similar to CBM in OOD settings, effectively reducing the gap between interpretability (intervention) and generalization (performance).<\/jats:p>","DOI":"10.1007\/978-3-032-08317-3_3","type":"book-chapter","created":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T03:37:10Z","timestamp":1760153830000},"page":"48-67","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["V-CEM: Bridging Performance and\u00a0Intervenability in\u00a0Concept-Based Models"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0008-3501-963X","authenticated-orcid":false,"given":"Francesco","family":"De Santis","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6799-1043","authenticated-orcid":false,"given":"Gabriele","family":"Ciravegna","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3772-4817","authenticated-orcid":false,"given":"Philippe","family":"Bich","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6987-2064","authenticated-orcid":false,"given":"Danilo","family":"Giordano","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9039-6226","authenticated-orcid":false,"given":"Tania","family":"Cerquitelli","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,10,12]]},"reference":[{"key":"3_CR1","first-page":"17582","volume":"35","author":"ED Abraham","year":"2022","unstructured":"Abraham, E.D., et al.: Cebab: estimating the causal effects of real-world concepts on nlp model behavior. Adv. Neural. Inf. Process. Syst. 35, 17582\u201317596 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"3_CR2","unstructured":"Alvarez\u00a0Melis, D., Jaakkola, T.: Towards robust interpretability with self-explaining neural networks. In: Advances in Neural Information Processing Systems, vol. 31 (2018)"},{"key":"3_CR3","unstructured":"Bai, H., Zhang, J., Nowak, R.: Aha: Human-assisted out-of-distribution generalization and detection. arXiv preprint arXiv:2410.08000 (2024)"},{"key":"3_CR4","doi-asserted-by":"publisher","first-page":"103822","DOI":"10.1016\/j.artint.2022.103822","volume":"314","author":"G Ciravegna","year":"2023","unstructured":"Ciravegna, G., et al.: Logic explained networks. Artif. Intell. 314, 103822 (2023)","journal-title":"Artif. Intell."},{"key":"3_CR5","unstructured":"De\u00a0Felice, G., et al.: Causally reliable concept bottleneck models. arXiv preprint arXiv:2503.04363 (2025)"},{"key":"3_CR6","unstructured":"Dominici, G., et al.: Causal concept graph models: beyond causal opacity in deep learning. arXiv preprint arXiv:2405.16507 (2024)"},{"key":"3_CR7","unstructured":"Gulrajani, I., Lopez-Paz, D.: In search of lost domain generalization. arXiv preprint arXiv:2007.01434 (2020)"},{"key":"3_CR8","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"3_CR9","unstructured":"Hendrycks, D., Mazeika, M., Dietterich, T.: Deep anomaly detection with outlier exposure. arXiv preprint arXiv:1812.04606 (2018)"},{"key":"3_CR10","unstructured":"Kim, B., Wattenberg, M., Gilmer, J., Cai, C., Wexler, J., Viegas, F., et\u00a0al.: Interpretability beyond feature attribution: quantitative testing with concept activation vectors (tcav). In: International Conference on Machine Learning, pp. 2668\u20132677. PMLR (2018)"},{"key":"3_CR11","unstructured":"Kim, E., Jung, D., Park, S., Kim, S., Yoon, S.: Probabilistic concept bottleneck models. In: International Conference on Machine Learning, pp. 16521\u201316540 (2023)"},{"key":"3_CR12","unstructured":"Kingma, D.P.: Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114 (2013)"},{"key":"3_CR13","unstructured":"Koh, P.W., et al.: Concept bottleneck models. In: International Conference on Machine Learning, pp. 5338\u20135348. PMLR (2020)"},{"issue":"11","key":"3_CR14","doi-asserted-by":"publisher","first-page":"2278","DOI":"10.1109\/5.726791","volume":"86","author":"Y Lecun","year":"1998","unstructured":"Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition. Proc. IEEE 86(11), 2278\u20132324 (1998). https:\/\/doi.org\/10.1109\/5.726791","journal-title":"Proc. IEEE"},{"key":"3_CR15","unstructured":"Liang, S., Li, Y.: Enhancing the reliability of out-of-distribution image detection in neural networks. arXiv preprint arXiv:1706.02690 (2017)"},{"key":"3_CR16","doi-asserted-by":"crossref","unstructured":"Liu, Z., Luo, P., Wang, X., Tang, X.: Deep learning face attributes in the wild. In: Proceedings of International Conference on Computer Vision (ICCV), December 2015","DOI":"10.1109\/ICCV.2015.425"},{"key":"3_CR17","unstructured":"Louizos, C., Welling, M., Kingma, D.P.: Learning sparse neural networks through $$ l_0 $$ regularization. arXiv preprint arXiv:1712.01312 (2017)"},{"key":"3_CR18","unstructured":"Maas, A.L., Daly, R.E., Pham, P.T., Huang, D., Ng, A.Y., Potts, C.: Learning word vectors for sentiment analysis. In: Lin, D., Matsumoto, Y., Mihalcea, R. (eds.) Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, pp. 142\u2013150. ACL, Portland, Oregon, USA, June 2011, https:\/\/aclanthology.org\/P11-1015\/"},{"key":"3_CR19","unstructured":"Mahinpei, A., Clark, J., Lage, I., Doshi-Velez, F., Pan, W.: Promises and pitfalls of black-box concept learning models. arXiv preprint arXiv:2106.13314 (2021)"},{"key":"3_CR20","first-page":"21212","volume":"35","author":"E Marconato","year":"2022","unstructured":"Marconato, E., Passerini, A., Teso, S.: Glancenets: interpretable, leak-proof concept-based models. Adv. Neural. Inf. Process. Syst. 35, 21212\u201321227 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"3_CR21","unstructured":"Poeta, E., Ciravegna, G., Pastor, E., Cerquitelli, T., Baralis, E.: Concept-based explainable artificial intelligence: a survey. arXiv preprint arXiv:2312.12936 (2023)"},{"issue":"5","key":"3_CR22","doi-asserted-by":"publisher","first-page":"206","DOI":"10.1038\/s42256-019-0048-x","volume":"1","author":"C Rudin","year":"2019","unstructured":"Rudin, C.: Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 1(5), 206\u2013215 (2019)","journal-title":"Nat. Mach. Intell."},{"key":"3_CR23","unstructured":"Sanh, V.: Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108 (2019)"},{"key":"3_CR24","unstructured":"Vishwakarma, H., Lin, H., Vinayak, R.: Human-in-the-loop out-of-distribution detection with false positive rate control. In: NeurIPS Workshop on Adaptive Experimental Design and Active Learning in the Real World (2023)"},{"key":"3_CR25","unstructured":"Zarlenga, M.E., et\u00a0al.: Concept embedding models. In: NeurIPS 2022-36th Conference on Neural Information Processing Systems (2022)"}],"container-title":["Communications in Computer and Information Science","Explainable Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-032-08317-3_3","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T03:37:16Z","timestamp":1760153836000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-032-08317-3_3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,12]]},"ISBN":["9783032083166","9783032083173"],"references-count":25,"URL":"https:\/\/doi.org\/10.1007\/978-3-032-08317-3_3","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"value":"1865-0929","type":"print"},{"value":"1865-0937","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,10,12]]},"assertion":[{"value":"12 October 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"xAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"World Conference on Explainable Artificial Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Istanbul","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"T\u00fcrkiye","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"9 July 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"11 July 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"3","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"xai2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/xaiworldconference.com\/2025\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}