Climbing the Ladder of Interpretability with Counterfactual Concept Bottleneck Models
Submitted to IJCAI 2024, 2023
In this paper, we introduce CounterFactual Concept Bottleneck Models (CF-CBMs), a class of models designed to efficiently address three fundamental questions all at once without the need to run post-hoc searches: predict class labels to solve a given classification task (the “What?”), explain task predictions (the “Why?”), and imagine alternative scenarios that could result in different predictions (the “What if?”).
Recommended citation: Gabriele Dominici, Pietro Barbiero, Francesco Giannini, Martin Gjoreski, Marc Langheinrich & Giuseppe Marra. (2024). Climbing the Ladder of Interpretability with Counterfactual Concept Bottleneck Models