Home
Uni-Logo
 

Selective Concept Bottleneck Models Without Predefined Concepts

Transactions on Machine Learning Research (TMLR), May 2025
Abstract: Concept-based models like Concept Bottleneck Models (CBMs) have garnered significant interest for improving model interpretability by first predicting human-understandable concepts before mapping them to the output classes. Early approaches required costly concept annotations. To alleviate this, recent methods utilized large language models to automatically generate class-specific concept descriptions and learned mappings from a pretrained black-box model’s raw features to these concepts using vision-language models. However, these approaches assume prior knowledge of which concepts the black-box model has learned. In this work, we discover the concepts encoded by the model through unsupervised concept discovery techniques instead. We further leverage a simple input-dependent concept selection mechanism that dynamically retains a sparse set of relevant concepts of each input, enhancing both sparsity and interpretability. Our approach not only improves downstream performance, but also needs significantly fewer concepts for accurate classification. Lastly, we show how large vision-language models can guide the editing of our models' weights to correct model errors.
Paper Publisher's link Downloads

Images and movies

 

BibTex reference

@Article{SAB25,
  author       = "S. Schrodi and J. Schur and M. Argus and T. Brox",
  title        = "Selective Concept Bottleneck Models Without Predefined Concepts",
  journal      = "Transactions on Machine Learning Research (TMLR)",
  month        = "May",
  year         = "2025",
  url          = "http://lmbweb.informatik.uni-freiburg.de/Publications/2025/SAB25"
}

Other publications in the database