Abstract
Concept Bottleneck Models (CBMs) enhance the interpretability of neural networks by basing predictions on human-understandable concepts. However, current CBMs typically rely on concept sets extracted from large language models or extensive image corpora, limiting their effectiveness in data-sparse scenarios. We propose Data-efficient CBMs (DCBMs), which reduce the need for large sample sizes during concept generation while preserving interpretability. DCBMs define concepts as image regions detected by segmentation or detection foundation models, allowing each image to generate multiple concepts across different granularities. Exclusively containing dataset-specific concepts, DCBMs are well suited for fine-grained classification and out-of-distribution tasks. Attribution analysis using Grad-CAM demonstrates that DCBMs deliver visual concepts that can be localized in test images. By leveraging dataset-specific concepts instead of predefined or general ones, DCBMs enhance adaptability to new domains.
Core Technologies
- 🔍 Segmentation: SAM, SAM2, GroundingDINO, Mask R-CNN
- 💡 Backbone: CLIP ViT-L/14, CLIP ViT-B/16, ResNet50
- 🎯 Tasks: Fine-Grained & OOD Classification
Introduction
Deep neural networks achieve state-of-the-art performance but remain opaque, hindering trust in critical applications. Explainable AI (XAI) seeks transparency; Concept Bottleneck Models predict through human-understandable concepts, weighting these to form decisions. Early CBMs use manually defined concepts, while recent works leverage text-aligned sets from large corpora, which can misalign in fine-grained or low-data regimes. DCBMs define concepts as visual regions extracted by foundation models, generating multi-granularity, dataset-specific concepts robust to domain shifts.
Contributions
- A data-efficient CBM framework requiring as few as 50 labeled images per class.
- Dataset-specific concept generation via foundation segmentation and detection models.
- Comprehensive evaluation demonstrating interpretability and robustness across benchmarks.
Method
- Proposal: Sample n images per class; apply SAM2/GroundingDINO for region proposals and filter by area.
- Clustering: Encode proposals, cluster embeddings (k-means, k=2048) into visual concepts.
- Bottleneck Training: Project image embeddings onto centroids; train sparse linear classifier with L1.
- Naming: (Optional) Match centroids to CLIP text embeddings for labels; prune spurious concepts.

Experiments
DCBMs achieve competitive performance while requiring significantly less data for concept generation. Our experiments demonstrate:
Data Efficiency
reduction in concept generation data
Performance Gap
vs. CLIP linear probe
OOD Robustness
error increase on ImageNet-R
Key Findings
- ✓ Achieve comparable accuracy with only 50 images per class for concept generation
- ✓ Maintain performance across ImageNet, Places365, CUB, and CIFAR benchmarks
- ✓ Superior out-of-distribution robustness compared to baseline methods
- ✓ Dataset-specific concepts improve fine-grained classification accuracy

Conclusion
DCBMs offer a visually grounded, data-efficient concept bottleneck approach with minimal samples and strong interpretability. Future directions include extending to regression and refining concept naming.