Title

Interpretable Zero-shot Classification with Concept Bottleneck Models

Abstract

Abstract

Concept Bottleneck Models (CBMs) have shown promise in improving the interpretability of neural networks by generating intermediate, human-understandable concepts before making predictions. Existing approaches focus on training downstream classifiers jointly with CBMs to produce interpretable classification results, but they often struggle to generalize to unseen classes. In this work, we propose Zero-shot Concept Bottleneck Models (ZS-CBMs), which enable interpretable open-vocabulary recognition. Our methods leverage pre-trained vision-language models to identify attributes related to given image and text inputs and project these inputs into a shared concept space. This allows for zero-shot classification of unseen classes while providing interpretable justifications for model decisions. The accuracy and interpretability of the proposed approaches are evaluated on widely used datasets, including CIFAR-100, CUB-200-2011, CC3M, and ImageNet-100.

Supervisor(s)

Supervisor(s)

ONAT OZDEMIR

Date and Location

Date and Location

2025-08-05 11:00:00

Category

Category

MSc_Thesis