Do Concept Bottleneck Models Learn as Intended?
View / Open Files
Publication Date
2021-05-10Journal Title
CoRR
Conference Name
ICLR-21 Workshop on Responsible AI
Type
Article
This Version
AM
Metadata
Show full item recordCitation
Margeloiu, A., Ashman, M., Bhatt, U., Chen, Y., Jamnik, M., & Weller, A. (2021). Do Concept Bottleneck Models Learn as Intended?. CoRR https://doi.org/10.17863/CAM.80941
Abstract
Concept bottleneck models map from raw inputs to concepts, and then from
concepts to targets. Such models aim to incorporate pre-specified, high-level
concepts into the learning procedure, and have been motivated to meet three
desiderata: interpretability, predictability, and intervenability. However, we
find that concept bottleneck models struggle to meet these goals. Using post
hoc interpretability methods, we demonstrate that concepts do not correspond to
anything semantically meaningful in input space, thus calling into question the
usefulness of concept bottleneck models in their current form.
Keywords
cs.LG, cs.LG, cs.AI
Identifiers
This record's DOI: https://doi.org/10.17863/CAM.80941
This record's URL: https://www.repository.cam.ac.uk/handle/1810/333521
Statistics
Total file downloads (since January 2020). For more information on metrics see the
IRUS guide.
Recommended or similar items
The current recommendation prototype on the Apollo Repository will be turned off on 03 February 2023. Although the pilot has been fruitful for both parties, the service provider IKVA is focusing on horizon scanning products and so the recommender service can no longer be supported. We recognise the importance of recommender services in supporting research discovery and are evaluating offerings from other service providers. If you would like to offer feedback on this decision please contact us on: support@repository.cam.ac.uk