foundation-models

Align Once to Explain: Feature Alignment for Scalable B-cosification of Foundational Vision Transformers

Foundational vision models have become the de facto standard for many vision tasks due to their strong performance. However, they are notoriously opaque and remain hard to interpret. We present ALOE (ALign Once to Explain), a one-time, label-free …

CFM: Language-aligned Concept Foundation Model for Vision

Language-aligned vision foundation models perform strongly across diverse downstream tasks. Yet, their learned representations remain opaque, making interpreting their decision-making difficult. Recent work decompose these representations into …