computer-vision

Good Teachers Explain: Explanation-Enhanced Knowledge Distillation

Knowledge Distillation (KD) has proven effective for compressing large teacher models into smaller student models. While it is well known that student models can achieve similar accuracies as the teachers, it has also been shown that they nonetheless …

Studying How to Efficiently and Effectively Guide Models with Explanations

Despite being highly performant, deep neural networks might base their decisions on features that spuriously correlate with the provided labels, thus hurting generalization. To mitigate this, ‘model guidance’ has recently gained popularity, i.e. the …

Model Guidance

Studying How to Efficiently and Effectively Guide Models with Explanations