Explainable machine learning significantly improves the transparency of deep neural networks (DNN). However, existing work is constrained to explaining the behavior of individual model predictions, and lacks the ability to transfer the explanation across various models and tasks. This limitation results in explaining various tasks being time- and resource-consuming. To address this problem, we develop a DNN-based transferable vision explainer, named TVE, pre-trained on large-scale image datasets, and leverage its transferability to explain various vision models for downstream tasks.
The development of transferable explainers introduces two non-trivial two challenges:
CH1: Without task-specific exposure during the pre-training, how to ensure the universal effectiveness of explainer for various downstream tasks?
CH2: How to adapt the explainer to a specific task without fine-tuning on the task-specific data?
torch
torchvision
pillow
transformers
datasets
accelerate
Run our pre-trained explainer to generate heatmaps for explaining ViT classification models
python3 heatmap_demo.py
The heatmap of local explanation will be