[NAACL 2025] From redundancy to relevance: Enhancing explainability in multimodal large language models
conda env create -f environment.yml
conda activate redundancy
python -m pip install -e transformers-4.29.2
retain_grad()
required_grad()=True
The following evaluation requires for MSCOCO 2014 dataset. Please download here and extract it in your data path.
Besides, it needs you to prepare the following checkpoints of 7B base models:
- Download LLaVA-1.5 merged 7B model and specify it at Line 14 of
eval_configs/llava-1.5_eval.yaml
.
@article{zhang2024redundancy,
title={From Redundancy to Relevance: Enhancing Explainability in Multimodal Large Language Models},
author={Zhang, Xiaofeng and Quan, Yihao and Shen, Chen and Yuan, Xiaosong and Yan, Shaotian and Xie, Liang and Wang, Wenxiao and Gu, Chaochen and Tang, Hao and Ye, Jieping},
journal={Annual Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics},
year={2025}
}