Official implementation of Vision-Amplified Semantic Entropy for Hallucination Detection in Medical Visual Question Answering (MICCAI 2025).
This repo contains the official implementation of our paper: Vision-Amplified Semantic Entropy for Hallucination Detection in Medical Visual Question Answering, which proposed a hallucination detection method for medical VQA.
We suggest using virtual env to configure the experimental environment. Parts of the Semantic Entropy computation code are adapted from https://github.com/lorenzkuhn/semantic_uncertainty.
- Clone this repo:
git clone https://github.com/Merrical/VASE.git-
Create an environment 'env_main' for medical MLLMs ( MedGemma-4b-it used in this project).
-
Create another dedicated environment 'env_green' for the GREEN model.
-
Activate env_main, generate answers for open-ended test VQA samples of the RAD-VQA dataset with MedGemma-4b-it (temperature = 0.1), and compute hallucination scores (RadFlag, Semantic Entropy, and our VASE).
python main_hall_det.py- Activate env_green, and obtain GREEN scores (hallucination labels) for all test samples.
CUDA_VISIBLE_DEVICES=0 torchrun --nproc_per_node=1 green_eval.py- Evaluate hallucination detection performance (AUC, AUG) of RadFlag, SE, and VASE.
python hall_det_eval.py@inproceedings{Liao2025VASE,
title={Vision-Amplified Semantic Entropy for Hallucination Detection in Medical Visual Question Answering},
author={Liao, Zehui and Hu, Shishuai and Zou, Ke and Fu, Huazhu and Zhen, Liangli and Xia, Yong},
booktitle={International Conference on Medical Image Computing and Computer-Assisted Intervention},
year={2025},
organization={Springer}
}If you have any questions, please contact us ( merrical@mail.nwpu.edu.cn ).
