CVPR 2026
While large vision–language models (LVLMs) achieve strong performance on multimodal tasks, they frequently generate hallucinations—unfaithful outputs misaligned with the visual input. To address this issue, we introduce CIPHER (Counterfactual Image Perturbations for Hallucination Extraction and Removal), a training-free method that suppresses vision-induced hallucinations via lightweight feature-level correction. Unlike prior training-free approaches that primarily focus on text-induced hallucinations, CIPHER explicitly targets hallucinations arising from the visual modality.
CIPHER operates in two phases. In the offline phase, we construct OHC-25K (Object-Hallucinated Counterfactuals, 25,000 samples), a counterfactual dataset consisting of diffusion-edited images that intentionally contradict the original ground-truth captions. We pair these edited images with the unchanged ground-truth captions and process them through an LVLM to extract hallucination-related representations. Contrasting these representations with those from authentic (image, caption) pairs reveals structured, systematic shifts spanning a low-rank subspace characterizing vision-induced hallucination.
In the inference phase, CIPHER suppresses hallucinations by projecting intermediate hidden states away from this subspace. Experiments across multiple benchmarks show that CIPHER significantly reduces hallucination rates while preserving task performance.
CIPHER estimates feature directions triggered by hallucination-inducing visual cues and suppresses these components during inference. The method operates in two stages: (1) an offline phase, where we construct a counterfactual dataset (OHC-25K) and estimate the hallucination subspaces from the LVLM’s hidden representations; and (2) an inference phase, where we nullify hallucination-prone components in hidden states during generation by projecting onto the subspace orthogonal to the learned hallucination directions.
| Method | LLaVA-1.5 | MiniGPT-4 | mPLUG-Owl2 | ||||||
|---|---|---|---|---|---|---|---|---|---|
| CHAIRS↓ | CHAIRI↓ | BLEU↑ | CHAIRS↓ | CHAIRI↓ | BLEU↑ | CHAIRS↓ | CHAIRI↓ | BLEU↑ | |
| Greedy | 20.40±2.80 | 7.08±0.33 | 15.72±0.10 | 32.40±2.20 | 12.20±0.42 | 14.57±0.11 | 22.90±0.90 | 8.62±0.11 | 15.01±0.24 |
| Beam Search | 19.50±2.30 | 6.84±0.79 | 15.99±0.14 | 30.10±0.30 | 11.87±0.37 | 15.35±0.24 | 20.30±0.70 | 7.62±0.19 | 15.43±0.05 |
| DoLa (ICLR'24) | 20.20±2.80 | 6.75±0.54 | 15.68±0.10 | 31.90±3.30 | 12.15±0.89 | 14.54±0.12 | 22.40±1.80 | 8.36±0.04 | 15.13±0.21 |
| OPERA (CVPR'24) | 17.50±0.50 | 6.07±0.32 | 16.02±0.02 | 29.70±0.30 | 11.96±0.29 | 14.82±0.05 | 20.07±2.07 | 7.18±0.39 | 15.41±0.12 |
| VCD (CVPR'24) | 20.30±1.10 | 7.28±0.10 | 14.53±0.01 | 29.00±2.80 | 12.64±1.19 | 14.42±0.01 | 22.80±0.80 | 8.68±0.17 | 15.14±0.13 |
| Woodpecker (SCIS'24) | 23.85±4.62 | 7.50±0.01 | 17.05±0.00 | 28.87±2.20 | 10.20±0.85 | 15.30±0.01 | 26.33±1.98 | 8.43±0.80 | 16.43±0.00 |
| LURE (ICLR'24) | 19.48±2.35 | 6.50±0.38 | 15.97±0.01 | 27.88±2.25 | 10.20±0.85 | 15.03±0.01 | 21.27±0.06 | 7.67±0.16 | 15.65±0.15 |
| HALC (ICML'24) | 16.90±2.10 | 5.72±0.55 | 16.02±0.04 | 25.20±2.00 | 9.42±0.41 | 14.91±0.13 | 18.80±1.20 | 7.00±0.01 | 15.33±0.24 |
| Nullu (CVPR'25) | 15.20±0.60 | 5.30±0.03 | 15.69±0.04 | 21.40±1.00 | 8.99±0.36 | 14.81±0.06 | 15.60±1.20 | 5.77±0.01 | 15.45±0.01 |
| CIPHER (Ours) | 13.05±0.57 | 4.53±0.38 | 15.82±0.25 | 18.48±1.20 | 8.33±0.17 | 15.10±0.43 | 13.60±1.06 | 4.92±0.15 | 16.25±0.47 |
Table 1: CHAIR and BLEU scores across LVLMs; lower CHAIR = less hallucination, higher BLEU = better fluency.
@inproceedings{dastmalchi2026cipher,
title = {Fighting Hallucinations with Counterfactuals: Diffusion-Guided Perturbations for LVLM Hallucination Suppression},
author = {Dastmalchi, Hamidreza and An, Aijun and Cheraghian, Ali and Barzamini, Hamed},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
year = {2026}
}