Abstract
The Draft and Refine (DnR) agent framework uses a question-conditioned utilization metric to improve visual grounding in large vision-language models, reducing hallucinations and increasing accuracy.
While recent Large Vision-Language Models (LVLMs) exhibit strong multimodal reasoning abilities, they often produce ungrounded or hallucinated responses because they rely too heavily on linguistic priors instead of visual evidence. This limitation highlights the absence of a quantitative measure of how much these models actually use visual information during reasoning. We propose Draft and Refine (DnR), an agent framework driven by a question-conditioned utilization metric. The metric quantifies the model's reliance on visual evidence by first constructing a query-conditioned relevance map to localize question-specific cues and then measuring dependence through relevance-guided probabilistic masking. Guided by this metric, the DnR agent refines its initial draft using targeted feedback from external visual experts. Each expert's output (such as boxes or masks) is rendered as visual cues on the image, and the model is re-queried to select the response that yields the largest improvement in utilization. This process strengthens visual grounding without retraining or architectural changes. Experiments across VQA and captioning benchmarks show consistent accuracy gains and reduced hallucination, demonstrating that measuring visual utilization provides a principled path toward more interpretable and evidence-driven multimodal agent systems.
Community
This paper introduces Draft-and-Refine (DnR), a framework that measures question-conditioned visual utilization in LVLMs and refines answers using external visual experts. The pipeline mitigates hallucination without retraining and strengthens visual grounding through relevance-driven feedback.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Small Drafts, Big Verdict: Information-Intensive Visual Reasoning via Speculation (2025)
- Diagnosing Visual Reasoning: Challenges, Insights, and a Path Forward (2025)
- VisRAG 2.0: Evidence-Guided Multi-Image Reasoning in Visual Retrieval-Augmented Generation (2025)
- Token-Level Inference-Time Alignment for Vision-Language Models (2025)
- Where MLLMs Attend and What They Rely On: Explaining Autoregressive Token Generation (2025)
- Taming a Retrieval Framework to Read Images in Humanlike Manner for Augmenting Generation of MLLMs (2025)
- Mitigating Hallucination in Multimodal Reasoning via Functional Attention Control (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper