Verb Mirage: Unveiling and Assessing Verb Concept Hallucinations in Multimodal Large Language Models

arXiv — cs.CVThursday, November 20, 2025 at 5:00:00 AM
  • The study highlights the issue of verb hallucination in Multimodal Large Language Models (MLLMs), marking a significant step in understanding their limitations.
  • Addressing verb hallucination is crucial for improving the reliability of MLLMs, as verbs are essential for interpreting human actions and interactions.
  • This research contributes to ongoing discussions about the robustness of MLLMs, emphasizing the need for comprehensive evaluation methods that address both object and verb hallucinations.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended apps based on your readingExplore all apps
Continue Readings
Where Does Vision Meet Language? Understanding and Refining Visual Fusion in MLLMs via Contrastive Attention
PositiveArtificial Intelligence
A recent study has explored the integration of visual and textual information in Multimodal Large Language Models (MLLMs), revealing that visual-text fusion occurs at specific layers within these models rather than uniformly across the network. The research highlights a late-stage
Incentivizing Cardiologist-Like Reasoning in MLLMs for Interpretable Echocardiographic Diagnosis
PositiveArtificial Intelligence
A novel approach has been proposed to enhance echocardiographic diagnosis through the integration of a Cardiac Reasoning Template (CRT) and CardiacMind, aimed at improving the reasoning capabilities of multimodal large language models (MLLMs). This method addresses the challenges faced by existing models in capturing the relationship between quantitative measurements and clinical manifestations in cardiac screening.
Ground What You See: Hallucination-Resistant MLLMs via Caption Feedback, Diversity-Aware Sampling, and Conflict Regularization
PositiveArtificial Intelligence
A recent study has introduced a framework aimed at mitigating hallucination issues in Multimodal Large Language Models (MLLMs) during Reinforcement Learning (RL) optimization. The research identifies key factors contributing to hallucinations, including over-reliance on visual reasoning and insufficient exploration diversity. The proposed framework incorporates modules for caption feedback, diversity-aware sampling, and conflict regularization to enhance model reliability.
KidVis: Do Multimodal Large Language Models Possess the Visual Perceptual Capabilities of a 6-Year-Old?
NeutralArtificial Intelligence
A new benchmark called KidVis has been introduced to evaluate the visual perceptual capabilities of Multimodal Large Language Models (MLLMs), specifically assessing their performance against that of 6-7 year old children across six atomic visual capabilities. The results reveal a significant performance gap, with human children scoring an average of 95.32 compared to GPT-5's score of 67.33.
UR-Bench: A Benchmark for Multi-Hop Reasoning over Ultra-High-Resolution Images
NeutralArtificial Intelligence
The introduction of the Ultra-high-resolution Reasoning Benchmark (UR-Bench) aims to evaluate the reasoning capabilities of multimodal large language models (MLLMs) specifically on ultra-high-resolution images, which have been largely unexplored in existing visual question answering benchmarks. This benchmark features two main categories, Humanistic Scenes and Natural Scenes, with images ranging from hundreds of megapixels to gigapixels, accompanied by structured questions.
M3CoTBench: Benchmark Chain-of-Thought of MLLMs in Medical Image Understanding
PositiveArtificial Intelligence
The introduction of M3CoTBench marks a significant advancement in the evaluation of Chain-of-Thought (CoT) reasoning within Multimodal Large Language Models (MLLMs) specifically for medical image understanding, addressing the limitations of existing benchmarks that focus solely on final answers without considering the reasoning process.
PRISM: Self-Pruning Intrinsic Selection Method for Training-Free Multimodal Data Selection
PositiveArtificial Intelligence
A new method called PRISM has been introduced to optimize the selection of training data for Multimodal Large Language Models (MLLMs), addressing the redundancy in rapidly growing datasets that increases computational costs. This self-pruning intrinsic selection method aims to enhance efficiency without the need for extensive training or proxy-based inference techniques.
MoHoBench: Assessing Honesty of Multimodal Large Language Models via Unanswerable Visual Questions
NeutralArtificial Intelligence
A recent study introduced MoHoBench, a benchmark designed to assess the honesty of Multimodal Large Language Models (MLLMs) when confronted with unanswerable visual questions. This research highlights the need for a systematic evaluation of MLLMs' response behaviors, as their trustworthiness in generating content remains underexplored.

Ready to build your own newsroom?

Subscribe to unlock a personalised feed, podcasts, newsletters, and notifications tailored to the topics you actually care about