MoHoBench: Assessing Honesty of Multimodal Large Language Models via Unanswerable Visual Questions
NeutralArtificial Intelligence
- A recent study introduced MoHoBench, a benchmark designed to assess the honesty of Multimodal Large Language Models (MLLMs) when confronted with unanswerable visual questions. This research highlights the need for a systematic evaluation of MLLMs' response behaviors, as their trustworthiness in generating content remains underexplored.
- The development of MoHoBench is significant as it provides a structured framework to evaluate the honesty of 28 popular MLLMs, revealing that many models struggle to provide reliable answers to visually unanswerable questions.
- This initiative is part of a broader effort to enhance the reliability and safety of MLLMs, addressing concerns over their performance in various tasks, including deception detection and visual interpretation, which have shown limitations in previous studies.
— via World Pulse Now AI Editorial System
