Quantification and object perception in Multimodal Large Language Models deviate from human linguistic cognition

arXiv — cs.CLWednesday, November 12, 2025 at 5:00:00 AM
The study published on arXiv investigates the challenges faced by Multimodal Large Language Models (MLLMs) in understanding quantification, a complex linguistic phenomenon. It reveals that MLLMs exhibit clear differences from human cognition in representing quantification, particularly regarding the ordering of quantifiers and biases in numerical perception. By examining these discrepancies, the research aims to enhance our understanding of MLLMs as semantic and pragmatic agents. This exploration is vital for advancing AI language models, as it highlights the need for improvements in their architecture to better align with human linguistic capabilities.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended apps based on your readingExplore all apps

Ready to build your own newsroom?

Subscribe to unlock a personalised feed, podcasts, newsletters, and notifications tailored to the topics you actually care about