AraLingBench A Human-Annotated Benchmark for Evaluating Arabic Linguistic Capabilities of Large Language Models

arXiv — cs.LGWednesday, November 19, 2025 at 5:00:00 AM
  • AraLingBench has been launched as a comprehensive benchmark for assessing the Arabic linguistic capabilities of large language models, featuring 150 expert
  • The benchmark is significant as it addresses the need for a diagnostic tool that can guide the development of more effective Arabic LLMs, ensuring they move beyond mere memorization to achieve genuine comprehension of the language.
  • This development reflects ongoing challenges in the field of AI, where models often excel in surface
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended Readings
ArbESC+: Arabic Enhanced Edit Selection System Combination for Grammatical Error Correction Resolving conflict and improving system combination in Arabic GEC
PositiveArtificial Intelligence
The paper presents ArbESC+, a novel multi-system approach for Grammatical Error Correction (GEC) in Arabic. Given the complexity of Arabic's morphological and syntactic structures, previous models often operated in isolation. ArbESC+ combines multiple models to generate correction proposals, which are evaluated by a classifier that selects the most appropriate corrections. The framework enhances output quality by filtering overlapping corrections and assessing decision reliability, marking a significant advancement in Arabic GEC.
Automatic Fact-checking in English and Telugu
NeutralArtificial Intelligence
The research paper explores the challenge of false information and the effectiveness of large language models (LLMs) in verifying factual claims in English and Telugu. It presents a bilingual dataset and evaluates various approaches for classifying the veracity of claims. The study aims to enhance the efficiency of fact-checking processes, which are often labor-intensive and time-consuming.
Start Small, Think Big: Curriculum-based Relative Policy Optimization for Visual Grounding
PositiveArtificial Intelligence
The article presents a novel training strategy called Curriculum-based Relative Policy Optimization (CuRPO) aimed at improving Visual Grounding tasks. It highlights the limitations of Chain-of-Thought (CoT) prompting, particularly when outputs become lengthy or complex, which can degrade performance. The study reveals that simply increasing dataset size does not guarantee better results due to varying complexities. CuRPO utilizes CoT length and generalized Intersection over Union (gIoU) rewards to structure training data progressively from simpler to more challenging examples, demonstrating ef…
MoHoBench: Assessing Honesty of Multimodal Large Language Models via Unanswerable Visual Questions
NeutralArtificial Intelligence
MoHoBench is a newly developed benchmark aimed at assessing the honesty of Multimodal Large Language Models (MLLMs) when confronted with unanswerable visual questions. Despite advancements in vision-language tasks, MLLMs often produce unreliable content. This study systematically evaluates the honesty of 28 popular MLLMs using a dataset of over 12,000 visual questions, revealing that many models struggle to provide honest responses. The findings highlight the need for improved trustworthiness in AI systems.
Large Language Models and 3D Vision for Intelligent Robotic Perception and Autonomy
PositiveArtificial Intelligence
The integration of Large Language Models (LLMs) with 3D vision is revolutionizing robotic perception and autonomy. This approach enhances robotic sensing technologies, allowing machines to understand and interact with complex environments using natural language and spatial awareness. The review discusses the foundational principles of LLMs and 3D data, examines critical 3D sensing technologies, and highlights advancements in scene understanding, text-to-3D generation, and embodied agents, while addressing the challenges faced in this evolving field.
MedBench v4: A Robust and Scalable Benchmark for Evaluating Chinese Medical Language Models, Multimodal Models, and Intelligent Agents
PositiveArtificial Intelligence
MedBench v4 is a new benchmarking infrastructure designed to evaluate Chinese medical language models, multimodal models, and intelligent agents. It features over 700,000 expert-curated tasks across various specialties, with evaluations conducted by clinicians from more than 500 institutions. The study assessed 15 advanced models, revealing that base LLMs scored an average of 54.1/100, while safety and ethics ratings were notably low at 18.4/100. Multimodal models performed even worse, indicating a need for improved evaluation frameworks in medical AI.
SERL: Self-Examining Reinforcement Learning on Open-Domain
PositiveArtificial Intelligence
Self-Examining Reinforcement Learning (SERL) is a proposed framework that addresses challenges in applying Reinforcement Learning (RL) to open-domain tasks. Traditional methods face issues with subjectivity and reliance on external rewards. SERL innovatively positions large language models (LLMs) as both Actor and Judge, utilizing internal reward mechanisms. It employs Copeland-style pairwise comparisons to enhance the Actor's capabilities and introduces a self-consistency reward to improve the Judge's reliability, aiming to advance RL applications in open domains.
GenRecal: Generation after Recalibration from Large to Small Vision-Language Models
PositiveArtificial Intelligence
Recent advancements in vision-language models (VLMs) have utilized large language models (LLMs) to achieve performance comparable to proprietary systems like GPT-4V. However, deploying these models on resource-constrained devices poses challenges due to high computational requirements. To address this, a new framework called Generation after Recalibration (GenRecal) has been introduced, which distills knowledge from large VLMs into smaller, more efficient models by aligning feature representations across diverse architectures.