Studies with impossible languages falsify LMs as models of human language

arXiv — cs.CLMonday, November 17, 2025 at 5:00:00 AM
A study published on arXiv examines the learning capabilities of infants and language models (LMs) regarding attested versus impossible languages. The research indicates that both groups find attested languages easier to learn than those with unnatural structures. However, the findings reveal that LMs can learn many impossible languages as effectively as attested ones. The study suggests that the complexity of these languages, rather than their impossibility, contributes to the challenges faced by LMs, which lack the human inductive biases essential for language acquisition.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended Readings
On the Entropy Calibration of Language Models
NeutralArtificial Intelligence
The paper examines entropy calibration in language models, focusing on whether their entropy aligns with log loss on human text. Previous studies indicated that as text generation lengthens, entropy increases while text quality declines, highlighting a fundamental issue in autoregressive models. The authors investigate whether miscalibration can improve with scale and if calibration without tradeoffs is theoretically feasible, analyzing the scaling behavior concerning dataset size and power law exponents.
Are language models rational? The case of coherence norms and belief revision
NeutralArtificial Intelligence
The paper titled 'Are language models rational? The case of coherence norms and belief revision' explores the application of rationality norms, specifically coherence norms, to language models. It distinguishes between logical coherence norms and those related to the strength of belief. The authors introduce the Minimal Assent Connection (MAC), a new framework for understanding credence in language models based on internal token probabilities. The findings suggest that while some language models adhere to these rational norms, others do not, raising important questions about AI behavior and safety.