Do LLMs produce texts with "human-like" lexical diversity?

arXiv — cs.CLMonday, November 24, 2025 at 5:00:00 AM
  • A recent study has examined the lexical diversity of texts generated by various ChatGPT models, including ChatGPT-3.5, ChatGPT-4, ChatGPT-o4 mini, and ChatGPT-4.5, comparing them to texts written by native and non-native English speakers. The findings indicate significant differences in lexical diversity metrics, suggesting that LLMs do not produce writing that is truly human-like.
  • This research is crucial as it highlights the limitations of current large language models in mimicking human writing styles, which could impact their adoption in fields requiring nuanced language use, such as creative writing and education.
  • The ongoing debate about the capabilities of AI in creative domains continues, with studies indicating that while LLMs have advanced in reasoning and problem-solving, they still fall short of matching the creativity and lexical richness of human authors, raising questions about the future roles of AI in creative industries.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended apps based on your readingExplore all apps
Continue Readings
Why the long interface? AI systems don't 'get' the joke, research reveals
NeutralArtificial Intelligence
A recent study indicates that advanced AI systems like ChatGPT and Gemini simulate an understanding of humor but do not genuinely comprehend jokes. This finding highlights a significant limitation in the capabilities of these AI models, which are often perceived as more intelligent than they are.
Five crucial ways LLMs can endanger your privacy
NegativeArtificial Intelligence
Privacy concerns surrounding large language models (LLMs) like ChatGPT, Anthropic, and Gemini have escalated, as highlighted by a Northeastern University computer science expert. The issues extend beyond the data these algorithms process, raising alarms about user privacy and data security.
Meet the Group Breaking People Out of AI Delusions
NegativeArtificial Intelligence
A group is actively working to help individuals recognize and break free from their delusions related to artificial intelligence, particularly those who have become overly reliant on AI tools like ChatGPT. This phenomenon highlights a growing concern about the psychological impact of AI on users, as some individuals no longer feel the need for human interaction.
A Research Leader Behind ChatGPT’s Mental Health Work Is Leaving OpenAI
NeutralArtificial Intelligence
A key research leader involved in ChatGPT's mental health initiatives is departing from OpenAI, which raises questions about the future direction of AI safety research, particularly in how the chatbot interacts with users in crisis situations. This change comes at a time when OpenAI is expanding its features, including group chats and a free version for educators.
‘Holy S***… I’m Not Going Back to ChatGPT,’ Says Marc Benioff After Using Gemini 3
PositiveArtificial Intelligence
Marc Benioff, CEO of Salesforce, expressed his strong preference for Google's Gemini 3 over OpenAI's ChatGPT, stating, 'Holy S***… I’m Not Going Back to ChatGPT' after experiencing the new AI model. This statement highlights the growing competition between Google and OpenAI in the AI landscape.
How Well Do LLMs Understand Tunisian Arabic?
NegativeArtificial Intelligence
A recent study highlights the limitations of Large Language Models (LLMs) in understanding Tunisian Arabic, also known as Tunizi. This research introduces a new dataset that includes parallel translations in Tunizi, standard Tunisian Arabic, and English, aiming to benchmark LLMs on their comprehension of this low-resource language. The findings indicate that the neglect of such dialects may hinder millions of Tunisians from engaging with AI in their native language.
How Language Directions Align with Token Geometry in Multilingual LLMs
PositiveArtificial Intelligence
A recent study on multilingual large language models (LLMs) reveals that language information is distinctly organized within their internal representation space, particularly showing significant separation in the first transformer block. This comprehensive probing study analyzed six multilingual LLMs across all 268 transformer layers, utilizing both linear and nonlinear probes alongside a new Token-Language Alignment analysis.
MUCH: A Multilingual Claim Hallucination Benchmark
PositiveArtificial Intelligence
A new benchmark named MUCH has been introduced to assess Claim-level Uncertainty Quantification (UQ) in Large Language Models (LLMs). This benchmark includes 4,873 samples in English, French, Spanish, and German, and provides 24 generation logits per token, enhancing the evaluation of UQ methods under realistic conditions.