Bolmo’s architecture unlocks efficient byte‑level LM training without sacrificing quality

VentureBeat — AIMonday, December 15, 2025 at 5:00:00 AM
Bolmo’s architecture unlocks efficient byte‑level LM training without sacrificing quality
  • The Allen Institute of AI (Ai2) has launched Bolmo, a new family of byte-level language models designed to operate without predefined vocabularies or tokenizers, thus enhancing efficiency in multilingual model training. The models, Bolmo 7B and Bolmo 1B, are noted for their competitive performance against existing byte-level and character-based models.
  • This development positions Ai2 as a leader in the field of language modeling, particularly for enterprises seeking robust, tokenizer-free solutions that can handle noisy or low-resource text effectively. Bolmo's architecture aims to reduce brittleness while maintaining quality.
  • The introduction of Bolmo reflects a growing trend in AI towards models that prioritize efficiency and adaptability, especially in high-stakes applications like finance and medicine. This shift aligns with broader advancements in AI, such as the Olmo 3 family, which emphasizes customization and transparency, indicating a significant evolution in how AI systems are developed and deployed.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended apps based on your readingExplore all apps
Continue Readings
Allen Institute for AI launches Bolmo 7B and Bolmo 1B, claiming they are "the first fully open byte-level language models", built on its Olmo 3 models (Emilia David/VentureBeat)
PositiveArtificial Intelligence
The Allen Institute for AI has launched Bolmo 7B and Bolmo 1B, which it claims are the first fully open byte-level language models, built on its Olmo 3 models. These models aim to provide tokenizer-free multilingual capabilities, addressing the need for more robust language processing in noisy or low-resource text environments.
How Tech’s Biggest Companies Are Offloading the Risks of the A.I. Boom
NeutralArtificial Intelligence
Tech giants, including Meta, are investing heavily in new data centers to support artificial intelligence initiatives, with a significant project underway in Eagle Mountain, Utah. These facilities can cost tens of billions to build, prompting companies to seek ways to mitigate financial risks associated with these investments.
The case against predicting tokens to build AGI
NeutralArtificial Intelligence
In a recent debate, Yann LeCun, Meta's Chief AI Scientist, argued against the efficacy of Large Language Models (LLMs) in achieving human-like intelligence, stating that their predictive nature represents a dead end in the quest for Artificial General Intelligence (AGI). This discussion took place alongside DeepMind researcher Adam Brown, highlighting fundamental concerns regarding the limitations of current AI models.
Meta-Statistical Learning: Supervised Learning of Statistical Estimators
NeutralArtificial Intelligence
A new framework called meta-statistical learning has been introduced, which utilizes supervised learning to optimize the design of statistical estimators. This approach addresses the challenges of crafting estimators with desirable properties, such as bias and variance, by employing permutation-invariant neural networks like Set Transformers to analyze entire datasets.

Ready to build your own newsroom?

Subscribe to unlock a personalised feed, podcasts, newsletters, and notifications tailored to the topics you actually care about