Instella: Fully Open Language Models with Stellar Performance

arXiv — cs.CLMonday, November 17, 2025 at 5:00:00 AM
  • Instella has launched a new series of fully open language models that excel in performance, trained on publicly available data and utilizing advanced AMD GPUs. These models, including Instella
  • The introduction of fully open models like Instella is significant as it promotes transparency and reproducibility in AI research, potentially influencing the development of future AI technologies and encouraging broader access to high
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended Readings
The Anatomy of a Triton Attention Kernel
PositiveArtificial Intelligence
The article discusses the development of a portable and efficient LLM inference platform using a state-of-the-art paged attention kernel built on the Triton language. This kernel is designed to work across various hardware architectures, specifically NVIDIA and AMD GPUs, without the need for extensive low-level tuning. The authors detail their approach, algorithmic improvements, and the necessary integrations to enhance performance, achieving a significant increase from 19.7% to 105% of the state-of-the-art efficiency.