EMAformer: Enhancing Transformer through Embedding Armor for Time Series Forecasting

arXiv — cs.LGWednesday, November 12, 2025 at 5:00:00 AM
The introduction of EMAformer marks a significant advancement in multivariate time series forecasting, particularly in enhancing the Transformer architecture, which has previously struggled against MLP-based models. By addressing unstable inter-channel relationships, EMAformer incorporates three key inductive biases: global stability, phase sensitivity, and cross-axis specificity. These innovations have led to state-of-the-art performance on 12 real-world benchmarks, with a notable reduction in forecasting errors—2.73% in MSE and 5.15% in MAE. This progress not only highlights the potential of Transformer-based approaches but also underscores the importance of continuous improvement in forecasting models, which are crucial across various sectors. The code for EMAformer is publicly available on GitHub, encouraging further exploration and application of this model in real-world scenarios.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended apps based on your readingExplore all apps
Continue Readings
Softpick: No Attention Sink, No Massive Activations with Rectified Softmax
PositiveArtificial Intelligence
The introduction of softpick, a novel drop-in replacement for softmax in transformer attention mechanisms, addresses issues of attention sink and massive activations, achieving a consistent 0% sink rate in experiments with large models. This advancement allows for the production of hidden states with lower kurtosis and sparser attention maps.

Ready to build your own newsroom?

Subscribe to unlock a personalised feed, podcasts, newsletters, and notifications tailored to the topics you actually care about