LBMamba: Locally Bi-directional Mamba

arXiv — cs.CVThursday, November 13, 2025 at 5:00:00 AM
The recent introduction of LBMamba, a locally bi-directional State Space Model (SSM), represents a pivotal development in the field of artificial intelligence, particularly in computer vision. Traditional Mamba models, while efficient, faced limitations due to their unidirectional nature, which restricted their ability to access future states. LBMamba addresses this by integrating a lightweight backward scan into the forward scan, effectively maintaining computational efficiency without the burden of additional scans. This innovation is further exemplified in LBVim, a backbone that alternates scan directions every two layers, achieving notable improvements in accuracy across various datasets. For instance, LBVim demonstrates a 1.2% increase in top-1 accuracy on ImageNet-1K, a 1.65% improvement in mean Intersection over Union (mIoU) on ADE20K, and enhancements in detection metrics on COCO. These advancements not only highlight the superior performance-throughput trade-off offered by LBM…
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended Readings
ERMoE: Eigen-Reparameterized Mixture-of-Experts for Stable Routing and Interpretable Specialization
PositiveArtificial Intelligence
The article introduces ERMoE, a new Mixture-of-Experts (MoE) architecture designed to enhance model capacity by addressing challenges in routing and expert specialization. ERMoE reparameterizes experts in an orthonormal eigenbasis and utilizes an 'Eigenbasis Score' for routing, which stabilizes expert utilization and improves interpretability. This approach aims to overcome issues of misalignment and load imbalances that have hindered previous MoE architectures.
MADiff: Motion-Aware Mamba Diffusion Models for Hand Trajectory Prediction on Egocentric Videos
PositiveArtificial Intelligence
The article presents MADiff, a novel method for predicting hand trajectories in egocentric videos using diffusion models. This approach aims to enhance the understanding of human intentions and actions, which is crucial for advancements in embodied artificial intelligence. The challenges of capturing high-level human intentions and the effects of camera egomotion interference are addressed, making this method significant for applications in extended reality and robot manipulation.
PrivDFS: Private Inference via Distributed Feature Sharing against Data Reconstruction Attacks
PositiveArtificial Intelligence
The paper introduces PrivDFS, a distributed feature-sharing framework designed for input-private inference in image classification. It addresses vulnerabilities in split inference that allow Data Reconstruction Attacks (DRAs) to reconstruct inputs with high fidelity. By fragmenting the intermediate representation and processing these fragments independently across a majority-honest set of servers, PrivDFS limits the reconstruction capability while maintaining accuracy within 1% of non-private methods.
Out-of-Distribution Detection with Positive and Negative Prompt Supervision Using Large Language Models
PositiveArtificial Intelligence
The paper discusses advancements in out-of-distribution (OOD) detection, focusing on the integration of visual and textual modalities through large language models (LLMs). It introduces a method called Positive and Negative Prompt Supervision, which aims to improve OOD detection by using class-specific prompts that capture inter-class features. This approach addresses the limitations of negative prompts that may include non-ID features, potentially leading to suboptimal outcomes.
OpenUS: A Fully Open-Source Foundation Model for Ultrasound Image Analysis via Self-Adaptive Masked Contrastive Learning
PositiveArtificial Intelligence
OpenUS is a newly proposed open-source foundation model for ultrasound image analysis, addressing the challenges of operator-dependent interpretation and variability in ultrasound imaging. This model utilizes a vision Mamba backbone and introduces a self-adaptive masking framework that enhances pre-training through contrastive learning and masked image modeling. With a dataset comprising 308,000 images from 42 datasets, OpenUS aims to improve the generalizability and efficiency of ultrasound AI models.