PlantBiMoE: A Bidirectional Foundation Model with SparseMoE for Plant Genomes

arXiv — cs.LGTuesday, December 9, 2025 at 5:00:00 AM
  • A new plant genome language model named PlantBiMoE has been introduced, which integrates a bidirectional Mamba and a Sparse Mixture-of-Experts (SparseMoE) framework. This model aims to overcome the limitations of previous models like AgroNT and PDLLMs by effectively capturing structural dependencies in DNA strands while reducing the number of active parameters for improved computational efficiency.
  • The development of PlantBiMoE is significant as it enhances the ability to analyze plant genomes, potentially leading to advancements in computational biology and agricultural research. Its efficiency could facilitate more extensive genomic studies, benefiting researchers and institutions focused on plant genetics.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended apps based on your readingExplore all apps
Continue Readings
TinyViM: Frequency Decoupling for Tiny Hybrid Vision Mamba
PositiveArtificial Intelligence
A new study introduces TinyViM, a model that enhances the Mamba architecture by decoupling features based on frequency, allowing for improved performance in computer vision tasks such as image classification and semantic segmentation. This innovation addresses the limitations of existing lightweight Mamba-based models that have struggled to compete with Convolution and Transformer methods.
TextMamba: Scene Text Detector with Mamba
PositiveArtificial Intelligence
A novel scene text detector named TextMamba has been developed, leveraging the Mamba state space model to enhance long-range dependency modeling in text detection. This approach integrates a selection mechanism with attention layers, addressing limitations in traditional Transformer-based methods that often overlook critical information in lengthy sequences.
JambaTalk: Speech-Driven 3D Talking Head Generation Based on Hybrid Transformer-Mamba Model
PositiveArtificial Intelligence
JambaTalk has been introduced as a hybrid Transformer-Mamba model aimed at enhancing the generation of 3D talking heads, focusing on improving lip-sync, facial expressions, and head poses in animated videos. This model addresses the limitations of traditional Transformers by utilizing a Structured State Space Model (SSM) to manage long sequences effectively.
Always Keep Your Promises: DynamicLRP, A Model-Agnostic Solution To Layer-Wise Relevance Propagation
PositiveArtificial Intelligence
DynamicLRP has been introduced as a model-agnostic framework for Layer-wise Relevance Propagation (LRP), allowing for attribution in neural networks without the need for architecture-specific modifications. This innovation operates at the tensor operation level, utilizing a Promise System for deferred activation resolution, thereby enhancing the generality and sustainability of LRP implementations.