Language Models for Controllable DNA Sequence Design

arXiv — cs.LGWednesday, December 10, 2025 at 5:00:00 AM
  • Researchers have introduced ATGC-Gen, an Automated Transformer Generator designed for controllable DNA sequence design, which generates sequences based on specific biological properties. This model utilizes cross-modal encoding and can operate under various transformer architectures, enhancing its flexibility in training and generation tasks, particularly in promoter and enhancer sequence design.
  • The development of ATGC-Gen is significant as it represents a novel approach to DNA sequence generation, leveraging advanced language model techniques to address specific biological challenges. This could lead to more precise and efficient designs in genetic engineering and synthetic biology.
  • The introduction of ATGC-Gen aligns with ongoing advancements in artificial intelligence and its applications in biology, echoing trends in memory-augmented systems and hierarchical representations in other domains. As AI continues to evolve, the integration of language models into biological research may reshape methodologies and enhance the understanding of complex biological systems.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended apps based on your readingExplore all apps
Continue Readings
Shape and Texture Recognition in Large Vision-Language Models
NeutralArtificial Intelligence
The Large Shapes and Textures dataset (LAS&T) has been introduced to enhance the capabilities of Large Vision-Language Models (LVLMs) in recognizing and representing shapes and textures across various contexts. This dataset, created through unsupervised extraction from natural images, serves as a benchmark for evaluating the performance of leading models like CLIP and DINO in shape recognition tasks.
EEG-to-Text Translation: A Model for Deciphering Human Brain Activity
PositiveArtificial Intelligence
Researchers have introduced the R1 Translator model, which aims to enhance the decoding of EEG signals into text by combining a bidirectional LSTM encoder with a pretrained transformer-based decoder. This model addresses the limitations of existing EEG-to-text translation models, such as T5 and Brain Translator, and demonstrates superior performance in ROUGE metrics.
Using Text-Based Life Trajectories from Swedish Register Data to Predict Residential Mobility with Pretrained Transformers
PositiveArtificial Intelligence
A recent study has transformed extensive Swedish register data into textual life trajectories to predict residential mobility, utilizing data from 6.9 million individuals between 2001 and 2013. By converting demographic and life changes into semantically rich texts, the research employs various NLP architectures, including LSTM and BERT, to enhance prediction accuracy for residential moves from 2013 to 2017.
LUNA: Linear Universal Neural Attention with Generalization Guarantees
PositiveArtificial Intelligence
A new linear attention mechanism named LUNA has been introduced, addressing the computational bottleneck of traditional softmax attention, which operates at a quadratic cost. LUNA achieves linear cost while maintaining or exceeding the accuracy of quadratic attention by learning the kernel feature map tailored to specific data and tasks.
Less Is More for Multi-Step Logical Reasoning of LLM Generalisation Under Rule Removal, Paraphrasing, and Compression
NeutralArtificial Intelligence
Recent research has introduced a controlled evaluation framework to assess the generalization capabilities of large language models (LLMs) like BERT, Qwen2, and LLaMA under various logical perturbations, including rule deletion and contradictory evidence. The findings indicate that these models maintain high accuracy despite structural changes in reasoning tasks.
ContextualSHAP : Enhancing SHAP Explanations Through Contextual Language Generation
PositiveArtificial Intelligence
A new Python package named ContextualSHAP has been proposed to enhance SHAP (SHapley Additive exPlanations) by integrating it with OpenAI's GPT, allowing for the generation of contextualized textual explanations tailored to user-defined parameters. This development aims to bridge the gap in providing meaningful explanations for end-users, particularly those lacking technical expertise.
First Attentions Last: Better Exploiting First Attentions for Efficient Transformer Training
PositiveArtificial Intelligence
A new transformer architecture called FAL (First Attentions Last) has been proposed to enhance the efficiency of training billion-scale transformers by bypassing the MHA-MLP connections, which traditionally require significant communication overhead. This innovation allows for the first layer's attention output to be redirected to the MLP inputs of subsequent layers, facilitating parallel execution on a single GPU.
ARSS: Taming Decoder-only Autoregressive Visual Generation for View Synthesis From Single View
PositiveArtificial Intelligence
A novel framework named ARSS has been introduced, leveraging a GPT-style decoder-only autoregressive model to generate novel views from a single image, conditioned on a predefined camera trajectory. This approach addresses the limitations of existing diffusion-based methods in generating target views along a camera trajectory.