SmileyLlama: Modifying Large Language Models for Directed Chemical Space Exploration

arXiv — cs.LGThursday, November 13, 2025 at 5:00:00 AM
The recent transformation of the Llama-3.1-8B-Instruct model into SmileyLlama marks a significant advancement in the application of large language models (LLMs) for chemical exploration. By employing supervised fine-tuning of engineered prompts, researchers have created a chemical language model that can generate novel drug-like molecules tailored to user specifications. This capability was benchmarked against traditional CLMs trained from extensive ChEMBL data, showcasing SmileyLlama's ability to produce valid and innovative compounds. Furthermore, the integration of direct preference optimization enhances the model's adherence to prompts, while the iMiner reinforcement learning framework aids in predicting new drug molecules with optimized 3D conformations and high binding affinity to targets, such as the SARS-Cov-2 Main Protease. Although the current dataset focuses on drug discovery, the methodologies developed could be extended to various chemical applications, including chemical …
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended Readings
Synergizing Multigrid Algorithms with Vision Transformer: A Novel Approach to Enhance the Seismic Foundation Model
PositiveArtificial Intelligence
A novel approach to enhancing seismic foundation models has been introduced, synergizing multigrid algorithms with vision transformers. This method addresses the unique characteristics of seismic data, which require specialized processing techniques. The proposed adaptive two-grid foundation model training strategy (ADATG) utilizes Hilbert encoding to effectively capture both high- and low-frequency features in seismogram data, improving the efficiency of seismic data analysis and model training.
Beyond the Surface: Probing the Ideological Depth of Large Language Models
PositiveArtificial Intelligence
Large language models (LLMs) exhibit distinct political leanings, but their consistency in representing these orientations varies. This study introduces the concept of ideological depth, defined by a model's ability to follow political instructions reliably and the richness of its internal political representations, assessed using sparse autoencoders. The research compares Llama-3.1-8B-Instruct and Gemma-2-9B-IT, revealing that Gemma is significantly more steerable and activates approximately 7.3 times more distinct political features than Llama.
RAG-Enhanced Collaborative LLM Agents for Drug Discovery
PositiveArtificial Intelligence
Recent advancements in large language models (LLMs) have demonstrated significant potential to enhance drug discovery processes. However, the specialized nature of biochemical data often requires expensive domain-specific fine-tuning, which poses challenges for the application of general-purpose LLMs. To overcome these obstacles, the proposed CLADD system utilizes retrieval-augmented generation (RAG) to facilitate dynamic information retrieval from biomedical knowledge bases, thereby improving the efficiency and effectiveness of drug discovery tasks.