MTA: A Merge-then-Adapt Framework for Personalized Large Language Model

arXiv — cs.CLThursday, November 27, 2025 at 5:00:00 AM
  • The Merge-then-Adapt (MTA) framework has been introduced to enhance Personalized Large Language Models (PLLMs) by addressing scalability and performance issues associated with traditional fine-tuning methods. MTA operates through three stages: creating a shared Meta-LoRA Bank, implementing Adaptive LoRA Fusion, and enabling dynamic personalization, which collectively aim to optimize user-specific model outputs.
  • This development is significant as it allows for a more efficient and scalable approach to personalizing language models, reducing storage costs and improving performance for users with limited data. By leveraging a shared bank of meta-personalization traits, MTA can adapt to diverse user preferences without the need for extensive individual model fine-tuning.
  • The introduction of MTA reflects a broader trend in AI towards more adaptable and efficient frameworks that can handle user heterogeneity and data sparsity. Similar innovations, such as federated learning and parameter-efficient fine-tuning methods, are emerging to tackle challenges in model training and deployment, emphasizing the importance of dynamic adaptation in AI systems to meet varying user needs.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended apps based on your readingExplore all apps
Continue Readings
Dual LoRA: Enhancing LoRA with Magnitude and Direction Updates
PositiveArtificial Intelligence
A novel method called Dual LoRA has been proposed to enhance the performance of Low-Rank Adaptation (LoRA) in fine-tuning large language models (LLMs). This method introduces two distinct groups within low-rank matrices: a magnitude group for controlling the extent of parameter updates and a direction group for determining the update direction, thereby improving the adaptation process.
NAS-LoRA: Empowering Parameter-Efficient Fine-Tuning for Visual Foundation Models with Searchable Adaptation
PositiveArtificial Intelligence
The introduction of NAS-LoRA represents a significant advancement in the adaptation of the Segment Anything Model (SAM) for specialized tasks, particularly in medical and agricultural imaging. This new Parameter-Efficient Fine-Tuning (PEFT) method integrates a Neural Architecture Search (NAS) block to enhance SAM's performance by addressing its limitations in acquiring high-level semantic information due to the lack of spatial priors in its Transformer encoder.
LoRA Patching: Exposing the Fragility of Proactive Defenses against Deepfakes
NegativeArtificial Intelligence
A recent study highlights the vulnerabilities of proactive defenses against deepfakes, revealing that these defenses often lack the necessary robustness and reliability. The research introduces a novel technique called Low-Rank Adaptation (LoRA) patching, which effectively bypasses existing defenses by injecting adaptable patches into deepfake generators. This method also includes a Multi-Modal Feature Alignment loss to ensure semantic consistency in outputs.
Delta Sampling: Data-Free Knowledge Transfer Across Diffusion Models
PositiveArtificial Intelligence
Delta Sampling (DS) has been introduced as a novel method for enabling data-free knowledge transfer across different diffusion models, particularly addressing the challenges faced when upgrading base models like Stable Diffusion. This method operates at inference time, utilizing the delta between model predictions before and after adaptation, thus facilitating the reuse of adaptation components across varying architectures.
PersonaAgent with GraphRAG: Community-Aware Knowledge Graphs for Personalized LLM
PositiveArtificial Intelligence
A novel framework called PersonaAgent with GraphRAG has been proposed to create personalized AI agents that adapt to individual user preferences by embodying the user's persona and utilizing a large language model (LLM). This system integrates a Knowledge-Graph-enhanced Retrieval-Augmented Generation mechanism to summarize community-related information and generate personalized prompts based on user behavior and global interaction patterns.
Glance: Accelerating Diffusion Models with 1 Sample
PositiveArtificial Intelligence
Recent advancements in diffusion models have led to the development of a phase-aware strategy that accelerates image generation by applying different speedups to various stages of the process. This approach utilizes lightweight LoRA adapters, named Slow-LoRA and Fast-LoRA, to enhance efficiency without extensive retraining of models.
PERP: Rethinking the Prune-Retrain Paradigm in the Era of LLMs
PositiveArtificial Intelligence
A recent study titled 'PERP: Rethinking the Prune-Retrain Paradigm in the Era of LLMs' reveals that neural networks can be effectively compressed through pruning, which reduces storage and compute demands while maintaining performance. The research indicates that instead of retraining all parameters, updating a small subset of highly expressive parameters can restore or even enhance performance after pruning, particularly in large language models (LLMs) like GPT.
Pre-trained Language Models Improve the Few-shot Prompt Ability of Decision Transformer
PositiveArtificial Intelligence
The introduction of the Language model-initialized Prompt Decision Transformer (LPDT) framework marks a significant advancement in offline reinforcement learning (RL) by enhancing the few-shot prompt ability of Decision Transformers. This framework utilizes pre-trained language models to improve performance on unseen tasks, addressing challenges related to data collection and the limitations of traditional Prompt-DT methods.