CONGRAD:Conflicting Gradient Filtering for Multilingual Preference Alignment

arXiv — cs.CLWednesday, November 12, 2025 at 5:00:00 AM
The introduction of CONGRAD marks a pivotal advancement in the field of multilingual preference alignment for large language models (LLMs). Traditional joint training methods often suffer from negative interference due to conflicting objectives, which can degrade performance. CONGRAD addresses this issue by employing a filtering technique that selects high-quality preference samples, thereby minimizing gradient conflicts across languages. This innovative approach utilizes gradient surgery and a sublinear gradient compression strategy to enhance efficiency during training. Evaluated on the LLaMA3-8B and Gemma2-2B models across 10 languages, the results demonstrate that CONGRAD consistently outperforms strong baselines, indicating its effectiveness in improving multilingual training outcomes. The implications of this research are significant, as it opens new avenues for enhancing the performance of multilingual models, which are increasingly vital in our globalized world.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended Readings
W2S-AlignTree: Weak-to-Strong Inference-Time Alignment for Large Language Models via Monte Carlo Tree Search
PositiveArtificial Intelligence
The paper titled 'W2S-AlignTree: Weak-to-Strong Inference-Time Alignment for Large Language Models via Monte Carlo Tree Search' introduces a novel framework aimed at improving the alignment of large language models (LLMs) with human preferences. The proposed W2S-AlignTree framework integrates Monte Carlo Tree Search with the Weak-to-Strong Generalization paradigm, addressing the limitations of existing training-time alignment methods. This approach seeks to provide a scalable and adaptable solution for enhancing LLM performance during inference.