Quantifying Edits Decay in Fine-tuned LLMs
NegativeArtificial Intelligence
The study titled 'Quantifying Edits Decay in Fine-tuned LLMs' explores how fine-tuning affects knowledge edits in large language models (LLMs). It addresses a critical question: do edits survive after fine-tuning? The findings indicate that edits do decay, with the extent of survival varying across different configurations. This decay poses challenges for knowledge editing methods, as it could necessitate re-editing fine-tuned models, thereby increasing costs. Additionally, the risk of propagating hidden malicious edits raises serious safety concerns. The research evaluated state-of-the-art editing methods like MEMIT and AlphaEdit, alongside fine-tuning approaches such as full-parameter, LoRA, and DoRA, across multiple LLMs. Notably, AlphaEdit edits were found to decay more than those from MEMIT. The study suggests a selective-layer fine-tuning method as a potential solution to mitigate these issues. As LLMs become increasingly integrated into various applications, understanding the im…
— via World Pulse Now AI Editorial System
