Diffusion-SDPO: Safeguarded Direct Preference Optimization for Diffusion Models

arXiv — cs.CVThursday, November 6, 2025 at 5:00:00 AM
A recent study on text-to-image diffusion models highlights challenges in aligning generated images with human preferences. The research revisits Direct Preference Optimization (DPO) and reveals that simply increasing the preference margin may not enhance image quality, potentially leading to higher reconstruction errors. This finding is significant as it prompts a reevaluation of current optimization strategies, which could influence future developments in AI-generated imagery.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended Readings
SCALEX: Scalable Concept and Latent Exploration for Diffusion Models
PositiveArtificial Intelligence
SCALEX is a newly introduced framework designed for scalable and automated exploration of latent spaces in diffusion models. It addresses the issue of social biases, such as gender and racial stereotypes, that are often encoded in image generation models. By utilizing natural language prompts, SCALEX enables zero-shot interpretation, allowing for systematic comparisons across various concepts and facilitating the discovery of internal model associations without the need for retraining or labeling.
Optimizing Input of Denoising Score Matching is Biased Towards Higher Score Norm
NeutralArtificial Intelligence
The paper titled 'Optimizing Input of Denoising Score Matching is Biased Towards Higher Score Norm' discusses the implications of using denoising score matching in optimizing diffusion models. It reveals that this optimization disrupts the equivalence between denoising score matching and exact score matching, resulting in a bias that favors higher score norms. The study also highlights similar biases in optimizing data distributions with pre-trained diffusion models, affecting various applications such as MAR, PerCo, and DreamFusion.