VeCoR - Velocity Contrastive Regularization for Flow Matching

arXiv — cs.CVTuesday, November 25, 2025 at 5:00:00 AM
  • The introduction of Velocity Contrastive Regularization (VeCoR) enhances Flow Matching (FM) by implementing a balanced attract-repel scheme, which guides the learned velocity field towards stable directions while avoiding off-manifold errors. This development aims to improve stability and generalization in generative modeling, particularly in lightweight configurations.
  • VeCoR's implementation is significant as it addresses the limitations of standard FM, which can lead to perceptual degradation in generative models. By providing explicit guidance on both positive and negative directions, VeCoR aims to refine the generative process, potentially leading to higher quality outputs in various applications.
  • The advancement of VeCoR reflects a broader trend in artificial intelligence where enhancing generative models is crucial for applications ranging from image synthesis to speech recognition. This aligns with ongoing research efforts to improve model robustness and accuracy, as seen in related works that explore multi-modal integration and out-of-distribution detection.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended apps based on your readingExplore all apps
Continue Readings
Cross-modal Proxy Evolving for OOD Detection with Vision-Language Models
PositiveArtificial Intelligence
A new framework named CoEvo has been proposed for zero-shot out-of-distribution (OOD) detection in vision-language models, addressing the challenges posed by the absence of labeled negatives. CoEvo employs a bidirectional adaptation mechanism for both textual and visual proxies, dynamically refining them based on contextual information from test images. This innovation aims to enhance the reliability of OOD detection in open-world applications.
DGAE: Diffusion-Guided Autoencoder for Efficient Latent Representation Learning
PositiveArtificial Intelligence
The introduction of the Diffusion-Guided Autoencoder (DGAE) marks a significant advancement in latent representation learning, enhancing the decoder's expressiveness and effectively addressing training instability associated with GANs. This model achieves state-of-the-art performance while utilizing a latent space that is twice as compact, thus improving efficiency in image and video generative tasks.

Ready to build your own newsroom?

Subscribe to unlock a personalised feed, podcasts, newsletters, and notifications tailored to the topics you actually care about