D4C: Data-free Quantization for Contrastive Language-Image Pre-training Models

arXiv — cs.LGThursday, November 20, 2025 at 5:00:00 AM
  • The introduction of D4C marks a significant advancement in Data
  • This development is crucial as it enhances model compression techniques without compromising data privacy, which is increasingly important in various applications.
  • The broader implications of this research highlight ongoing challenges in model efficiency and robustness, particularly in the context of Vision
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended apps based on your readingExplore all apps
Continue Readings
Franca: Nested Matryoshka Clustering for Scalable Visual Representation Learning
PositiveArtificial Intelligence
Franca, the first fully open-source vision foundation model, has been introduced, showcasing performance that matches or exceeds proprietary models like DINOv2 and CLIP. This model utilizes a transparent training pipeline and publicly available datasets, addressing limitations in current self-supervised learning clustering methods through a novel nested Matryoshka clustering approach.
SWAGSplatting: Semantic-guided Water-scene Augmented Gaussian Splatting
PositiveArtificial Intelligence
The introduction of SWAGSplatting, a novel framework for underwater 3D reconstruction, addresses the challenges posed by light attenuation and limited visibility in aquatic environments. This approach integrates semantic understanding with 3D Gaussian Splatting, enhancing the accuracy and fidelity of underwater scene reconstruction.
NOVAK: Unified adaptive optimizer for deep neural networks
PositiveArtificial Intelligence
The recent introduction of NOVAK, a unified adaptive optimizer for deep neural networks, combines several advanced techniques including adaptive moment estimation and lookahead synchronization, aiming to enhance the performance and efficiency of neural network training.
FigEx2: Visual-Conditioned Panel Detection and Captioning for Scientific Compound Figures
PositiveArtificial Intelligence
The recent introduction of FigEx2, a visual-conditioned framework, aims to enhance the understanding of scientific compound figures by localizing panels and generating detailed captions directly from the images. This addresses the common issue of missing or inadequate captions that hinder panel-level comprehension.
The Role of Noisy Data in Improving CNN Robustness for Image Classification
PositiveArtificial Intelligence
A recent study highlights the importance of data quality in enhancing the robustness of convolutional neural networks (CNNs) for image classification, specifically through the introduction of controlled noise during training. Utilizing the CIFAR-10 dataset, the research demonstrates that incorporating just 10% noisy data can significantly reduce test loss and improve accuracy under corrupted conditions without adversely affecting performance on clean data.
An Explainable Two Stage Deep Learning Framework for Pericoronitis Assessment in Panoramic Radiographs Using YOLOv8 and ResNet-50
PositiveArtificial Intelligence
A new study has introduced an explainable two-stage deep learning framework for assessing pericoronitis in panoramic radiographs, utilizing YOLOv8 for anatomical localization and a modified ResNet-50 for pathological classification. The system achieved high precision and alignment with radiologists' diagnostic impressions, enhancing interpretability through Grad-CAM visualizations.
MMLGNet: Cross-Modal Alignment of Remote Sensing Data using CLIP
PositiveArtificial Intelligence
A novel multimodal framework, MMLGNet, has been introduced to align heterogeneous remote sensing modalities, such as Hyperspectral Imaging and LiDAR, with natural language semantics using vision-language models like CLIP. This framework employs modality-specific encoders and bi-directional contrastive learning to enhance the understanding of complex Earth observation data.
Cross-modal Proxy Evolving for OOD Detection with Vision-Language Models
PositiveArtificial Intelligence
A new framework named CoEvo has been proposed for zero-shot out-of-distribution (OOD) detection in vision-language models, addressing the challenges posed by the absence of labeled negatives. CoEvo employs a bidirectional adaptation mechanism for both textual and visual proxies, dynamically refining them based on contextual information from test images. This innovation aims to enhance the reliability of OOD detection in open-world applications.

Ready to build your own newsroom?

Subscribe to unlock a personalised feed, podcasts, newsletters, and notifications tailored to the topics you actually care about