CHIPS: Efficient CLIP Adaptation via Curvature-aware Hybrid Influence-based Data Selection

arXiv — cs.LGTuesday, November 25, 2025 at 5:00:00 AM
  • The recent introduction of CHIPS (Curvature-aware Hybrid Influence in Projection Subspace) presents a novel approach to adapting the CLIP model for specific vertical domains by focusing on effective data selection rather than relying solely on large-scale datasets. This method integrates utility scores based on faithfulness, scalability, and retention, aiming to enhance the adaptation process significantly.
  • This development is crucial as it addresses the limitations of traditional fine-tuning strategies and continual pre-training, potentially leading to more efficient and effective adaptations of CLIP in various applications, thereby improving performance in domain-specific tasks.
  • The emergence of CHIPS highlights a growing trend in AI research towards data-centric methodologies, emphasizing the importance of data selection in model training. This shift aligns with ongoing discussions about the balance between model complexity and data efficiency, as seen in other recent advancements in CLIP adaptations and related frameworks that seek to enhance model robustness and generalization capabilities.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended apps based on your readingExplore all apps
Continue Readings
UStyle: Waterbody Style Transfer of Underwater Scenes by Depth-Guided Feature Synthesis
NeutralArtificial Intelligence
The introduction of UStyle represents a significant advancement in underwater imaging, focusing on waterbody style transfer through a novel depth-aware feature synthesis mechanism. This framework addresses the challenges of traditional style transfer methods that struggle with high-scattering mediums, ensuring that underwater images maintain their geometric integrity while achieving artistic stylization.
Vision-Language Models for Infrared Industrial Sensing in Additive Manufacturing Scene Description
PositiveArtificial Intelligence
A new framework named VLM-IRIS has been introduced to enhance infrared industrial sensing in additive manufacturing, addressing the limitations of conventional vision systems in low-light environments. By preprocessing infrared images into RGB-compatible inputs for CLIP-based encoders, this zero-shot learning approach enables effective workpiece presence detection without the need for extensive labeled datasets.
Task-Specific Distance Correlation Matching for Few-Shot Action Recognition
PositiveArtificial Intelligence
A new framework named Task-Specific Distance Correlation Matching for Few-Shot Action Recognition (TS-FSAR) has been proposed to enhance few-shot action recognition by addressing limitations in existing set matching metrics and the adaptation of CLIP models. TS-FSAR includes a visual Ladder Side Network for efficient fine-tuning and aims to capture complex patterns beyond linear dependencies.
Kinetic Mining in Context: Few-Shot Action Synthesis via Text-to-Motion Distillation
PositiveArtificial Intelligence
KineMIC (Kinetic Mining In Context) has been introduced as a transfer learning framework aimed at enhancing few-shot action synthesis for Human Activity Recognition (HAR). This framework addresses the significant domain gap between general Text-to-Motion (T2M) models and the precise requirements of HAR classifiers, leveraging semantic correspondences in text encoding for kinematic distillation.
Depth-Copy-Paste: Multimodal and Depth-Aware Compositing for Robust Face Detection
PositiveArtificial Intelligence
A new framework called Depth Copy Paste has been introduced to enhance face detection systems by utilizing multimodal and depth-aware compositing techniques. This approach aims to improve data augmentation by generating realistic training samples that account for occlusion and varying illumination conditions, addressing limitations of traditional methods that often yield unrealistic composites.
Free-Lunch Color-Texture Disentanglement for Stylized Image Generation
PositiveArtificial Intelligence
A new study presents a tuning-free approach for color-texture disentanglement in stylized image generation, addressing challenges in controlling multiple style attributes in Text-to-Image diffusion models. This method utilizes the Image-Prompt Additivity property in the CLIP image embedding space to extract Color-Texture Embeddings from reference images, enhancing the Disentangled Stylized Image Generation process.
Noise Matters: Optimizing Matching Noise for Diffusion Classifiers
NeutralArtificial Intelligence
Recent advancements in diffusion classifiers (DC) have highlighted the challenges of noise instability, which significantly affects classification performance. The study proposes a method to optimize matching noise, aiming to enhance the stability and speed of DCs by reducing the reliance on ensemble results from numerous sampled noises.
The Finer the Better: Towards Granular-aware Open-set Domain Generalization
PositiveArtificial Intelligence
The Semantic-enhanced CLIP (SeeCLIP) framework has been proposed to address challenges in Open-Set Domain Generalization (OSDG), where models face both domain shifts and novel object categories. This framework enhances fine-grained semantic understanding, allowing for better differentiation between known and unknown classes, particularly those with visual similarities.

Ready to build your own newsroom?

Subscribe to unlock a personalised feed, podcasts, newsletters, and notifications tailored to the topics you actually care about