Boosting Medical Visual Understanding From Multi-Granular Language Learning

arXiv — cs.CVFriday, November 21, 2025 at 5:00:00 AM

Was this article worth reading? Share it

Recommended apps based on your readingExplore all apps
Continue Readings
The Finer the Better: Towards Granular-aware Open-set Domain Generalization
PositiveArtificial Intelligence
The recent introduction of the Semantic-enhanced CLIP (SeeCLIP) framework addresses the challenges of Open-Set Domain Generalization (OSDG), particularly the risks associated with distinguishing known and unknown classes in vision-language models. SeeCLIP enhances semantic understanding by decomposing images into detailed semantic tokens, improving model performance in recognizing novel object categories amidst domain shifts.
Sparse Reasoning is Enough: Biological-Inspired Framework for Video Anomaly Detection with Large Pre-trained Models
PositiveArtificial Intelligence
A novel framework named ReCoVAD has been proposed for video anomaly detection (VAD), inspired by the human nervous system's dual pathways. This framework allows for selective frame processing, significantly reducing computational costs associated with dense frame-level inference. The approach leverages large pre-trained models, enhancing VAD's efficiency in applications such as security surveillance and autonomous driving.
SpatialGeo:Boosting Spatial Reasoning in Multimodal LLMs via Geometry-Semantics Fusion
PositiveArtificial Intelligence
SpatialGeo has been introduced as a novel vision encoder that enhances the spatial reasoning capabilities of multimodal large language models (MLLMs) by integrating geometry and semantics features. This advancement addresses the limitations of existing MLLMs, particularly in interpreting spatial arrangements in three-dimensional space, which has been a significant challenge in the field.
ATAC: Augmentation-Based Test-Time Adversarial Correction for CLIP
PositiveArtificial Intelligence
A new method called Augmentation-Based Test-Time Adversarial Correction (ATAC) has been proposed to enhance the robustness of the CLIP model against adversarial perturbations in images. This approach operates in the embedding space of CLIP, utilizing augmentation-induced drift vectors to correct embeddings based on angular consistency. The method has shown to outperform previous state-of-the-art techniques by nearly 50% in robustness across various benchmarks.
MindShot: A Few-Shot Brain Decoding Framework via Transferring Cross-Subject Prior and Distilling Frequency Domain Knowledge
PositiveArtificial Intelligence
A new framework named MindShot has been introduced to enhance brain decoding by reconstructing visual stimuli from brain signals, addressing challenges like individual differences and high data collection costs. This two-stage framework includes a Multi-Subject Pretraining (MSP) stage and a Fourier-based cross-subject Knowledge Distillation (FKD) stage, aiming to improve adaptability for clinical applications.
SafeR-CLIP: Mitigating NSFW Content in Vision-Language Models While Preserving Pre-Trained Knowledge
PositiveArtificial Intelligence
The introduction of SaFeR-CLIP marks a significant advancement in enhancing the safety of vision-language models like CLIP by employing a proximity-aware approach to redirect unsafe concepts to semantically similar safe alternatives. This method minimizes representational changes while improving zero-shot accuracy by up to 8.0% compared to previous techniques.