Defending Multimodal Backdoored Models by Repulsive Visual Prompt Tuning
NeutralArtificial Intelligence
A recent study highlights the vulnerabilities of multimodal contrastive learning models, particularly CLIP, to backdoor attacks. These models, which learn from extensive image-text datasets, can inadvertently encode features that make them susceptible to input perturbations. This research is crucial as it sheds light on the safety concerns surrounding AI models, emphasizing the need for improved defenses against such vulnerabilities.
— Curated by the World Pulse Now AI Editorial System


