Decorrelation Speeds Up Vision Transformers
PositiveArtificial Intelligence
- Recent advancements in Vision Transformers (ViTs) have been achieved through the integration of Decorrelated Backpropagation (DBP) into Masked Autoencoder (MAE) pre-training, resulting in a 21.1% reduction in wall-clock time and a 21.4% decrease in carbon emissions during training on datasets like ImageNet-1K and ADE20K.
- This development is significant as it enhances the efficiency of ViTs in low-label data scenarios, making them more practical for industrial applications where computational resources are limited.
- The ongoing evolution of ViTs highlights a broader trend in AI research towards optimizing model performance while addressing environmental concerns, as seen in various approaches that aim to reduce computational costs and improve model generalization across different tasks.
— via World Pulse Now AI Editorial System
