Which Layer Causes Distribution Deviation? Entropy-Guided Adaptive Pruning for Diffusion and Flow Models

arXiv — cs.CVThursday, November 27, 2025 at 5:00:00 AM
  • A new framework called EntPruner has been introduced to address parameter redundancy in large-scale vision generative models, specifically diffusion and flow models. This framework employs an entropy-guided automatic progressive pruning strategy, which assesses the importance of model blocks based on Conditional Entropy Deviation (CED) to optimize performance across various downstream tasks.
  • The development of EntPruner is significant as it enhances the efficiency of generative models, ensuring that the diversity and fidelity of output distributions are maintained while reducing unnecessary complexity. This could lead to improved performance in visual generation tasks, making it a valuable tool for researchers and developers in the field.
  • This advancement reflects a broader trend in artificial intelligence where optimizing model efficiency is crucial. As generative models become increasingly complex, methods like EntPruner highlight the ongoing efforts to balance performance and computational resource management, paralleling other innovations such as frequency-decoupled diffusion methods and self-distillation techniques that aim to streamline generative processes.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended apps based on your readingExplore all apps
Continue Readings
MoGAN: Improving Motion Quality in Video Diffusion via Few-Step Motion Adversarial Post-Training
PositiveArtificial Intelligence
MoGAN has been introduced as a motion-centric post-training framework aimed at enhancing motion quality in video diffusion models, which often struggle with issues like jitter and ghosting. This framework utilizes a DiT-based optical-flow discriminator to improve motion realism without relying on reward models or human preference data.
Filter Like You Test: Data-Driven Data Filtering for CLIP Pretraining
PositiveArtificial Intelligence
The introduction of Filter Like You Test (FLYT) presents a novel algorithm for curating large-scale vision-language datasets, enhancing the selection of pretraining examples by learning the usefulness of each data point through gradient signals from downstream tasks. This is complemented by Mixing-FLYT (M-FLYT) and Soft Cap Sampling (SCS), which improve dataset filtering and accuracy.
Mechanisms of Non-Monotonic Scaling in Vision Transformers
NeutralArtificial Intelligence
A recent study on Vision Transformers (ViTs) reveals a non-monotonic scaling behavior, where deeper models like ViT-L may underperform compared to shallower variants such as ViT-S and ViT-B. This research identifies a three-phase pattern—Cliff-Plateau-Climb—indicating how representation quality evolves with depth, particularly noting the diminishing role of the [CLS] token in favor of patch tokens for better performance.
LTD: Low Temperature Distillation for Gradient Masking-free Adversarial Training
PositiveArtificial Intelligence
A novel approach called Low-Temperature Distillation (LTD) has been introduced to enhance adversarial training in neural networks, addressing the vulnerabilities associated with one-hot label representations in image classification. LTD utilizes a lower temperature in the teacher model while keeping the student model's temperature fixed, refining label representations and improving model robustness against adversarial attacks.
DP-MicroAdam: Private and Frugal Algorithm for Training and Fine-tuning
PositiveArtificial Intelligence
The introduction of DP-MicroAdam marks a significant advancement in the realm of adaptive optimizers for differentially private training, demonstrating superior performance and convergence rates compared to traditional methods like DP-SGD. This new algorithm is designed to be memory-efficient and sparsity-aware, addressing the challenges of extensive compute and hyperparameter tuning typically associated with differential privacy.
ModHiFi: Identifying High Fidelity predictive components for Model Modification
PositiveArtificial Intelligence
A recent study titled 'ModHiFi: Identifying High Fidelity predictive components for Model Modification' explores methods to modify open weight models without access to training data or loss functions. The research focuses on identifying critical components that influence predictive performance using only distributional access, such as synthetic data.
MGAS: Multi-Granularity Architecture Search for Trade-Off Between Model Effectiveness and Efficiency
PositiveArtificial Intelligence
The introduction of Multi-Granularity Differentiable Architecture Search (MG-DARTS) marks a significant advancement in neural architecture search (NAS), focusing on optimizing both model effectiveness and efficiency. This framework addresses limitations in existing differentiable architecture search methods by incorporating finer-grained structures, enhancing the balance between model performance and size.
DeCo: Frequency-Decoupled Pixel Diffusion for End-to-End Image Generation
PositiveArtificial Intelligence
The newly proposed DeCo framework introduces a frequency-decoupled pixel diffusion method for end-to-end image generation, addressing the inefficiencies of existing models that combine high and low-frequency signal modeling within a single diffusion transformer. This innovation allows for improved training and inference speeds by separating the generation processes of high-frequency details and low-frequency semantics.