Mixture of Ranks with Degradation-Aware Routing for One-Step Real-World Image Super-Resolution

arXiv — cs.CVFriday, November 21, 2025 at 5:00:00 AM
  • The introduction of the Mixture
  • This development is crucial as it addresses the limitations of traditional dense models in effectively handling heterogeneous degraded samples, thereby enhancing the quality of high
  • The ongoing exploration of adaptive frameworks like MoR reflects a broader trend in artificial intelligence towards optimizing model efficiency and performance across various applications, including federated learning and multimodal systems.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Continue Readings
ILoRA: Federated Learning with Low-Rank Adaptation for Heterogeneous Client Aggregation
PositiveArtificial Intelligence
ILoRA, or Federated Learning with Low-Rank Adaptation, addresses three significant challenges in client heterogeneity: initialization instability, rank incompatibility, and client drift under non-IID data. The proposed framework integrates a QR-based initialization, a concatenated QR aggregation mechanism, and an AdamW optimizer with rank-aware control variates. These innovations aim to enhance the stability and performance of federated learning models across diverse client environments.
CAMS: Towards Compositional Zero-Shot Learning via Gated Cross-Attention and Multi-Space Disentanglement
PositiveArtificial Intelligence
CAMS, a new approach to compositional zero-shot learning (CZSL), aims to enhance the understanding of attributes and objects in unseen compositions. By utilizing Gated Cross-Attention and multi-space disentanglement, CAMS improves the extraction of semantic features from visual data, addressing limitations in existing CLIP-based methods that struggle with complete disentanglement. This advancement is expected to enhance generalization capabilities in recognizing novel attribute-object combinations.
Erase to Retain: Low Rank Adaptation Guided Selective Unlearning in Medical Segmentation Networks
PositiveArtificial Intelligence
The study introduces 'Erase to Retain', a framework for selectively unlearning knowledge in medical segmentation networks. This method allows for targeted forgetting of specific representations without the need for complete retraining, utilizing a teacher-student distillation approach combined with Low-Rank Adaptation (LoRA). The framework enhances privacy compliance and ethical deployment in medical imaging by enabling the erasure of sensitive information while maintaining overall anatomical understanding.
Dataset Distillation for Pre-Trained Self-Supervised Vision Models
PositiveArtificial Intelligence
The paper discusses dataset distillation, aiming to create a small set of synthetic images that can train a model to match the performance of one trained on a larger dataset. Unlike previous methods that focus on randomly initialized models, this research targets pre-trained self-supervised vision models. The proposed Linear Gradient Matching method optimizes synthetic images to produce similar gradients in a linear classifier as real data, enhancing the training process.
How Noise Benefits AI-generated Image Detection
PositiveArtificial Intelligence
The rapid advancement of generative models has made it increasingly difficult to distinguish between real and AI-generated images. Researchers have identified that out-of-distribution generalization remains a challenge due to spurious shortcuts used during training. To combat this, they propose the Positive-Incentive Noise for CLIP (PiN-CLIP), which trains a noise generator alongside a detection network to enhance the detection of AI-generated images by mitigating shortcut dominance.
InfoCLIP: Bridging Vision-Language Pretraining and Open-Vocabulary Semantic Segmentation via Information-Theoretic Alignment Transfer
PositiveArtificial Intelligence
InfoCLIP is a novel approach that enhances open-vocabulary semantic segmentation by transferring alignment knowledge from the pretrained CLIP model. It addresses the issue of overfitting during fine-tuning on limited categories by employing an information-theoretic perspective. The method aims to stabilize modality alignment and improve segmentation performance by maximizing mutual information between the pretrained and fine-tuned models.
Segmenting Collision Sound Sources in Egocentric Videos
PositiveArtificial Intelligence
The proposed task of Collision Sound Source Segmentation (CS3) aims to identify and segment objects responsible for collision sounds in egocentric videos. This task addresses challenges such as cluttered visual scenes and brief interactions, utilizing a weakly-supervised method that leverages audio cues and foundation models like CLIP and SAM2. The focus on egocentric video allows for clearer sound identification despite visual complexity.
LoRA on the Go: Instance-level Dynamic LoRA Selection and Merging
PositiveArtificial Intelligence
LoRA on the Go (LoGo) introduces a training-free framework for dynamic selection and merging of Low-Rank Adaptation (LoRA) adapters at the instance level. This approach addresses the limitations of conventional LoRA adapters, which are typically trained for single tasks. By leveraging signals from a single forward pass, LoGo identifies the most relevant adapters for diverse tasks, enhancing performance across multiple NLP benchmarks without the need for additional labeled data.