BD-Net: Has Depth-Wise Convolution Ever Been Applied in Binary Neural Networks?

arXiv — cs.CVTuesday, November 25, 2025 at 5:00:00 AM
  • A recent study introduces BD-Net, which successfully applies depth-wise convolution in Binary Neural Networks (BNNs) by proposing a 1.58-bit convolution and a pre-BN residual connection to enhance expressiveness and stabilize training. This innovation marks a significant advancement in model compression techniques, achieving a new state-of-the-art performance on ImageNet with MobileNet V1 and outperforming previous methods across various datasets.
  • The development of BD-Net is crucial as it addresses the limitations of extreme quantization in BNNs, which often destabilizes training and reduces representational capacity. By enhancing the optimization process, this approach not only improves the efficiency of lightweight architectures but also opens new avenues for deploying BNNs in resource-constrained environments.
  • This advancement in BNNs aligns with ongoing efforts in the field of artificial intelligence to create more efficient neural network architectures. Techniques such as structured pruning and adaptive fine-tuning are gaining traction, highlighting a broader trend towards optimizing model performance while maintaining low computational costs. The integration of various pruning strategies and quantization methods reflects a growing recognition of the need for efficient AI solutions in diverse applications.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended apps based on your readingExplore all apps
Continue Readings
In Search of Goodness: Large Scale Benchmarking of Goodness Functions for the Forward-Forward Algorithm
PositiveArtificial Intelligence
The Forward-Forward (FF) algorithm presents a biologically plausible alternative to traditional backpropagation in neural networks, focusing on local updates through a scalar measure of 'goodness'. Recent benchmarking of 21 distinct goodness functions across four standard image datasets revealed that certain alternatives significantly outperform the conventional sum-of-squares metric, with notable accuracy improvements on datasets like MNIST and FashionMNIST.
Flow Map Distillation Without Data
PositiveArtificial Intelligence
A new approach to flow map distillation has been introduced, which eliminates the need for external datasets traditionally used in the sampling process. This method aims to mitigate the risks associated with Teacher-Data Mismatch by relying solely on the prior distribution, ensuring that the teacher's generative capabilities are accurately represented without data dependency.
DeCo: Frequency-Decoupled Pixel Diffusion for End-to-End Image Generation
PositiveArtificial Intelligence
The newly proposed DeCo framework introduces a frequency-decoupled pixel diffusion method for end-to-end image generation, addressing the inefficiencies of existing models that combine high and low-frequency signal modeling within a single diffusion transformer. This innovation allows for improved training and inference speeds by separating the generation processes of high-frequency details and low-frequency semantics.
Temporal-adaptive Weight Quantization for Spiking Neural Networks
PositiveArtificial Intelligence
A new study introduces Temporal-adaptive Weight Quantization (TaWQ) for Spiking Neural Networks (SNNs), which aims to reduce energy consumption while maintaining accuracy. This method leverages temporal dynamics to allocate ultra-low-bit weights, demonstrating minimal quantization loss of 0.22% on ImageNet and high energy efficiency in extensive experiments.
Annotation-Free Class-Incremental Learning
PositiveArtificial Intelligence
A new paradigm in continual learning, Annotation-Free Class-Incremental Learning (AFCIL), has been introduced, addressing the challenge of learning from unlabeled data that arrives sequentially. This approach allows systems to adapt to new classes without supervision, marking a significant shift from traditional methods reliant on labeled data.
TSRE: Channel-Aware Typical Set Refinement for Out-of-Distribution Detection
PositiveArtificial Intelligence
A new method called Channel-Aware Typical Set Refinement (TSRE) has been proposed for Out-of-Distribution (OOD) detection, addressing the limitations of existing activation-based methods that often neglect channel characteristics, leading to inaccurate typical set estimations. This method enhances the separation between in-distribution and OOD data, improving model reliability in open-world environments.
EVCC: Enhanced Vision Transformer-ConvNeXt-CoAtNet Fusion for Classification
PositiveArtificial Intelligence
The introduction of EVCC (Enhanced Vision Transformer-ConvNeXt-CoAtNet) marks a significant advancement in hybrid vision architectures, integrating Vision Transformers, lightweight ConvNeXt, and CoAtNet. This multi-branch architecture employs innovative techniques such as adaptive token pruning and gated bidirectional cross-attention, achieving state-of-the-art accuracy on various datasets while reducing computational costs by 25 to 35% compared to existing models.
Understanding, Accelerating, and Improving MeanFlow Training
PositiveArtificial Intelligence
Recent advancements in MeanFlow training have clarified the dynamics between instantaneous and average velocity fields, revealing that effective learning of average velocity relies on the prior establishment of accurate instantaneous velocities. This understanding has led to the design of a new training scheme that accelerates the formation of these velocities, enhancing the overall training process.