Network Pruning

212 papers with code • 5 benchmarks • 5 datasets

Network Pruning is a popular approach to reduce a heavy network to obtain a light-weight form by removing redundancy in the heavy network. In this approach, a complex over-parameterized network is first trained, then pruned based on come criterions, and finally fine-tuned to achieve comparable performance with reduced parameters.

Source: Ensemble Knowledge Distillation for Learning Improved and Efficient Networks

Libraries

Use these libraries to find Network Pruning models and implementations

Latest papers with no code

FedP3: Federated Personalized and Privacy-friendly Network Pruning under Model Heterogeneity

no code yet • 15 Apr 2024

The interest in federated learning has surged in recent research due to its unique ability to train a global model using privacy-secured information held locally on each client.

Aggressive or Imperceptible, or Both: Network Pruning Assisted Hybrid Byzantines in Federated Learning

no code yet • 9 Apr 2024

Hence, inspired by the sparse neural networks, we introduce a hybrid sparse Byzantine attack that is composed of two parts: one exhibiting a sparse nature and attacking only certain NN locations with higher sensitivity, and the other being more silent but accumulating over time, where each ideally targets a different type of defence mechanism, and together they form a strong but imperceptible attack.

FedMef: Towards Memory-efficient Federated Dynamic Pruning

no code yet • 21 Mar 2024

To address these challenges, we propose FedMef, a novel and memory-efficient federated dynamic pruning framework.

LNPT: Label-free Network Pruning and Training

no code yet • 19 Mar 2024

Pruning before training enables the deployment of neural networks on smart devices.

Structurally Prune Anything: Any Architecture, Any Framework, Any Time

no code yet • 3 Mar 2024

However, the diverse patterns for coupling parameters, such as residual connections and group convolutions, the diverse deep learning frameworks, and the various time stages at which pruning can be performed make existing pruning methods less adaptable to different architectures, frameworks, and pruning criteria.

SequentialAttention++ for Block Sparsification: Differentiable Pruning Meets Combinatorial Optimization

no code yet • 27 Feb 2024

Neural network pruning is a key technique towards engineering large yet scalable, interpretable, and generalizable models.

Discriminative Adversarial Unlearning

no code yet • 10 Feb 2024

We consider the scenario of two networks, the attacker $\mathbf{A}$ and the trained defender $\mathbf{D}$ pitted against each other in an adversarial objective, wherein the attacker aims at teasing out the information of the data to be unlearned in order to infer membership, and the defender unlearns to defend the network against the attack, whilst preserving its general performance.

EPSD: Early Pruning with Self-Distillation for Efficient Model Compression

no code yet • 31 Jan 2024

Instead of a simple combination of pruning and SD, EPSD enables the pruned network to favor SD by keeping more distillable weights before training to ensure better distillation of the pruned network.

Enhancing Scalability in Recommender Systems through Lottery Ticket Hypothesis and Knowledge Distillation-based Neural Network Pruning

no code yet • 19 Jan 2024

This study introduces an innovative approach aimed at the efficient pruning of neural networks, with a particular focus on their deployment on edge devices.

GD doesn't make the cut: Three ways that non-differentiability affects neural network training

no code yet • 16 Jan 2024

This paper investigates the distinctions between gradient methods applied to non-differentiable functions (NGDMs) and classical gradient descents (GDs) designed for differentiable functions.