Network Pruning

213 papers with code • 5 benchmarks • 5 datasets

Network Pruning is a popular approach to reduce a heavy network to obtain a light-weight form by removing redundancy in the heavy network. In this approach, a complex over-parameterized network is first trained, then pruned based on come criterions, and finally fine-tuned to achieve comparable performance with reduced parameters.

Source: Ensemble Knowledge Distillation for Learning Improved and Efficient Networks

Libraries

Use these libraries to find Network Pruning models and implementations

Neural Network Pruning by Gradient Descent

3riccc/neural_pruning 21 Nov 2023

The rapid increase in the parameters of deep learning models has led to significant costs, challenging computational efficiency and model interpretability.

2
21 Nov 2023

Do Localization Methods Actually Localize Memorized Data in LLMs? A Tale of Two Benchmarks

terarachang/memdata 15 Nov 2023

On the other hand, even successful methods identify neurons that are not specific to a single memorized sequence.

1
15 Nov 2023

Beyond Size: How Gradients Shape Pruning Decisions in Large Language Models

vila-lab/gblm-pruner 8 Nov 2023

GBLM-Pruner leverages the first-order term of the Taylor expansion, operating in a training-free manner by harnessing properly normalized gradients from a few calibration samples to determine the pruning metric, and substantially outperforms competitive counterparts like SparseGPT and Wanda in multiple benchmarks.

26
08 Nov 2023

Dynamic Sparse No Training: Training-Free Fine-tuning for Sparse LLMs

zyxxmu/dsnot 13 Oct 2023

Inspired by the Dynamic Sparse Training, DSnoT minimizes the reconstruction error between the dense and sparse LLMs, in the fashion of performing iterative weight pruning-and-growing on top of sparse LLMs.

26
13 Oct 2023

Filter Pruning For CNN With Enhanced Linear Representation Redundancy

bojue-wang/ccm-lrr 10 Oct 2023

In this paper, we propose a new structured pruning method.

1
10 Oct 2023

Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity

luuyin/owl 8 Oct 2023

Large Language Models (LLMs), renowned for their remarkable performance across diverse domains, present a challenge when it comes to practical deployment due to their colossal model size.

39
08 Oct 2023

SWAP: Sparse Entropic Wasserstein Regression for Robust Network Pruning

youlei202/entropic-wasserstein-pruning 7 Oct 2023

This study addresses the challenge of inaccurate gradients in computing the empirical Fisher Information Matrix during neural network pruning.

0
07 Oct 2023

Feather: An Elegant Solution to Effective DNN Sparsification

athglentis/feather 3 Oct 2023

Neural Network pruning is an increasingly popular way for producing compact and efficient models, suitable for resource-limited environments, while preserving high performance.

7
03 Oct 2023

EDAC: Efficient Deployment of Audio Classification Models For COVID-19 Detection

edac-ml4h/edac-ml4h 11 Sep 2023

Various researchers made use of machine learning methods in an attempt to detect COVID-19.

0
11 Sep 2023

A Survey on Deep Neural Network Pruning-Taxonomy, Comparison, Analysis, and Recommendations

hrcheng1066/awesome-pruning 13 Aug 2023

Modern deep neural networks, particularly recent large language models, come with massive model sizes that require significant computational and storage resources.

99
13 Aug 2023