ResearchTrend.AI
  • Papers
  • Communities
  • Organizations
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1510.00149
  4. Cited By
Deep Compression: Compressing Deep Neural Networks with Pruning, Trained
  Quantization and Huffman Coding
v1v2v3v4v5 (latest)

Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding

1 October 2015
Song Han
Huizi Mao
W. Dally
    3DGS
ArXiv (abs)PDFHTML

Papers citing "Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding"

50 / 3,481 papers shown
Title
An Efficient Method of Training Small Models for Regression Problems
  with Knowledge Distillation
An Efficient Method of Training Small Models for Regression Problems with Knowledge Distillation
M. Takamoto
Yusuke Morishita
Hitoshi Imaoka
65
33
0
28 Feb 2020
Learning in the Frequency Domain
Learning in the Frequency Domain
Kai Xu
Minghai Qin
Fei Sun
Yuhao Wang
Yen-kuang Chen
Fengbo Ren
165
412
0
27 Feb 2020
Advances in Collaborative Filtering and Ranking
Advances in Collaborative Filtering and Ranking
Liwei Wu
55
7
0
27 Feb 2020
Train Large, Then Compress: Rethinking Model Size for Efficient Training
  and Inference of Transformers
Train Large, Then Compress: Rethinking Model Size for Efficient Training and Inference of Transformers
Zhuohan Li
Eric Wallace
Sheng Shen
Kevin Lin
Kurt Keutzer
Dan Klein
Joseph E. Gonzalez
147
151
0
26 Feb 2020
Moniqua: Modulo Quantized Communication in Decentralized SGD
Moniqua: Modulo Quantized Communication in Decentralized SGD
Yucheng Lu
Christopher De Sa
MQ
82
50
0
26 Feb 2020
Quantized Neural Network Inference with Precision Batching
Quantized Neural Network Inference with Precision Batching
Maximilian Lam
Zachary Yedidia
Colby R. Banbury
Vijay Janapa Reddi
MQ
50
1
0
26 Feb 2020
Optimal Gradient Quantization Condition for Communication-Efficient
  Distributed Training
Optimal Gradient Quantization Condition for Communication-Efficient Distributed Training
An Xu
Zhouyuan Huo
Heng-Chiao Huang
MQ
40
6
0
25 Feb 2020
HYDRA: Pruning Adversarially Robust Neural Networks
HYDRA: Pruning Adversarially Robust Neural Networks
Vikash Sehwag
Shiqi Wang
Prateek Mittal
Suman Jana
AAML
75
25
0
24 Feb 2020
HRank: Filter Pruning using High-Rank Feature Map
HRank: Filter Pruning using High-Rank Feature Map
Mingbao Lin
Rongrong Ji
Yan Wang
Yichen Zhang
Baochang Zhang
Yonghong Tian
Ling Shao
142
730
0
24 Feb 2020
Gradual Channel Pruning while Training using Feature Relevance Scores
  for Convolutional Neural Networks
Gradual Channel Pruning while Training using Feature Relevance Scores for Convolutional Neural Networks
Sai Aparna Aketi
Sourjya Roy
A. Raghunathan
Kaushik Roy
65
22
0
23 Feb 2020
Communication-Efficient Edge AI: Algorithms and Systems
Communication-Efficient Edge AI: Algorithms and Systems
Yuanming Shi
Kai Yang
Tao Jiang
Jun Zhang
Khaled B. Letaief
GNN
101
335
0
22 Feb 2020
Post-training Quantization with Multiple Points: Mixed Precision without
  Mixed Precision
Post-training Quantization with Multiple Points: Mixed Precision without Mixed Precision
Xingchao Liu
Mao Ye
Dengyong Zhou
Qiang Liu
MQ
92
43
0
20 Feb 2020
SpArch: Efficient Architecture for Sparse Matrix Multiplication
SpArch: Efficient Architecture for Sparse Matrix Multiplication
Zhekai Zhang
Hanrui Wang
Song Han
W. Dally
102
234
0
20 Feb 2020
Performance Aware Convolutional Neural Network Channel Pruning for
  Embedded GPUs
Performance Aware Convolutional Neural Network Channel Pruning for Embedded GPUs
Valentin Radu
Kuba Kaszyk
Yuan Wen
Jack Turner
José Cano
Elliot J. Crowley
Björn Franke
Amos Storkey
Michael F. P. O'Boyle
CVBM
70
38
0
20 Feb 2020
NeuroFabric: Identifying Ideal Topologies for Training A Priori Sparse
  Networks
NeuroFabric: Identifying Ideal Topologies for Training A Priori Sparse Networks
Mihailo Isakov
Michel A. Kinsy
98
1
0
19 Feb 2020
RTMobile: Beyond Real-Time Mobile Acceleration of RNNs for Speech
  Recognition
RTMobile: Beyond Real-Time Mobile Acceleration of RNNs for Speech Recognition
Zhaoyang Han
Siyue Wang
Wei Niu
Chengming Zhang
Sheng Lin
...
Yifan Gong
Bin Ren
Xinyu Lin
Yanzhi Wang
Dingwen Tao
68
46
0
19 Feb 2020
DeepLight: Deep Lightweight Feature Interactions for Accelerating CTR
  Predictions in Ad Serving
DeepLight: Deep Lightweight Feature Interactions for Accelerating CTR Predictions in Ad Serving
Wei Deng
Junwei Pan
Tian Zhou
Deguang Kong
Aaron Flores
Guang Lin
34
4
0
17 Feb 2020
BinaryDuo: Reducing Gradient Mismatch in Binary Activation Network by
  Coupling Binary Activations
BinaryDuo: Reducing Gradient Mismatch in Binary Activation Network by Coupling Binary Activations
Hyungjun Kim
Kyungsu Kim
Jinseok Kim
Jae-Joon Kim
MQ
80
48
0
16 Feb 2020
Boosted Locality Sensitive Hashing: Discriminative Binary Codes for
  Source Separation
Boosted Locality Sensitive Hashing: Discriminative Binary Codes for Source Separation
Sunwoo Kim
Haici Yang
Minje Kim
39
9
0
14 Feb 2020
Improving Efficiency in Neural Network Accelerator Using Operands
  Hamming Distance optimization
Improving Efficiency in Neural Network Accelerator Using Operands Hamming Distance optimization
Meng Li
Yilei Li
P. Chuang
Liangzhen Lai
Vikas Chandra
25
3
0
13 Feb 2020
Retrain or not retrain? -- efficient pruning methods of deep CNN
  networks
Retrain or not retrain? -- efficient pruning methods of deep CNN networks
Marcin Pietroñ
Maciej Wielgosz
CVBM
36
17
0
12 Feb 2020
PCNN: Pattern-based Fine-Grained Regular Pruning towards Optimizing CNN
  Accelerators
PCNN: Pattern-based Fine-Grained Regular Pruning towards Optimizing CNN Accelerators
Zhanhong Tan
Jiebo Song
Xiaolong Ma
S. Tan
Hongyang Chen
...
Yifu Wu
Shaokai Ye
Yanzhi Wang
Dehui Li
Kaisheng Ma
84
24
0
11 Feb 2020
A study of local optima for learning feature interactions using neural
  networks
A study of local optima for learning feature interactions using neural networks
Yangzi Guo
Adrian Barbu
120
1
0
11 Feb 2020
Network Pruning via Annealing and Direct Sparsity Control
Network Pruning via Annealing and Direct Sparsity Control
Yangzi Guo
Yiyuan She
Adrian Barbu
30
0
0
11 Feb 2020
Understanding and Improving Knowledge Distillation
Understanding and Improving Knowledge Distillation
Jiaxi Tang
Rakesh Shivanna
Zhe Zhao
Dong Lin
Anima Singh
Ed H. Chi
Sagar Jain
116
134
0
10 Feb 2020
Convolutional Neural Network Pruning Using Filter Attenuation
Convolutional Neural Network Pruning Using Filter Attenuation
Morteza Mousa Pasandi
M. Hajabdollahi
N. Karimi
S. Samavi
S. Shirani
3DPC
27
3
0
09 Feb 2020
PixelHop++: A Small Successive-Subspace-Learning-Based (SSL-based) Model
  for Image Classification
PixelHop++: A Small Successive-Subspace-Learning-Based (SSL-based) Model for Image Classification
Yueru Chen
Mozhdeh Rouhsedaghat
Suya You
Raghuveer Rao
C.-C. Jay Kuo
57
70
0
08 Feb 2020
Activation Density driven Energy-Efficient Pruning in Training
Activation Density driven Energy-Efficient Pruning in Training
Timothy Foldy-Porto
Yeshwanth Venkatesha
Priyadarshini Panda
59
4
0
07 Feb 2020
BERT-of-Theseus: Compressing BERT by Progressive Module Replacing
BERT-of-Theseus: Compressing BERT by Progressive Module Replacing
Canwen Xu
Wangchunshu Zhou
Tao Ge
Furu Wei
Ming Zhou
372
201
0
07 Feb 2020
BABO: Background Activation Black-Out for Efficient Object Detection
BABO: Background Activation Black-Out for Efficient Object Detection
Byungseok Roh
Hankyu Cho
Myung-Ho Ju
Soon Hyung Pyo
ObjD
22
1
0
05 Feb 2020
Towards Explainable Bit Error Tolerance of Resistive RAM-Based Binarized
  Neural Networks
Towards Explainable Bit Error Tolerance of Resistive RAM-Based Binarized Neural Networks
Sebastian Buschjäger
Jian-Jia Chen
Kuan-Hsun Chen
Mario Günzel
Christian Hakert
K. Morik
Rodion Novkin
Lukas Pfahler
Mikail Yayla
MQAAML
32
2
0
03 Feb 2020
CMOS-Free Multilayer Perceptron Enabled by Four-Terminal MTJ Device
CMOS-Free Multilayer Perceptron Enabled by Four-Terminal MTJ Device
Wesley H. Brigner
Naimul Hassan
Xuan Hu
C. Bennett
F. García-Sánchez
M. Marinella
J. Incorvia
Joseph S. Friedman
17
4
0
03 Feb 2020
Exponential discretization of weights of neural network connections in
  pre-trained neural networks
Exponential discretization of weights of neural network connections in pre-trained neural networks
M. Malsagov
E. Khayrov
M. Pushkareva
I. Karandashev
21
5
0
03 Feb 2020
Widening and Squeezing: Towards Accurate and Efficient QNNs
Widening and Squeezing: Towards Accurate and Efficient QNNs
Chuanjian Liu
Kai Han
Yunhe Wang
Hanting Chen
Qi Tian
Chunjing Xu
MQ
31
0
0
03 Feb 2020
An Equivalence between Bayesian Priors and Penalties in Variational
  Inference
An Equivalence between Bayesian Priors and Penalties in Variational Inference
Pierre Wolinski
Guillaume Charpiat
Yann Ollivier
BDL
51
1
0
01 Feb 2020
Analysis of Deep Feature Loss based Enhancement for Speaker Verification
Analysis of Deep Feature Loss based Enhancement for Speaker Verification
Saurabh Kataria
P. S. Nidadavolu
Jesús Villalba
Najim Dehak
103
13
0
01 Feb 2020
Post-Training Piecewise Linear Quantization for Deep Neural Networks
Post-Training Piecewise Linear Quantization for Deep Neural Networks
Jun Fang
Ali Shafiee
Hamzah Abdel-Aziz
D. Thorsley
Georgios Georgiadis
Joseph Hassoun
MQ
122
149
0
31 Jan 2020
How Does BN Increase Collapsed Neural Network Filters?
How Does BN Increase Collapsed Neural Network Filters?
Sheng Zhou
Xinjiang Wang
Ping Luo
Xue Jiang
Wenjie Li
Wei Zhang
57
1
0
30 Jan 2020
MSE-Optimal Neural Network Initialization via Layer Fusion
MSE-Optimal Neural Network Initialization via Layer Fusion
Ramina Ghods
Andrew Lan
Tom Goldstein
Christoph Studer
FedML
33
1
0
28 Jan 2020
Efficient and Stable Graph Scattering Transforms via Pruning
Efficient and Stable Graph Scattering Transforms via Pruning
V. Ioannidis
Siheng Chen
G. Giannakis
71
11
0
27 Jan 2020
Compressing Language Models using Doped Kronecker Products
Compressing Language Models using Doped Kronecker Products
Urmish Thakker
Paul Whatamough
Zhi-Gang Liu
Matthew Mattina
Jesse G. Beu
91
9
0
24 Jan 2020
Progressive Local Filter Pruning for Image Retrieval Acceleration
Progressive Local Filter Pruning for Image Retrieval Acceleration
Xiaodong Wang
Zhedong Zheng
Yang He
Fei Yan
Zhi-qiang Zeng
Yi Yang
85
36
0
24 Jan 2020
SPACE: Structured Compression and Sharing of Representational Space for
  Continual Learning
SPACE: Structured Compression and Sharing of Representational Space for Continual Learning
Gobinda Saha
Isha Garg
Aayush Ankit
Kaushik Roy
CLL
65
22
0
23 Jan 2020
RPN: A Residual Pooling Network for Efficient Federated Learning
RPN: A Residual Pooling Network for Efficient Federated Learning
Anbu Huang
Yuanyuan Chen
Yang Liu
Tianjian Chen
Qiang Yang
FedML
74
11
0
23 Jan 2020
DCT-Conv: Coding filters in convolutional networks with Discrete Cosine
  Transform
DCT-Conv: Coding filters in convolutional networks with Discrete Cosine Transform
Karol Cheinski
Pawel Wawrzynski
34
11
0
23 Jan 2020
Pruning CNN's with linear filter ensembles
Pruning CNN's with linear filter ensembles
Csanád Sándor
Szabolcs Pável
Lehel Csató
3DPC
50
3
0
22 Jan 2020
Are Accelerometers for Activity Recognition a Dead-end?
Are Accelerometers for Activity Recognition a Dead-end?
C. Tong
Shyam A. Tailor
Nicholas D. Lane
HAI
51
33
0
22 Jan 2020
Deep Learning for Sensor-based Human Activity Recognition: Overview,
  Challenges and Opportunities
Deep Learning for Sensor-based Human Activity Recognition: Overview, Challenges and Opportunities
Kaixuan Chen
Dalin Zhang
Lina Yao
Bin Guo
Zhiwen Yu
Yunhao Liu
HAI
108
648
0
21 Jan 2020
Generalization Bounds for High-dimensional M-estimation under Sparsity
  Constraint
Generalization Bounds for High-dimensional M-estimation under Sparsity Constraint
Xiao-Tong Yuan
Ping Li
88
2
0
20 Jan 2020
An Image Enhancing Pattern-based Sparsity for Real-time Inference on
  Mobile Devices
An Image Enhancing Pattern-based Sparsity for Real-time Inference on Mobile Devices
Xiaolong Ma
Wei Niu
Tianyun Zhang
Sijia Liu
Sheng Lin
...
Xiang Chen
Jian Tang
Kaisheng Ma
Bin Ren
Yanzhi Wang
103
27
0
20 Jan 2020
Previous
123...444546...686970
Next