ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2101.09671
  4. Cited By
Pruning and Quantization for Deep Neural Network Acceleration: A Survey

Pruning and Quantization for Deep Neural Network Acceleration: A Survey

24 January 2021
Tailin Liang
C. Glossner
Lei Wang
Shaobo Shi
Xiaotong Zhang
    MQ
ArXivPDFHTML

Papers citing "Pruning and Quantization for Deep Neural Network Acceleration: A Survey"

50 / 202 papers shown
Title
Survey on Evolutionary Deep Learning: Principles, Algorithms,
  Applications and Open Issues
Survey on Evolutionary Deep Learning: Principles, Algorithms, Applications and Open Issues
Nan Li
Lianbo Ma
Guo-Ding Yu
Bing Xue
Mengjie Zhang
Yaochu Jin
27
70
0
23 Aug 2022
Efficient High-Resolution Deep Learning: A Survey
Efficient High-Resolution Deep Learning: A Survey
Arian Bakhtiarnia
Qi Zhang
Alexandros Iosifidis
MedIm
21
19
0
26 Jul 2022
Bitwidth-Adaptive Quantization-Aware Neural Network Training: A
  Meta-Learning Approach
Bitwidth-Adaptive Quantization-Aware Neural Network Training: A Meta-Learning Approach
Jiseok Youn
Jaehun Song
Hyung-Sin Kim
S. Bahk
MQ
9
8
0
20 Jul 2022
DASS: Differentiable Architecture Search for Sparse neural networks
DASS: Differentiable Architecture Search for Sparse neural networks
H. Mousavi
Mohammad Loni
Mina Alibeigi
Masoud Daneshtalab
38
9
0
14 Jul 2022
CPrune: Compiler-Informed Model Pruning for Efficient Target-Aware DNN
  Execution
CPrune: Compiler-Informed Model Pruning for Efficient Target-Aware DNN Execution
Taeho Kim
Yongin Kwon
Jemin Lee
Taeho Kim
Sangtae Ha
27
2
0
04 Jul 2022
Automatic autism spectrum disorder detection using artificial
  intelligence methods with MRI neuroimaging: A review
Automatic autism spectrum disorder detection using artificial intelligence methods with MRI neuroimaging: A review
Parisa Moridian
Navid Ghassemi
M. Jafari
S. Salloum-Asfar
Delaram Sadeghi
...
A. Subasi
R. Alizadehsani
Juan M Gorriz
Sara A. Abdulla
U. Acharya
8
74
0
20 Jun 2022
Why Quantization Improves Generalization: NTK of Binary Weight Neural
  Networks
Why Quantization Improves Generalization: NTK of Binary Weight Neural Networks
Kaiqi Zhang
Ming Yin
Yu-Xiang Wang
MQ
24
4
0
13 Jun 2022
Differentially Private Model Compression
Differentially Private Model Compression
Fatemehsadat Mireshghallah
A. Backurs
Huseyin A. Inan
Lukas Wutschitz
Janardhan Kulkarni
SyDa
21
13
0
03 Jun 2022
Deep Learning Workload Scheduling in GPU Datacenters: Taxonomy,
  Challenges and Vision
Deep Learning Workload Scheduling in GPU Datacenters: Taxonomy, Challenges and Vision
Wei Gao
Qi Hu
Zhisheng Ye
Peng Sun
Xiaolin Wang
Yingwei Luo
Tianwei Zhang
Yonggang Wen
83
26
0
24 May 2022
Dynamic Split Computing for Efficient Deep Edge Intelligence
Dynamic Split Computing for Efficient Deep Edge Intelligence
Arian Bakhtiarnia
Nemanja Milošević
Qi Zhang
Dragana Bajović
Alexandros Iosifidis
25
24
0
23 May 2022
What Do Compressed Multilingual Machine Translation Models Forget?
What Do Compressed Multilingual Machine Translation Models Forget?
Alireza Mohammadshahi
Vassilina Nikoulina
Alexandre Berard
Caroline Brun
James Henderson
Laurent Besacier
AI4CE
42
9
0
22 May 2022
Perturbation of Deep Autoencoder Weights for Model Compression and
  Classification of Tabular Data
Perturbation of Deep Autoencoder Weights for Model Compression and Classification of Tabular Data
Manar D. Samad
Sakib Abrar
22
12
0
17 May 2022
Automation Slicing and Testing for in-App Deep Learning Models
Automation Slicing and Testing for in-App Deep Learning Models
Hao Wu
Yuhang Gong
Xiaopeng Ke
Hanzhong Liang
Minghao Li
Fengyuan Xu
Yunxin Liu
Sheng Zhong
49
1
0
15 May 2022
A Comprehensive Survey on Model Quantization for Deep Neural Networks in
  Image Classification
A Comprehensive Survey on Model Quantization for Deep Neural Networks in Image Classification
Babak Rokh
A. Azarpeyvand
Alireza Khanteymoori
MQ
30
82
0
14 May 2022
A Survey on AI Sustainability: Emerging Trends on Learning Algorithms
  and Research Challenges
A Survey on AI Sustainability: Emerging Trends on Learning Algorithms and Research Challenges
Zhenghua Chen
Min-man Wu
Alvin Chan
Xiaoli Li
Yew-Soon Ong
19
6
0
08 May 2022
Resource-efficient domain adaptive pre-training for medical images
Resource-efficient domain adaptive pre-training for medical images
Y. Mehmood
U. I. Bajwa
Xianfang Sun
14
1
0
28 Apr 2022
Boosting Pruned Networks with Linear Over-parameterization
Boosting Pruned Networks with Linear Over-parameterization
Yundi Qian
Siyuan Pan
Xiaoshuang Li
Jie Zhang
Liang Hou
Xiaobing Tu
11
2
0
25 Apr 2022
MIME: Adapting a Single Neural Network for Multi-task Inference with
  Memory-efficient Dynamic Pruning
MIME: Adapting a Single Neural Network for Multi-task Inference with Memory-efficient Dynamic Pruning
Abhiroop Bhattacharjee
Yeshwanth Venkatesha
Abhishek Moitra
Priyadarshini Panda
19
6
0
11 Apr 2022
Deep neural network goes lighter: A case study of deep compression
  techniques on automatic RF modulation recognition for Beyond 5G networks
Deep neural network goes lighter: A case study of deep compression techniques on automatic RF modulation recognition for Beyond 5G networks
Anu Jagannath
Jithin Jagannath
Yanzhi Wang
Tommaso Melodia
23
3
0
09 Apr 2022
Enabling All In-Edge Deep Learning: A Literature Review
Enabling All In-Edge Deep Learning: A Literature Review
Praveen Joshi
Mohammed Hasanuzzaman
Chandra Thapa
Haithem Afli
T. Scully
31
22
0
07 Apr 2022
Optimizing the Consumption of Spiking Neural Networks with Activity
  Regularization
Optimizing the Consumption of Spiking Neural Networks with Activity Regularization
Simon Narduzzi
Siavash Bigdeli
Shih-Chii Liu
L. A. Dunbar
18
8
0
04 Apr 2022
Adaptive and Cascaded Compressive Sensing
Adaptive and Cascaded Compressive Sensing
Chenxi Qiu
Tao Yue
Xue-mei Hu
25
2
0
21 Mar 2022
Hardware Approximate Techniques for Deep Neural Network Accelerators: A
  Survey
Hardware Approximate Techniques for Deep Neural Network Accelerators: A Survey
Giorgos Armeniakos
Georgios Zervakis
Dimitrios Soudris
J. Henkel
211
93
0
16 Mar 2022
Improvements to Gradient Descent Methods for Quantum Tensor Network
  Machine Learning
Improvements to Gradient Descent Methods for Quantum Tensor Network Machine Learning
F. Barratt
J. Dborin
L. Wright
16
11
0
03 Mar 2022
F8Net: Fixed-Point 8-bit Only Multiplication for Network Quantization
F8Net: Fixed-Point 8-bit Only Multiplication for Network Quantization
Qing Jin
Jian Ren
Richard Zhuang
Sumant Hanumante
Zhengang Li
Zhiyu Chen
Yanzhi Wang
Kai-Min Yang
Sergey Tulyakov
MQ
24
48
0
10 Feb 2022
Quantization in Layer's Input is Matter
Quantization in Layer's Input is Matter
Daning Cheng
Wenguang Chen
MQ
11
0
0
10 Feb 2022
Local Feature Matching with Transformers for low-end devices
Local Feature Matching with Transformers for low-end devices
Kyrylo Kolodiazhnyi
16
0
0
01 Feb 2022
COIN++: Neural Compression Across Modalities
COIN++: Neural Compression Across Modalities
Emilien Dupont
H. Loya
Milad Alizadeh
Adam Goliñski
Yee Whye Teh
Arnaud Doucet
53
82
0
30 Jan 2022
Object Detection in Autonomous Vehicles: Status and Open Challenges
Object Detection in Autonomous Vehicles: Status and Open Challenges
Abhishek Balasubramaniam
S. Pasricha
47
54
0
19 Jan 2022
The Effect of Model Compression on Fairness in Facial Expression
  Recognition
The Effect of Model Compression on Fairness in Facial Expression Recognition
Samuil Stoychev
Hatice Gunes
CVBM
30
19
0
05 Jan 2022
Accurate Neural Training with 4-bit Matrix Multiplications at Standard
  Formats
Accurate Neural Training with 4-bit Matrix Multiplications at Standard Formats
Brian Chmiel
Ron Banner
Elad Hoffer
Hilla Ben Yaacov
Daniel Soudry
MQ
27
22
0
19 Dec 2021
A Survey on Green Deep Learning
A Survey on Green Deep Learning
Jingjing Xu
Wangchunshu Zhou
Zhiyi Fu
Hao Zhou
Lei Li
VLM
73
83
0
08 Nov 2021
When in Doubt, Summon the Titans: Efficient Inference with Large Models
When in Doubt, Summon the Titans: Efficient Inference with Large Models
A. S. Rawat
Manzil Zaheer
A. Menon
Amr Ahmed
Sanjiv Kumar
19
7
0
19 Oct 2021
BERMo: What can BERT learn from ELMo?
BERMo: What can BERT learn from ELMo?
Sangamesh Kodge
Kaushik Roy
38
3
0
18 Oct 2021
Training Deep Neural Networks with Joint Quantization and Pruning of
  Weights and Activations
Training Deep Neural Networks with Joint Quantization and Pruning of Weights and Activations
Xinyu Zhang
Ian Colbert
Ken Kreutz-Delgado
Srinjoy Das
MQ
32
11
0
15 Oct 2021
Shifting Capsule Networks from the Cloud to the Deep Edge
Shifting Capsule Networks from the Cloud to the Deep Edge
Miguel Costa
Diogo Costa
T. Gomes
Sandro Pinto
21
5
0
06 Oct 2021
On the Compression of Neural Networks Using $\ell_0$-Norm Regularization
  and Weight Pruning
On the Compression of Neural Networks Using ℓ0\ell_0ℓ0​-Norm Regularization and Weight Pruning
F. Oliveira
E. Batista
R. Seara
12
9
0
10 Sep 2021
Juvenile state hypothesis: What we can learn from lottery ticket
  hypothesis researches?
Juvenile state hypothesis: What we can learn from lottery ticket hypothesis researches?
Di Zhang
23
1
0
08 Sep 2021
How much pre-training is enough to discover a good subnetwork?
How much pre-training is enough to discover a good subnetwork?
Cameron R. Wolfe
Fangshuo Liao
Qihan Wang
J. Kim
Anastasios Kyrillidis
30
3
0
31 Jul 2021
Experiments on Properties of Hidden Structures of Sparse Neural Networks
Experiments on Properties of Hidden Structures of Sparse Neural Networks
Julian Stier
Harsh Darji
Michael Granitzer
16
2
0
27 Jul 2021
A Lightweight and Gradient-Stable Neural Layer
A Lightweight and Gradient-Stable Neural Layer
Yueyao Yu
Yin Zhang
29
0
0
08 Jun 2021
Model Compression
Model Compression
Arhum Ishtiaq
Sara Mahmood
M. Anees
Neha Mumtaz
15
0
0
20 May 2021
The Untapped Potential of Off-the-Shelf Convolutional Neural Networks
The Untapped Potential of Off-the-Shelf Convolutional Neural Networks
Matthew J. Inkawhich
Nathan Inkawhich
Eric K. Davis
H. Li
Yiran Chen
BDL
18
0
0
17 Mar 2021
Deep Model Compression based on the Training History
Deep Model Compression based on the Training History
S. H. Shabbeer Basha
M. Farazuddin
Viswanath Pulabaigari
S. Dubey
Snehasis Mukherjee
VLM
16
17
0
30 Jan 2021
DAIS: Automatic Channel Pruning via Differentiable Annealing Indicator
  Search
DAIS: Automatic Channel Pruning via Differentiable Annealing Indicator Search
Yushuo Guan
Ning Liu
Pengyu Zhao
Zhengping Che
Kaigui Bian
Yanzhi Wang
Jian Tang
20
38
0
04 Nov 2020
Group Sparsity: The Hinge Between Filter Pruning and Decomposition for
  Network Compression
Group Sparsity: The Hinge Between Filter Pruning and Decomposition for Network Compression
Yawei Li
Shuhang Gu
Christoph Mayer
Luc Van Gool
Radu Timofte
137
189
0
19 Mar 2020
What is the State of Neural Network Pruning?
What is the State of Neural Network Pruning?
Davis W. Blalock
Jose Javier Gonzalez Ortiz
Jonathan Frankle
John Guttag
191
1,027
0
06 Mar 2020
Forward and Backward Information Retention for Accurate Binary Neural
  Networks
Forward and Backward Information Retention for Accurate Binary Neural Networks
Haotong Qin
Ruihao Gong
Xianglong Liu
Mingzhu Shen
Ziran Wei
F. Yu
Jingkuan Song
MQ
131
324
0
24 Sep 2019
Training High-Performance and Large-Scale Deep Neural Networks with Full
  8-bit Integers
Training High-Performance and Large-Scale Deep Neural Networks with Full 8-bit Integers
Yukuan Yang
Shuang Wu
Lei Deng
Tianyi Yan
Yuan Xie
Guoqi Li
MQ
99
110
0
05 Sep 2019
MobileNets: Efficient Convolutional Neural Networks for Mobile Vision
  Applications
MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications
Andrew G. Howard
Menglong Zhu
Bo Chen
Dmitry Kalenichenko
Weijun Wang
Tobias Weyand
M. Andreetto
Hartwig Adam
3DH
950
20,567
0
17 Apr 2017
Previous
12345
Next