ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2006.03669
  4. Cited By
An Overview of Neural Network Compression
v1v2 (latest)

An Overview of Neural Network Compression

5 June 2020
James OÑeill
    AI4CE
ArXiv (abs)PDFHTML

Papers citing "An Overview of Neural Network Compression"

50 / 157 papers shown
Title
An Introductory Survey on Attention Mechanisms in NLP Problems
An Introductory Survey on Attention Mechanisms in NLP Problems
Dichao Hu
AIMat
72
247
0
12 Nov 2018
Knowledge Transfer via Distillation of Activation Boundaries Formed by
  Hidden Neurons
Knowledge Transfer via Distillation of Activation Boundaries Formed by Hidden Neurons
Byeongho Heo
Minsik Lee
Sangdoo Yun
J. Choi
55
527
0
08 Nov 2018
ReLeQ: A Reinforcement Learning Approach for Deep Quantization of Neural
  Networks
ReLeQ: A Reinforcement Learning Approach for Deep Quantization of Neural Networks
Ahmed T. Elthakeb
Prannoy Pilligundla
Fatemehsadat Mireshghallah
Amir Yazdanbakhsh
H. Esmaeilzadeh
MQ
124
68
0
05 Nov 2018
Trellis Networks for Sequence Modeling
Trellis Networks for Sequence Modeling
Shaojie Bai
J. Zico Kolter
V. Koltun
73
146
0
15 Oct 2018
Rethinking the Value of Network Pruning
Rethinking the Value of Network Pruning
Zhuang Liu
Mingjie Sun
Tinghui Zhou
Gao Huang
Trevor Darrell
40
1,475
0
11 Oct 2018
BERT: Pre-training of Deep Bidirectional Transformers for Language
  Understanding
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Jacob Devlin
Ming-Wei Chang
Kenton Lee
Kristina Toutanova
VLMSSLSSeg
1.8K
95,229
0
11 Oct 2018
SNIP: Single-shot Network Pruning based on Connection Sensitivity
SNIP: Single-shot Network Pruning based on Connection Sensitivity
Namhoon Lee
Thalaiyasingam Ajanthan
Philip Torr
VLM
269
1,211
0
04 Oct 2018
Gradient Descent Provably Optimizes Over-parameterized Neural Networks
Gradient Descent Provably Optimizes Over-parameterized Neural Networks
S. Du
Xiyu Zhai
Barnabás Póczós
Aarti Singh
MLTODL
236
1,276
0
04 Oct 2018
Post-training 4-bit quantization of convolution networks for
  rapid-deployment
Post-training 4-bit quantization of convolution networks for rapid-deployment
Ron Banner
Yury Nahshan
Elad Hoffer
Daniel Soudry
MQ
67
94
0
02 Oct 2018
Neural Architecture Optimization
Neural Architecture Optimization
Renqian Luo
Fei Tian
Tao Qin
Enhong Chen
Tie-Yan Liu
3DV
92
655
0
22 Aug 2018
Recurrent Stacking of Layers for Compact Neural Machine Translation
  Models
Recurrent Stacking of Layers for Compact Neural Machine Translation Models
Raj Dabre
Atsushi Fujita
57
60
0
14 Jul 2018
Universal Transformers
Universal Transformers
Mostafa Dehghani
Stephan Gouws
Oriol Vinyals
Jakob Uszkoreit
Lukasz Kaiser
87
756
0
10 Jul 2018
DARTS: Differentiable Architecture Search
DARTS: Differentiable Architecture Search
Hanxiao Liu
Karen Simonyan
Yiming Yang
206
4,368
0
24 Jun 2018
Quantizing deep convolutional networks for efficient inference: A
  whitepaper
Quantizing deep convolutional networks for efficient inference: A whitepaper
Raghuraman Krishnamoorthi
MQ
141
1,021
0
21 Jun 2018
Knowledge Distillation by On-the-Fly Native Ensemble
Knowledge Distillation by On-the-Fly Native Ensemble
Xu Lan
Xiatian Zhu
S. Gong
298
480
0
12 Jun 2018
A novel channel pruning method for deep neural network compression
A novel channel pruning method for deep neural network compression
Yiming Hu
Siyang Sun
Jianquan Li
Xingang Wang
Qingyi Gu
58
67
0
29 May 2018
PACT: Parameterized Clipping Activation for Quantized Neural Networks
PACT: Parameterized Clipping Activation for Quantized Neural Networks
Jungwook Choi
Zhuo Wang
Swagath Venkataramani
P. Chuang
Vijayalakshmi Srinivasan
K. Gopalakrishnan
MQ
75
955
0
16 May 2018
The loss landscape of overparameterized neural networks
The loss landscape of overparameterized neural networks
Y. Cooper
68
79
0
26 Apr 2018
Value-aware Quantization for Training and Inference of Neural Networks
Value-aware Quantization for Training and Inference of Neural Networks
Eunhyeok Park
S. Yoo
Peter Vajda
MQ
53
163
0
20 Apr 2018
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language
  Understanding
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding
Alex Jinpeng Wang
Amanpreet Singh
Julian Michael
Felix Hill
Omer Levy
Samuel R. Bowman
ELM
1.1K
7,201
0
20 Apr 2018
Co-teaching: Robust Training of Deep Neural Networks with Extremely
  Noisy Labels
Co-teaching: Robust Training of Deep Neural Networks with Extremely Noisy Labels
Bo Han
Quanming Yao
Xingrui Yu
Gang Niu
Miao Xu
Weihua Hu
Ivor Tsang
Masashi Sugiyama
NoLa
123
2,082
0
18 Apr 2018
Neural Compatibility Modeling with Attentive Knowledge Distillation
Neural Compatibility Modeling with Attentive Knowledge Distillation
Xuemeng Song
Fuli Feng
Xianjing Han
Xin Yang
Wen Liu
Liqiang Nie
99
145
0
17 Apr 2018
The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks
The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks
Jonathan Frankle
Michael Carbin
272
3,488
0
09 Mar 2018
Compressing Neural Networks using the Variational Information Bottleneck
Compressing Neural Networks using the Variational Information Bottleneck
Bin Dai
Chen Zhu
David Wipf
MLT
61
182
0
28 Feb 2018
Residual Dense Network for Image Super-Resolution
Residual Dense Network for Image Super-Resolution
Yulun Zhang
Yapeng Tian
Yu Kong
Bineng Zhong
Y. Fu
SupR
138
3,325
0
24 Feb 2018
Image Transformer
Image Transformer
Niki Parmar
Ashish Vaswani
Jakob Uszkoreit
Lukasz Kaiser
Noam M. Shazeer
Alexander Ku
Dustin Tran
ViT
147
1,687
0
15 Feb 2018
Model compression via distillation and quantization
Model compression via distillation and quantization
A. Polino
Razvan Pascanu
Dan Alistarh
MQ
86
732
0
15 Feb 2018
AMC: AutoML for Model Compression and Acceleration on Mobile Devices
AMC: AutoML for Model Compression and Acceleration on Mobile Devices
Yihui He
Ji Lin
Zhijian Liu
Hanrui Wang
Li Li
Song Han
100
1,349
0
10 Feb 2018
Mixed Precision Training of Convolutional Neural Networks using Integer
  Operations
Mixed Precision Training of Convolutional Neural Networks using Integer Operations
Dipankar Das
Naveen Mellempudi
Dheevatsa Mudigere
Dhiraj D. Kalamkar
Sasikanth Avancha
...
J. Corbal
N. Shustrov
R. Dubtsov
Evarist Fomenko
V. Pirogov
MQ
65
154
0
03 Feb 2018
Piggyback: Adapting a Single Network to Multiple Tasks by Learning to
  Mask Weights
Piggyback: Adapting a Single Network to Multiple Tasks by Learning to Mask Weights
Arun Mallya
Dillon Davis
Svetlana Lazebnik
CLL
64
35
0
19 Jan 2018
Faster gaze prediction with dense networks and Fisher pruning
Faster gaze prediction with dense networks and Fisher pruning
Lucas Theis
I. Korshunova
Alykhan Tejani
Ferenc Huszár
74
207
0
17 Jan 2018
Visualizing the Loss Landscape of Neural Nets
Visualizing the Loss Landscape of Neural Nets
Hao Li
Zheng Xu
Gavin Taylor
Christoph Studer
Tom Goldstein
264
1,901
0
28 Dec 2017
Quantization and Training of Neural Networks for Efficient
  Integer-Arithmetic-Only Inference
Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference
Benoit Jacob
S. Kligys
Bo Chen
Menglong Zhu
Matthew Tang
Andrew G. Howard
Hartwig Adam
Dmitry Kalenichenko
MQ
164
3,143
0
15 Dec 2017
Learning Compact Recurrent Neural Networks with Block-Term Tensor
  Decomposition
Learning Compact Recurrent Neural Networks with Block-Term Tensor Decomposition
Jinmian Ye
Linnan Wang
Guangxi Li
Di Chen
Shandian Zhe
Xinqi Chu
Zenglin Xu
98
132
0
14 Dec 2017
Towards Accurate Binary Convolutional Neural Network
Towards Accurate Binary Convolutional Neural Network
Xiaofan Lin
Cong Zhao
Wei Pan
MQ
91
649
0
30 Nov 2017
NISP: Pruning Networks using Neuron Importance Score Propagation
NISP: Pruning Networks using Neuron Importance Score Propagation
Ruichi Yu
Ang Li
Chun-Fu Chen
Jui-Hsin Lai
Vlad I. Morariu
Xintong Han
M. Gao
Ching-Yung Lin
L. Davis
74
800
0
16 Nov 2017
Apprentice: Using Knowledge Distillation Techniques To Improve
  Low-Precision Network Accuracy
Apprentice: Using Knowledge Distillation Techniques To Improve Low-Precision Network Accuracy
Asit K. Mishra
Debbie Marr
FedML
70
331
0
15 Nov 2017
Deep Rewiring: Training very sparse deep networks
Deep Rewiring: Training very sparse deep networks
G. Bellec
David Kappel
Wolfgang Maass
Robert Legenstein
BDL
189
279
0
14 Nov 2017
NeST: A Neural Network Synthesis Tool Based on a Grow-and-Prune Paradigm
NeST: A Neural Network Synthesis Tool Based on a Grow-and-Prune Paradigm
Xiaoliang Dai
Hongxu Yin
N. Jha
DD
87
239
0
06 Nov 2017
Long-term Forecasting using Higher Order Tensor RNNs
Long-term Forecasting using Higher Order Tensor RNNs
Rose Yu
Stephan Zheng
Anima Anandkumar
Yisong Yue
AI4TS
60
134
0
31 Oct 2017
Data-Free Knowledge Distillation for Deep Neural Networks
Data-Free Knowledge Distillation for Deep Neural Networks
Raphael Gontijo-Lopes
Stefano Fenu
Thad Starner
67
273
0
19 Oct 2017
Mixed Precision Training
Mixed Precision Training
Paulius Micikevicius
Sharan Narang
Jonah Alben
G. Diamos
Erich Elsen
...
Boris Ginsburg
Michael Houston
Oleksii Kuchaiev
Ganesh Venkatesh
Hao Wu
178
1,805
0
10 Oct 2017
To prune, or not to prune: exploring the efficacy of pruning for model
  compression
To prune, or not to prune: exploring the efficacy of pruning for model compression
Michael Zhu
Suyog Gupta
197
1,282
0
05 Oct 2017
N2N Learning: Network to Network Compression via Policy Gradient
  Reinforcement Learning
N2N Learning: Network to Network Compression via Policy Gradient Reinforcement Learning
A. Ashok
Nicholas Rhinehart
Fares N. Beainy
Kris Kitani
79
171
0
18 Sep 2017
WRPN: Wide Reduced-Precision Networks
WRPN: Wide Reduced-Precision Networks
Asit K. Mishra
Eriko Nurvitadhi
Jeffrey J. Cook
Debbie Marr
MQ
86
267
0
04 Sep 2017
Scalable Training of Artificial Neural Networks with Adaptive Sparse
  Connectivity inspired by Network Science
Scalable Training of Artificial Neural Networks with Adaptive Sparse Connectivity inspired by Network Science
Decebal Constantin Mocanu
Elena Mocanu
Peter Stone
Phuong H. Nguyen
M. Gibescu
A. Liotta
182
637
0
15 Jul 2017
Adversarial Dropout for Supervised and Semi-supervised Learning
Adversarial Dropout for Supervised and Semi-supervised Learning
Sungrae Park
Jun-Keon Park
Su-Jin Shin
Il-Chul Moon
GAN
81
174
0
12 Jul 2017
ShuffleNet: An Extremely Efficient Convolutional Neural Network for
  Mobile Devices
ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices
Xiangyu Zhang
Xinyu Zhou
Mengxiao Lin
Jian Sun
AI4TS
147
6,886
0
04 Jul 2017
Attention Is All You Need
Attention Is All You Need
Ashish Vaswani
Noam M. Shazeer
Niki Parmar
Jakob Uszkoreit
Llion Jones
Aidan Gomez
Lukasz Kaiser
Illia Polosukhin
3DV
803
132,454
0
12 Jun 2017
Bayesian Compression for Deep Learning
Bayesian Compression for Deep Learning
Christos Louizos
Karen Ullrich
Max Welling
UQCVBDL
195
481
0
24 May 2017
Previous
1234
Next