Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1312.6184
Cited By
Do Deep Nets Really Need to be Deep?
21 December 2013
Lei Jimmy Ba
R. Caruana
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Do Deep Nets Really Need to be Deep?"
50 / 379 papers shown
Title
The Description Length of Deep Learning Models
Léonard Blier
Yann Ollivier
32
97
0
20 Feb 2018
ThUnderVolt: Enabling Aggressive Voltage Underscaling and Timing Error Resilience for Energy Efficient Deep Neural Network Accelerators
Jeff Zhang
Kartheek Rangineni
Zahra Ghodsi
S. Garg
36
118
0
11 Feb 2018
Analyzing and Mitigating the Impact of Permanent Faults on a Systolic Array Based Neural Network Accelerator
Jeff Zhang
Tianyu Gu
K. Basu
S. Garg
14
134
0
11 Feb 2018
Few-shot learning of neural networks from scratch by pseudo example optimization
Akisato Kimura
Zoubin Ghahramani
Koh Takeuchi
Tomoharu Iwata
N. Ueda
35
52
0
08 Feb 2018
Digital Watermarking for Deep Neural Networks
Yuki Nagai
Yusuke Uchida
S. Sakazawa
Shiníchi Satoh
WIGM
31
143
0
06 Feb 2018
Recovering from Random Pruning: On the Plasticity of Deep Convolutional Neural Networks
Deepak Mittal
S. Bhardwaj
Mitesh M. Khapra
Balaraman Ravindran
VLM
36
65
0
31 Jan 2018
Focus: Querying Large Video Datasets with Low Latency and Low Cost
Kevin Hsieh
Ganesh Ananthanarayanan
P. Bodík
P. Bahl
Matthai Philipose
Phillip B. Gibbons
O. Mutlu
27
275
0
10 Jan 2018
Improving the Adversarial Robustness and Interpretability of Deep Neural Networks by Regularizing their Input Gradients
A. Ross
Finale Doshi-Velez
AAML
37
676
0
26 Nov 2017
Knowledge Concentration: Learning 100K Object Classifiers in a Single CNN
J. Gao
Zijian
Guo
Zerui Li
Ram Nevatia
VLM
26
20
0
21 Nov 2017
Moonshine: Distilling with Cheap Convolutions
Elliot J. Crowley
Gavia Gray
Amos Storkey
27
120
0
07 Nov 2017
Towards Effective Low-bitwidth Convolutional Neural Networks
Bohan Zhuang
Chunhua Shen
Mingkui Tan
Lingqiao Liu
Ian Reid
MQ
33
231
0
01 Nov 2017
Interpretation of Neural Networks is Fragile
Amirata Ghorbani
Abubakar Abid
James Zou
FAtt
AAML
80
857
0
29 Oct 2017
Knowledge Projection for Deep Neural Networks
Zhi Zhang
G. Ning
Zhihai He
38
15
0
26 Oct 2017
Trace norm regularization and faster inference for embedded speech recognition RNNs
Markus Kliegl
Siddharth Goyal
Kexin Zhao
Kavya Srinet
M. Shoeybi
40
8
0
25 Oct 2017
A Survey of Model Compression and Acceleration for Deep Neural Networks
Yu Cheng
Duo Wang
Pan Zhou
Zhang Tao
40
1,087
0
23 Oct 2017
Data-Free Knowledge Distillation for Deep Neural Networks
Raphael Gontijo-Lopes
Stefano Fenu
Thad Starner
25
270
0
19 Oct 2017
Distill-and-Compare: Auditing Black-Box Models Using Transparent Model Distillation
S. Tan
R. Caruana
Giles Hooker
Yin Lou
MLAU
20
180
0
17 Oct 2017
Deep Learning Techniques for Music Generation -- A Survey
Jean-Pierre Briot
Gaëtan Hadjeres
F. Pachet
MGen
37
297
0
05 Sep 2017
Sequence Prediction with Neural Segmental Models
Hao Tang
29
2
0
05 Sep 2017
Interpretability via Model Extraction
Osbert Bastani
Carolyn Kim
Hamsa Bastani
FAtt
21
129
0
29 Jun 2017
Deep Mutual Learning
Ying Zhang
Tao Xiang
Timothy M. Hospedales
Huchuan Lu
FedML
37
1,639
0
01 Jun 2017
Iterative Machine Teaching
Weiyang Liu
Bo Dai
Ahmad Humayun
C. Tay
Chen Yu
Linda B. Smith
James M. Rehg
Le Song
26
141
0
30 May 2017
Kronecker Recurrent Units
C. Jose
Moustapha Cissé
F. Fleuret
ODL
24
45
0
29 May 2017
Bayesian Compression for Deep Learning
Christos Louizos
Karen Ullrich
Max Welling
UQCV
BDL
23
479
0
24 May 2017
Interpreting Blackbox Models via Model Extraction
Osbert Bastani
Carolyn Kim
Hamsa Bastani
FAtt
35
170
0
23 May 2017
Compressing Recurrent Neural Network with Tensor Train
Andros Tjandra
S. Sakti
Satoshi Nakamura
31
109
0
23 May 2017
Hardware-Software Codesign of Accurate, Multiplier-free Deep Neural Networks
Hokchhay Tann
S. Hashemi
Iris Bahar
Sherief Reda
MQ
27
74
0
11 May 2017
Knowledge-Guided Deep Fractal Neural Networks for Human Pose Estimation
G. Ning
Zhi Zhang
Zhiquan He
GAN
29
169
0
05 May 2017
A Teacher-Student Framework for Zero-Resource Neural Machine Translation
Yun Chen
Yang Liu
Yong Cheng
V. Li
35
147
0
02 May 2017
The loss surface of deep and wide neural networks
Quynh N. Nguyen
Matthias Hein
ODL
51
283
0
26 Apr 2017
Deep Architectures for Modulation Recognition
Nathan E. West
Tim O'Shea
27
401
0
27 Mar 2017
Predicting Deeper into the Future of Semantic Segmentation
Pauline Luc
Natalia Neverova
Camille Couprie
Jakob Verbeek
Yann LeCun
23
242
0
22 Mar 2017
Knowledge distillation using unlabeled mismatched images
Mandar M. Kulkarni
Kalpesh Patil
Shirish S. Karande
48
16
0
21 Mar 2017
Chain-NN: An Energy-Efficient 1D Chain Architecture for Accelerating Deep Convolutional Neural Networks
Shihao Wang
Dajiang Zhou
Xushen Han
T. Yoshimura
3DV
11
51
0
04 Mar 2017
Neurogenesis-Inspired Dictionary Learning: Online Model Adaption in a Changing World
S. Garg
Irina Rish
Guillermo Cecchi
A. Lozano
OffRL
CLL
33
6
0
22 Jan 2017
Learning From Noisy Large-Scale Datasets With Minimal Supervision
Andreas Veit
N. Alldrin
Gal Chechik
Ivan Krasin
Abhinav Gupta
Serge J. Belongie
34
476
0
06 Jan 2017
Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer
Sergey Zagoruyko
N. Komodakis
19
2,552
0
12 Dec 2016
In Teacher We Trust: Learning Compressed Models for Pedestrian Detection
Jonathan Shen
Noranart Vesdapunt
Vishnu Boddeti
Kris Kitani
19
29
0
01 Dec 2016
Patient-Driven Privacy Control through Generalized Distillation
Z. Berkay Celik
David Lopez-Paz
Patrick McDaniel
22
18
0
26 Nov 2016
Training Sparse Neural Networks
Suraj Srinivas
Akshayvarun Subramanya
R. Venkatesh Babu
24
204
0
21 Nov 2016
Fast Video Classification via Adaptive Cascading of Deep Models
Haichen Shen
Seungyeop Han
Matthai Philipose
Arvind Krishnamurthy
34
78
0
20 Nov 2016
Deep Model Compression: Distilling Knowledge from Noisy Teachers
Bharat Bhusan Sau
V. Balasubramanian
20
181
0
30 Oct 2016
Small-footprint Highway Deep Neural Networks for Speech Recognition
Liang Lu
Steve Renals
32
15
0
18 Oct 2016
Distilling an Ensemble of Greedy Dependency Parsers into One MST Parser
A. Kuncoro
Miguel Ballesteros
Lingpeng Kong
Chris Dyer
Noah A. Smith
MoE
25
77
0
24 Sep 2016
Why does deep and cheap learning work so well?
Henry W. Lin
Max Tegmark
David Rolnick
40
603
0
29 Aug 2016
Lets keep it simple, Using simple architectures to outperform deeper and more complex architectures
S. H. HasanPour
Mohammad Rouhani
Mohsen Fayyaz
Mohammad Sabokrou
20
119
0
22 Aug 2016
Knowledge Distillation for Small-footprint Highway Networks
Liang Lu
Michelle Guo
Steve Renals
24
73
0
02 Aug 2016
Supervised learning based on temporal coding in spiking neural networks
Hesham Mostafa
31
350
0
27 Jun 2016
Sequence-Level Knowledge Distillation
Yoon Kim
Alexander M. Rush
47
1,099
0
25 Jun 2016
Active Long Term Memory Networks
Tommaso Furlanello
Jiaping Zhao
Andrew M. Saxe
Laurent Itti
B. Tjan
KELM
CLL
32
41
0
07 Jun 2016
Previous
1
2
3
4
5
6
7
8
Next