ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1312.6184
  4. Cited By
Do Deep Nets Really Need to be Deep?

Do Deep Nets Really Need to be Deep?

21 December 2013
Lei Jimmy Ba
R. Caruana
ArXivPDFHTML

Papers citing "Do Deep Nets Really Need to be Deep?"

50 / 379 papers shown
Title
The Description Length of Deep Learning Models
The Description Length of Deep Learning Models
Léonard Blier
Yann Ollivier
32
97
0
20 Feb 2018
ThUnderVolt: Enabling Aggressive Voltage Underscaling and Timing Error
  Resilience for Energy Efficient Deep Neural Network Accelerators
ThUnderVolt: Enabling Aggressive Voltage Underscaling and Timing Error Resilience for Energy Efficient Deep Neural Network Accelerators
Jeff Zhang
Kartheek Rangineni
Zahra Ghodsi
S. Garg
36
118
0
11 Feb 2018
Analyzing and Mitigating the Impact of Permanent Faults on a Systolic
  Array Based Neural Network Accelerator
Analyzing and Mitigating the Impact of Permanent Faults on a Systolic Array Based Neural Network Accelerator
Jeff Zhang
Tianyu Gu
K. Basu
S. Garg
14
134
0
11 Feb 2018
Few-shot learning of neural networks from scratch by pseudo example
  optimization
Few-shot learning of neural networks from scratch by pseudo example optimization
Akisato Kimura
Zoubin Ghahramani
Koh Takeuchi
Tomoharu Iwata
N. Ueda
35
52
0
08 Feb 2018
Digital Watermarking for Deep Neural Networks
Digital Watermarking for Deep Neural Networks
Yuki Nagai
Yusuke Uchida
S. Sakazawa
Shiníchi Satoh
WIGM
31
143
0
06 Feb 2018
Recovering from Random Pruning: On the Plasticity of Deep Convolutional
  Neural Networks
Recovering from Random Pruning: On the Plasticity of Deep Convolutional Neural Networks
Deepak Mittal
S. Bhardwaj
Mitesh M. Khapra
Balaraman Ravindran
VLM
36
65
0
31 Jan 2018
Focus: Querying Large Video Datasets with Low Latency and Low Cost
Focus: Querying Large Video Datasets with Low Latency and Low Cost
Kevin Hsieh
Ganesh Ananthanarayanan
P. Bodík
P. Bahl
Matthai Philipose
Phillip B. Gibbons
O. Mutlu
27
275
0
10 Jan 2018
Improving the Adversarial Robustness and Interpretability of Deep Neural
  Networks by Regularizing their Input Gradients
Improving the Adversarial Robustness and Interpretability of Deep Neural Networks by Regularizing their Input Gradients
A. Ross
Finale Doshi-Velez
AAML
37
676
0
26 Nov 2017
Knowledge Concentration: Learning 100K Object Classifiers in a Single
  CNN
Knowledge Concentration: Learning 100K Object Classifiers in a Single CNN
J. Gao
Zijian
Guo
Zerui Li
Ram Nevatia
VLM
26
20
0
21 Nov 2017
Moonshine: Distilling with Cheap Convolutions
Moonshine: Distilling with Cheap Convolutions
Elliot J. Crowley
Gavia Gray
Amos Storkey
27
120
0
07 Nov 2017
Towards Effective Low-bitwidth Convolutional Neural Networks
Towards Effective Low-bitwidth Convolutional Neural Networks
Bohan Zhuang
Chunhua Shen
Mingkui Tan
Lingqiao Liu
Ian Reid
MQ
33
231
0
01 Nov 2017
Interpretation of Neural Networks is Fragile
Interpretation of Neural Networks is Fragile
Amirata Ghorbani
Abubakar Abid
James Zou
FAtt
AAML
80
857
0
29 Oct 2017
Knowledge Projection for Deep Neural Networks
Knowledge Projection for Deep Neural Networks
Zhi Zhang
G. Ning
Zhihai He
38
15
0
26 Oct 2017
Trace norm regularization and faster inference for embedded speech
  recognition RNNs
Trace norm regularization and faster inference for embedded speech recognition RNNs
Markus Kliegl
Siddharth Goyal
Kexin Zhao
Kavya Srinet
M. Shoeybi
40
8
0
25 Oct 2017
A Survey of Model Compression and Acceleration for Deep Neural Networks
A Survey of Model Compression and Acceleration for Deep Neural Networks
Yu Cheng
Duo Wang
Pan Zhou
Zhang Tao
40
1,087
0
23 Oct 2017
Data-Free Knowledge Distillation for Deep Neural Networks
Data-Free Knowledge Distillation for Deep Neural Networks
Raphael Gontijo-Lopes
Stefano Fenu
Thad Starner
25
270
0
19 Oct 2017
Distill-and-Compare: Auditing Black-Box Models Using Transparent Model
  Distillation
Distill-and-Compare: Auditing Black-Box Models Using Transparent Model Distillation
S. Tan
R. Caruana
Giles Hooker
Yin Lou
MLAU
20
180
0
17 Oct 2017
Deep Learning Techniques for Music Generation -- A Survey
Deep Learning Techniques for Music Generation -- A Survey
Jean-Pierre Briot
Gaëtan Hadjeres
F. Pachet
MGen
37
297
0
05 Sep 2017
Sequence Prediction with Neural Segmental Models
Sequence Prediction with Neural Segmental Models
Hao Tang
29
2
0
05 Sep 2017
Interpretability via Model Extraction
Interpretability via Model Extraction
Osbert Bastani
Carolyn Kim
Hamsa Bastani
FAtt
21
129
0
29 Jun 2017
Deep Mutual Learning
Deep Mutual Learning
Ying Zhang
Tao Xiang
Timothy M. Hospedales
Huchuan Lu
FedML
37
1,639
0
01 Jun 2017
Iterative Machine Teaching
Iterative Machine Teaching
Weiyang Liu
Bo Dai
Ahmad Humayun
C. Tay
Chen Yu
Linda B. Smith
James M. Rehg
Le Song
26
141
0
30 May 2017
Kronecker Recurrent Units
Kronecker Recurrent Units
C. Jose
Moustapha Cissé
F. Fleuret
ODL
24
45
0
29 May 2017
Bayesian Compression for Deep Learning
Bayesian Compression for Deep Learning
Christos Louizos
Karen Ullrich
Max Welling
UQCV
BDL
23
479
0
24 May 2017
Interpreting Blackbox Models via Model Extraction
Interpreting Blackbox Models via Model Extraction
Osbert Bastani
Carolyn Kim
Hamsa Bastani
FAtt
35
170
0
23 May 2017
Compressing Recurrent Neural Network with Tensor Train
Compressing Recurrent Neural Network with Tensor Train
Andros Tjandra
S. Sakti
Satoshi Nakamura
31
109
0
23 May 2017
Hardware-Software Codesign of Accurate, Multiplier-free Deep Neural
  Networks
Hardware-Software Codesign of Accurate, Multiplier-free Deep Neural Networks
Hokchhay Tann
S. Hashemi
Iris Bahar
Sherief Reda
MQ
27
74
0
11 May 2017
Knowledge-Guided Deep Fractal Neural Networks for Human Pose Estimation
Knowledge-Guided Deep Fractal Neural Networks for Human Pose Estimation
G. Ning
Zhi Zhang
Zhiquan He
GAN
29
169
0
05 May 2017
A Teacher-Student Framework for Zero-Resource Neural Machine Translation
A Teacher-Student Framework for Zero-Resource Neural Machine Translation
Yun Chen
Yang Liu
Yong Cheng
V. Li
35
147
0
02 May 2017
The loss surface of deep and wide neural networks
The loss surface of deep and wide neural networks
Quynh N. Nguyen
Matthias Hein
ODL
51
283
0
26 Apr 2017
Deep Architectures for Modulation Recognition
Deep Architectures for Modulation Recognition
Nathan E. West
Tim O'Shea
27
401
0
27 Mar 2017
Predicting Deeper into the Future of Semantic Segmentation
Predicting Deeper into the Future of Semantic Segmentation
Pauline Luc
Natalia Neverova
Camille Couprie
Jakob Verbeek
Yann LeCun
23
242
0
22 Mar 2017
Knowledge distillation using unlabeled mismatched images
Knowledge distillation using unlabeled mismatched images
Mandar M. Kulkarni
Kalpesh Patil
Shirish S. Karande
48
16
0
21 Mar 2017
Chain-NN: An Energy-Efficient 1D Chain Architecture for Accelerating
  Deep Convolutional Neural Networks
Chain-NN: An Energy-Efficient 1D Chain Architecture for Accelerating Deep Convolutional Neural Networks
Shihao Wang
Dajiang Zhou
Xushen Han
T. Yoshimura
3DV
11
51
0
04 Mar 2017
Neurogenesis-Inspired Dictionary Learning: Online Model Adaption in a
  Changing World
Neurogenesis-Inspired Dictionary Learning: Online Model Adaption in a Changing World
S. Garg
Irina Rish
Guillermo Cecchi
A. Lozano
OffRL
CLL
33
6
0
22 Jan 2017
Learning From Noisy Large-Scale Datasets With Minimal Supervision
Learning From Noisy Large-Scale Datasets With Minimal Supervision
Andreas Veit
N. Alldrin
Gal Chechik
Ivan Krasin
Abhinav Gupta
Serge J. Belongie
34
476
0
06 Jan 2017
Paying More Attention to Attention: Improving the Performance of
  Convolutional Neural Networks via Attention Transfer
Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer
Sergey Zagoruyko
N. Komodakis
19
2,552
0
12 Dec 2016
In Teacher We Trust: Learning Compressed Models for Pedestrian Detection
In Teacher We Trust: Learning Compressed Models for Pedestrian Detection
Jonathan Shen
Noranart Vesdapunt
Vishnu Boddeti
Kris Kitani
19
29
0
01 Dec 2016
Patient-Driven Privacy Control through Generalized Distillation
Patient-Driven Privacy Control through Generalized Distillation
Z. Berkay Celik
David Lopez-Paz
Patrick McDaniel
22
18
0
26 Nov 2016
Training Sparse Neural Networks
Training Sparse Neural Networks
Suraj Srinivas
Akshayvarun Subramanya
R. Venkatesh Babu
24
204
0
21 Nov 2016
Fast Video Classification via Adaptive Cascading of Deep Models
Fast Video Classification via Adaptive Cascading of Deep Models
Haichen Shen
Seungyeop Han
Matthai Philipose
Arvind Krishnamurthy
34
78
0
20 Nov 2016
Deep Model Compression: Distilling Knowledge from Noisy Teachers
Deep Model Compression: Distilling Knowledge from Noisy Teachers
Bharat Bhusan Sau
V. Balasubramanian
20
181
0
30 Oct 2016
Small-footprint Highway Deep Neural Networks for Speech Recognition
Small-footprint Highway Deep Neural Networks for Speech Recognition
Liang Lu
Steve Renals
32
15
0
18 Oct 2016
Distilling an Ensemble of Greedy Dependency Parsers into One MST Parser
Distilling an Ensemble of Greedy Dependency Parsers into One MST Parser
A. Kuncoro
Miguel Ballesteros
Lingpeng Kong
Chris Dyer
Noah A. Smith
MoE
25
77
0
24 Sep 2016
Why does deep and cheap learning work so well?
Why does deep and cheap learning work so well?
Henry W. Lin
Max Tegmark
David Rolnick
40
603
0
29 Aug 2016
Lets keep it simple, Using simple architectures to outperform deeper and
  more complex architectures
Lets keep it simple, Using simple architectures to outperform deeper and more complex architectures
S. H. HasanPour
Mohammad Rouhani
Mohsen Fayyaz
Mohammad Sabokrou
20
119
0
22 Aug 2016
Knowledge Distillation for Small-footprint Highway Networks
Knowledge Distillation for Small-footprint Highway Networks
Liang Lu
Michelle Guo
Steve Renals
24
73
0
02 Aug 2016
Supervised learning based on temporal coding in spiking neural networks
Supervised learning based on temporal coding in spiking neural networks
Hesham Mostafa
31
350
0
27 Jun 2016
Sequence-Level Knowledge Distillation
Sequence-Level Knowledge Distillation
Yoon Kim
Alexander M. Rush
47
1,099
0
25 Jun 2016
Active Long Term Memory Networks
Active Long Term Memory Networks
Tommaso Furlanello
Jiaping Zhao
Andrew M. Saxe
Laurent Itti
B. Tjan
KELM
CLL
32
41
0
07 Jun 2016
Previous
12345678
Next