ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1803.03635
  4. Cited By
The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks

The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks

9 March 2018
Jonathan Frankle
Michael Carbin
ArXivPDFHTML

Papers citing "The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks"

50 / 729 papers shown
Title
A Framework for Neural Network Pruning Using Gibbs Distributions
A Framework for Neural Network Pruning Using Gibbs Distributions
Alex Labach
S. Valaee
9
5
0
08 Jun 2020
An Empirical Analysis of the Impact of Data Augmentation on Knowledge
  Distillation
An Empirical Analysis of the Impact of Data Augmentation on Knowledge Distillation
Deepan Das
Haley Massa
Abhimanyu Kulkarni
Theodoros Rekatsinas
29
18
0
06 Jun 2020
An Overview of Neural Network Compression
An Overview of Neural Network Compression
James OÑeill
AI4CE
45
98
0
05 Jun 2020
Feature Purification: How Adversarial Training Performs Robust Deep
  Learning
Feature Purification: How Adversarial Training Performs Robust Deep Learning
Zeyuan Allen-Zhu
Yuanzhi Li
MLT
AAML
39
147
0
20 May 2020
Dynamic Sparse Training: Find Efficient Sparse Network From Scratch With
  Trainable Masked Layers
Dynamic Sparse Training: Find Efficient Sparse Network From Scratch With Trainable Masked Layers
Junjie Liu
Zhe Xu
Runbin Shi
R. Cheung
Hayden Kwok-Hay So
17
119
0
14 May 2020
Data-Free Network Quantization With Adversarial Knowledge Distillation
Data-Free Network Quantization With Adversarial Knowledge Distillation
Yoojin Choi
Jihwan P. Choi
Mostafa El-Khamy
Jungwon Lee
MQ
27
119
0
08 May 2020
When BERT Plays the Lottery, All Tickets Are Winning
When BERT Plays the Lottery, All Tickets Are Winning
Sai Prasanna
Anna Rogers
Anna Rumshisky
MILM
16
186
0
01 May 2020
Pruning artificial neural networks: a way to find well-generalizing,
  high-entropy sharp minima
Pruning artificial neural networks: a way to find well-generalizing, high-entropy sharp minima
Enzo Tartaglione
Andrea Bragagnolo
Marco Grangetto
31
11
0
30 Apr 2020
Learning to Learn to Disambiguate: Meta-Learning for Few-Shot Word Sense
  Disambiguation
Learning to Learn to Disambiguate: Meta-Learning for Few-Shot Word Sense Disambiguation
Nithin Holla
Pushkar Mishra
H. Yannakoudakis
Ekaterina Shutova
37
28
0
29 Apr 2020
Random Features for Kernel Approximation: A Survey on Algorithms,
  Theory, and Beyond
Random Features for Kernel Approximation: A Survey on Algorithms, Theory, and Beyond
Fanghui Liu
Xiaolin Huang
Yudong Chen
Johan A. K. Suykens
BDL
44
172
0
23 Apr 2020
Network Adjustment: Channel Search Guided by FLOPs Utilization Ratio
Network Adjustment: Channel Search Guided by FLOPs Utilization Ratio
Zhengsu Chen
J. Niu
Lingxi Xie
Xuefeng Liu
Longhui Wei
Qi Tian
27
12
0
06 Apr 2020
CAZSL: Zero-Shot Regression for Pushing Models by Generalizing Through
  Context
CAZSL: Zero-Shot Regression for Pushing Models by Generalizing Through Context
Wenyu Zhang
Skyler Seto
Devesh K. Jha
20
5
0
26 Mar 2020
Born-Again Tree Ensembles
Born-Again Tree Ensembles
Thibaut Vidal
Toni Pacheco
Maximilian Schiffer
64
53
0
24 Mar 2020
SASL: Saliency-Adaptive Sparsity Learning for Neural Network
  Acceleration
SASL: Saliency-Adaptive Sparsity Learning for Neural Network Acceleration
Jun Shi
Jianfeng Xu
K. Tasaka
Zhibo Chen
6
25
0
12 Mar 2020
$Π-$nets: Deep Polynomial Neural Networks
Π−Π-Π−nets: Deep Polynomial Neural Networks
Grigorios G. Chrysos
Stylianos Moschoglou
Giorgos Bouritsas
Yannis Panagakis
Jiankang Deng
S. Zafeiriou
29
59
0
08 Mar 2020
Towards Practical Lottery Ticket Hypothesis for Adversarial Training
Towards Practical Lottery Ticket Hypothesis for Adversarial Training
Bai Li
Shiqi Wang
Yunhan Jia
Yantao Lu
Zhenyu Zhong
Lawrence Carin
Suman Jana
AAML
26
14
0
06 Mar 2020
Comparing Rewinding and Fine-tuning in Neural Network Pruning
Comparing Rewinding and Fine-tuning in Neural Network Pruning
Alex Renda
Jonathan Frankle
Michael Carbin
235
383
0
05 Mar 2020
Learned Threshold Pruning
Learned Threshold Pruning
K. Azarian
Yash Bhalgat
Jinwon Lee
Tijmen Blankevoort
MQ
28
38
0
28 Feb 2020
Learning in the Frequency Domain
Learning in the Frequency Domain
Kai Xu
Minghai Qin
Fei Sun
Yuhao Wang
Yen-kuang Chen
Fengbo Ren
39
395
0
27 Feb 2020
Deep Randomized Neural Networks
Deep Randomized Neural Networks
Claudio Gallicchio
Simone Scardapane
OOD
45
61
0
27 Feb 2020
Predicting Neural Network Accuracy from Weights
Predicting Neural Network Accuracy from Weights
Thomas Unterthiner
Daniel Keysers
Sylvain Gelly
Olivier Bousquet
Ilya O. Tolstikhin
30
101
0
26 Feb 2020
The Early Phase of Neural Network Training
The Early Phase of Neural Network Training
Jonathan Frankle
D. Schwab
Ari S. Morcos
21
170
0
24 Feb 2020
Neuron Shapley: Discovering the Responsible Neurons
Neuron Shapley: Discovering the Responsible Neurons
Amirata Ghorbani
James Zou
FAtt
TDI
25
108
0
23 Feb 2020
Identifying Critical Neurons in ANN Architectures using Mixed Integer
  Programming
Identifying Critical Neurons in ANN Architectures using Mixed Integer Programming
M. Elaraby
Guy Wolf
Margarida Carvalho
26
5
0
17 Feb 2020
A study of local optima for learning feature interactions using neural
  networks
A study of local optima for learning feature interactions using neural networks
Yangzi Guo
Adrian Barbu
16
1
0
11 Feb 2020
PixelHop++: A Small Successive-Subspace-Learning-Based (SSL-based) Model
  for Image Classification
PixelHop++: A Small Successive-Subspace-Learning-Based (SSL-based) Model for Image Classification
Yueru Chen
Mozhdeh Rouhsedaghat
Suya You
Raghuveer Rao
C.-C. Jay Kuo
14
68
0
08 Feb 2020
Activation Density driven Energy-Efficient Pruning in Training
Activation Density driven Energy-Efficient Pruning in Training
Timothy Foldy-Porto
Yeshwanth Venkatesha
Priyadarshini Panda
10
4
0
07 Feb 2020
Proving the Lottery Ticket Hypothesis: Pruning is All You Need
Proving the Lottery Ticket Hypothesis: Pruning is All You Need
Eran Malach
Gilad Yehudai
Shai Shalev-Shwartz
Ohad Shamir
64
272
0
03 Feb 2020
MEMO: A Deep Network for Flexible Combination of Episodic Memories
MEMO: A Deep Network for Flexible Combination of Episodic Memories
Andrea Banino
Adria Puigdomenech Badia
Raphael Köster
Martin Chadwick
V. Zambaldi
Demis Hassabis
Caswell Barry
M. Botvinick
D. Kumaran
Charles Blundell
KELM
26
33
0
29 Jan 2020
Progressive Local Filter Pruning for Image Retrieval Acceleration
Progressive Local Filter Pruning for Image Retrieval Acceleration
Xiaodong Wang
Zhedong Zheng
Yang He
Fei Yan
Zhi-qiang Zeng
Yi Yang
33
34
0
24 Jan 2020
Filter Sketch for Network Pruning
Filter Sketch for Network Pruning
Mingbao Lin
Liujuan Cao
Shaojie Li
QiXiang Ye
Yonghong Tian
Jianzhuang Liu
Q. Tian
Rongrong Ji
CLIP
3DPC
31
82
0
23 Jan 2020
Convolutional Neural Networks as a Model of the Visual System: Past,
  Present, and Future
Convolutional Neural Networks as a Model of the Visual System: Past, Present, and Future
Grace W. Lindsay
MedIm
35
424
0
20 Jan 2020
Least squares binary quantization of neural networks
Least squares binary quantization of neural networks
Hadi Pouransari
Zhucheng Tu
Oncel Tuzel
MQ
17
32
0
09 Jan 2020
Sparse Weight Activation Training
Sparse Weight Activation Training
Md Aamir Raihan
Tor M. Aamodt
34
73
0
07 Jan 2020
Lossless Compression of Deep Neural Networks
Lossless Compression of Deep Neural Networks
Thiago Serra
Abhinav Kumar
Srikumar Ramalingam
24
56
0
01 Jan 2020
Mixed-Precision Quantized Neural Network with Progressively Decreasing
  Bitwidth For Image Classification and Object Detection
Mixed-Precision Quantized Neural Network with Progressively Decreasing Bitwidth For Image Classification and Object Detection
Tianshu Chu
Qin Luo
Jie Yang
Xiaolin Huang
MQ
24
6
0
29 Dec 2019
Landscape Connectivity and Dropout Stability of SGD Solutions for
  Over-parameterized Neural Networks
Landscape Connectivity and Dropout Stability of SGD Solutions for Over-parameterized Neural Networks
A. Shevchenko
Marco Mondelli
27
37
0
20 Dec 2019
Optimization for deep learning: theory and algorithms
Optimization for deep learning: theory and algorithms
Ruoyu Sun
ODL
27
168
0
19 Dec 2019
QKD: Quantization-aware Knowledge Distillation
QKD: Quantization-aware Knowledge Distillation
Jangho Kim
Yash Bhalgat
Jinwon Lee
Chirag I. Patel
Nojun Kwak
MQ
21
63
0
28 Nov 2019
Learning Sparse Sharing Architectures for Multiple Tasks
Learning Sparse Sharing Architectures for Multiple Tasks
Tianxiang Sun
Yunfan Shao
Xiaonan Li
Pengfei Liu
Hang Yan
Xipeng Qiu
Xuanjing Huang
MoE
30
128
0
12 Nov 2019
Blockwise Self-Attention for Long Document Understanding
Blockwise Self-Attention for Long Document Understanding
J. Qiu
Hao Ma
Omer Levy
Scott Yih
Sinong Wang
Jie Tang
11
252
0
07 Nov 2019
Active Subspace of Neural Networks: Structural Analysis and Universal
  Attacks
Active Subspace of Neural Networks: Structural Analysis and Universal Attacks
Chunfeng Cui
Kaiqi Zhang
Talgat Daulbaev
Julia Gusak
Ivan Oseledets
Zheng-Wei Zhang
AAML
32
25
0
29 Oct 2019
Explainable Artificial Intelligence (XAI): Concepts, Taxonomies,
  Opportunities and Challenges toward Responsible AI
Explainable Artificial Intelligence (XAI): Concepts, Taxonomies, Opportunities and Challenges toward Responsible AI
Alejandro Barredo Arrieta
Natalia Díaz Rodríguez
Javier Del Ser
Adrien Bennetot
Siham Tabik
...
S. Gil-Lopez
Daniel Molina
Richard Benjamins
Raja Chatila
Francisco Herrera
XAI
41
6,119
0
22 Oct 2019
Spiking neural networks trained with backpropagation for low power
  neuromorphic implementation of voice activity detection
Spiking neural networks trained with backpropagation for low power neuromorphic implementation of voice activity detection
Flavio Martinelli
Giorgia Dellaferrera
Pablo Mainar
Milos Cernak
19
29
0
22 Oct 2019
Model Compression with Two-stage Multi-teacher Knowledge Distillation
  for Web Question Answering System
Model Compression with Two-stage Multi-teacher Knowledge Distillation for Web Question Answering System
Ze Yang
Linjun Shou
Ming Gong
Wutao Lin
Daxin Jiang
28
92
0
18 Oct 2019
ConfusionFlow: A model-agnostic visualization for temporal analysis of
  classifier confusion
ConfusionFlow: A model-agnostic visualization for temporal analysis of classifier confusion
A. Hinterreiter
Peter Ruch
Holger Stitz
Martin Ennemoser
J. Bernard
Hendrik Strobelt
M. Streit
24
43
0
02 Oct 2019
How does topology influence gradient propagation and model performance
  of deep networks with DenseNet-type skip connections?
How does topology influence gradient propagation and model performance of deep networks with DenseNet-type skip connections?
Kartikeya Bhardwaj
Guihong Li
R. Marculescu
38
1
0
02 Oct 2019
Optimizing Speech Recognition For The Edge
Optimizing Speech Recognition For The Edge
Yuan Shangguan
Jian Li
Qiao Liang
R. Álvarez
Ian McGraw
28
64
0
26 Sep 2019
Model Pruning Enables Efficient Federated Learning on Edge Devices
Model Pruning Enables Efficient Federated Learning on Edge Devices
Yuang Jiang
Shiqiang Wang
Victor Valls
Bongjun Ko
Wei-Han Lee
Kin K. Leung
Leandros Tassiulas
38
447
0
26 Sep 2019
Mixed Dimension Embeddings with Application to Memory-Efficient
  Recommendation Systems
Mixed Dimension Embeddings with Application to Memory-Efficient Recommendation Systems
Antonio A. Ginart
Maxim Naumov
Dheevatsa Mudigere
Jiyan Yang
James Zou
22
99
0
25 Sep 2019
Previous
123...131415
Next