Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1908.08520
Cited By
Adversarial-Based Knowledge Distillation for Multi-Model Ensemble and Noisy Data Refinement
22 August 2019
Zhiqiang Shen
Zhankui He
Wanyun Cui
Jiahui Yu
Yutong Zheng
Chenchen Zhu
Marios Savvides
AAML
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Adversarial-Based Knowledge Distillation for Multi-Model Ensemble and Noisy Data Refinement"
5 / 5 papers shown
Title
Knowledge Distillation by On-the-Fly Native Ensemble
Xu Lan
Xiatian Zhu
S. Gong
203
473
0
12 Jun 2018
Large scale distributed neural network training through online distillation
Rohan Anil
Gabriel Pereyra
Alexandre Passos
Róbert Ormándi
George E. Dahl
Geoffrey E. Hinton
FedML
278
404
0
09 Apr 2018
Image Generation from Scene Graphs
Justin Johnson
Agrim Gupta
Li Fei-Fei
GNN
223
815
0
04 Apr 2018
Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results
Antti Tarvainen
Harri Valpola
OOD
MoMe
264
1,275
0
06 Mar 2017
Simple and Scalable Predictive Uncertainty Estimation using Deep Ensembles
Balaji Lakshminarayanan
Alexander Pritzel
Charles Blundell
UQCV
BDL
276
5,675
0
05 Dec 2016
1