Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1612.03928
Cited By
Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer
12 December 2016
Sergey Zagoruyko
N. Komodakis
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer"
50 / 1,157 papers shown
Title
Highlight Every Step: Knowledge Distillation via Collaborative Teaching
Haoran Zhao
Xin Sun
Junyu Dong
Changrui Chen
Zihe Dong
27
57
0
23 Jul 2019
Bringing Giant Neural Networks Down to Earth with Unlabeled Data
Yehui Tang
Shan You
Chang Xu
Boxin Shi
Chao Xu
24
11
0
13 Jul 2019
Justifying Diagnosis Decisions by Deep Neural Networks
Graham Spinks
Marie-Francine Moens
37
13
0
12 Jul 2019
Distill-2MD-MTL: Data Distillation based on Multi-Dataset Multi-Domain Multi-Task Frame Work to Solve Face Related Tasksks, Multi Task Learning, Semi-Supervised Learning
Sepidehsadat Hosseini
M. Shabani
N. Cho
CVBM
36
3
0
08 Jul 2019
Understanding More about Human and Machine Attention in Deep Neural Networks
Qiuxia Lai
Salman Khan
Wenguan Wang
Jianbing Shen
Hanqiu Sun
Ling Shao
HAI
XAI
24
7
0
20 Jun 2019
GAN-Knowledge Distillation for one-stage Object Detection
Wanwei Wang
Jin ke Yu Fan Zong
ObjD
22
28
0
20 Jun 2019
A One-step Pruning-recovery Framework for Acceleration of Convolutional Neural Networks
Dong Wang
Lei Zhou
Xiao Bai
Jun Zhou
14
2
0
18 Jun 2019
BlockSwap: Fisher-guided Block Substitution for Network Compression on a Budget
Jack Turner
Elliot J. Crowley
Michael F. P. O'Boyle
Amos Storkey
Gavia Gray
22
37
0
10 Jun 2019
Distilling Object Detectors with Fine-grained Feature Imitation
Tao Wang
Li-xin Yuan
Xiaopeng Zhang
Jiashi Feng
ObjD
13
378
0
09 Jun 2019
Separable Layers Enable Structured Efficient Linear Substitutions
Gavia Gray
Elliot J. Crowley
Amos Storkey
30
3
0
03 Jun 2019
Deeply-supervised Knowledge Synergy
Dawei Sun
Anbang Yao
Aojun Zhou
Hao Zhao
12
63
0
03 Jun 2019
SCAN: A Scalable Neural Networks Framework Towards Compact and Efficient Models
Linfeng Zhang
Zhanhong Tan
Jiebo Song
Jingwei Chen
Chenglong Bao
Kaisheng Ma
16
71
0
27 May 2019
ShrinkTeaNet: Million-scale Lightweight Face Recognition via Shrinking Teacher-Student Networks
C. Duong
Khoa Luu
Kha Gia Quach
Ngan Le
CVBM
23
39
0
25 May 2019
Zero-shot Knowledge Transfer via Adversarial Belief Matching
P. Micaelli
Amos Storkey
19
228
0
23 May 2019
Network Pruning via Transformable Architecture Search
Xuanyi Dong
Yi Yang
3DPC
23
240
0
23 May 2019
Be Your Own Teacher: Improve the Performance of Convolutional Neural Networks via Self Distillation
Linfeng Zhang
Jiebo Song
Anni Gao
Jingwei Chen
Chenglong Bao
Kaisheng Ma
FedML
27
846
0
17 May 2019
Dream Distillation: A Data-Independent Model Compression Framework
Kartikeya Bhardwaj
Naveen Suda
R. Marculescu
DD
27
54
0
17 May 2019
Learning What and Where to Transfer
Yunhun Jang
Hankook Lee
Sung Ju Hwang
Jinwoo Shin
22
148
0
15 May 2019
Triplet Distillation for Deep Face Recognition
Yushu Feng
Huan Wang
Daniel T. Yi
Roland Hu
CVBM
11
45
0
11 May 2019
Creating Lightweight Object Detectors with Model Compression for Deployment on Edge Devices
Yiwu Yao
Weiqiang Yang
Haoqi Zhu
33
0
0
06 May 2019
Omni-Scale Feature Learning for Person Re-Identification
Kaiyang Zhou
Yongxin Yang
Andrea Cavallaro
Tao Xiang
30
820
0
02 May 2019
Knowledge Distillation via Route Constrained Optimization
Xiao Jin
Baoyun Peng
Yichao Wu
Yu Liu
Jiaheng Liu
Ding Liang
Junjie Yan
Xiaolin Hu
20
169
0
19 Apr 2019
Feature Fusion for Online Mutual Knowledge Distillation
Jangho Kim
Minsung Hyun
Inseop Chung
Nojun Kwak
FedML
26
91
0
19 Apr 2019
Biphasic Learning of GANs for High-Resolution Image-to-Image Translation
Jie Cao
Huaibo Huang
Yi Li
Jingtuo Liu
Ran He
Zhenan Sun
GAN
26
4
0
14 Apr 2019
Improved training of binary networks for human pose estimation and image recognition
Adrian Bulat
Georgios Tzimiropoulos
Jean Kossaifi
Maja Pantic
MQ
29
47
0
11 Apr 2019
Variational Information Distillation for Knowledge Transfer
Sungsoo Ahn
S. Hu
Andreas C. Damianou
Neil D. Lawrence
Zhenwen Dai
58
609
0
11 Apr 2019
Knowledge Squeezed Adversarial Network Compression
Changyong Shu
Li Peng
Xie Yuan
Yanyun Qu
Longquan Dai
Lizhuang Ma
GAN
45
11
0
10 Apr 2019
Relational Knowledge Distillation
Wonpyo Park
Dongju Kim
Yan Lu
Minsu Cho
16
1,387
0
10 Apr 2019
Learning Metrics from Teachers: Compact Networks for Image Embedding
Lu Yu
V. O. Yazici
Xialei Liu
Joost van de Weijer
Yongmei Cheng
Arnau Ramisa
11
105
0
07 Apr 2019
Self-supervised Spatio-temporal Representation Learning for Videos by Predicting Motion and Appearance Statistics
Jiangliu Wang
Jianbo Jiao
Linchao Bao
Shengfeng He
Yunhui Liu
Wen Liu
SSL
24
204
0
07 Apr 2019
Attention Distillation for Learning Video Representations
Miao Liu
Xin Chen
Yun C. Zhang
Yin Li
James M. Rehg
19
2
0
05 Apr 2019
A Comprehensive Overhaul of Feature Distillation
Byeongho Heo
Jeesoo Kim
Sangdoo Yun
Hyojin Park
Nojun Kwak
J. Choi
19
569
0
03 Apr 2019
Correlation Congruence for Knowledge Distillation
Baoyun Peng
Xiao Jin
Jiaheng Liu
Shunfeng Zhou
Yichao Wu
Yu Liu
Dongsheng Li
Zhaoning Zhang
63
507
0
03 Apr 2019
Improving Object Detection with Inverted Attention
Zeyi Huang
Wei Ke
Dong Huang
29
32
0
28 Mar 2019
Training Quantized Neural Networks with a Full-precision Auxiliary Module
Bohan Zhuang
Lingqiao Liu
Mingkui Tan
Chunhua Shen
Ian Reid
MQ
32
62
0
27 Mar 2019
Towards Optimal Structured CNN Pruning via Generative Adversarial Learning
Shaohui Lin
Rongrong Ji
Chenqian Yan
Baochang Zhang
Liujuan Cao
QiXiang Ye
Feiyue Huang
David Doermann
CVBM
16
505
0
22 Mar 2019
Spiking-YOLO: Spiking Neural Network for Energy-Efficient Object Detection
Seijoon Kim
Seongsik Park
Byunggook Na
Sungroh Yoon
ObjD
17
410
0
12 Mar 2019
Knowledge Adaptation for Efficient Semantic Segmentation
Tong He
Chunhua Shen
Zhi Tian
Dong Gong
Changming Sun
Youliang Yan
SSeg
8
225
0
12 Mar 2019
Structured Knowledge Distillation for Dense Prediction
Yifan Liu
Chris Liu
Jingdong Wang
Zhenbo Luo
27
576
0
11 Mar 2019
Graphical Contrastive Losses for Scene Graph Parsing
Ji Zhang
Kevin J. Shih
Ahmed Elgammal
Andrew Tao
Bryan Catanzaro
21
227
0
07 Mar 2019
Multi-loss-aware Channel Pruning of Deep Networks
Yiming Hu
Siyang Sun
Jianquan Li
Jiagang Zhu
Xingang Wang
Qingyi Gu
20
8
0
27 Feb 2019
Mockingbird: Defending Against Deep-Learning-Based Website Fingerprinting Attacks with Adversarial Traces
Mohammad Saidur Rahman
Mohsen Imani
Nate Mathews
M. Wright
AAML
14
80
0
18 Feb 2019
Improved Knowledge Distillation via Teacher Assistant
Seyed Iman Mirzadeh
Mehrdad Farajtabar
Ang Li
Nir Levine
Akihiro Matsukawa
H. Ghasemzadeh
53
1,067
0
09 Feb 2019
DELTA: DEep Learning Transfer using Feature Map with Attention for Convolutional Networks
Xingjian Li
Haoyi Xiong
Hanchao Wang
Yuxuan Rao
Liping Liu
Zeyu Chen
Jun Huan
21
171
0
26 Jan 2019
Attending Category Disentangled Global Context for Image Classification
Keke Tang
Guodong Wei
Runnan Chen
Jie Zhu
Zhaoquan Gu
Wenping Wang
17
0
0
17 Dec 2018
Learning Student Networks via Feature Embedding
Hanting Chen
Yunhe Wang
Chang Xu
Chao Xu
Dacheng Tao
19
94
0
17 Dec 2018
Optimizing speed/accuracy trade-off for person re-identification via knowledge distillation
Idoia Ruiz
Bogdan Raducanu
Rakesh Mehta
J. Amores
19
12
0
07 Dec 2018
An Embarrassingly Simple Approach for Knowledge Distillation
Mengya Gao
Yujun Shen
Quanquan Li
Junjie Yan
Liang Wan
Dahua Lin
Chen Change Loy
Xiaoou Tang
12
4
0
05 Dec 2018
Accelerating Large Scale Knowledge Distillation via Dynamic Importance Sampling
Minghan Li
Tanli Zuo
Ruicheng Li
Martha White
Weishi Zheng
29
3
0
03 Dec 2018
Unsupervised Domain Adaptation using Generative Models and Self-ensembling
Eman T. Hassan
Xin Chen
David J. Crandall
OOD
20
7
0
02 Dec 2018
Previous
1
2
3
...
21
22
23
24
Next