Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1612.03928
Cited By
Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer
12 December 2016
Sergey Zagoruyko
N. Komodakis
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer"
50 / 1,157 papers shown
Title
Knowledge Transfer via Dense Cross-Layer Mutual-Distillation
Anbang Yao
Dawei Sun
12
55
0
18 Aug 2020
Self-supervised Video Representation Learning by Pace Prediction
Jiangliu Wang
Jianbo Jiao
Yunhui Liu
SSL
AI4TS
24
233
0
13 Aug 2020
Transferring Inter-Class Correlation
Hui Wen
Yue-bo Wu
Chenming Yang
Jingjing Li
Yue Zhu
Xu Jiang
H. Duan
6
0
0
11 Aug 2020
Sharp Multiple Instance Learning for DeepFake Video Detection
Xiaodan Li
Yining Lang
YueFeng Chen
Xiaofeng Mao
Yuan He
Shuhui Wang
Hui Xue
Quan Lu
AAML
35
171
0
11 Aug 2020
HAPI: Hardware-Aware Progressive Inference
Stefanos Laskaridis
Stylianos I. Venieris
Hyeji Kim
Nicholas D. Lane
25
45
0
10 Aug 2020
Exploring Relations in Untrimmed Videos for Self-Supervised Learning
Dezhao Luo
Bo Fang
Yu Zhou
Yucan Zhou
Dayan Wu
Weiping Wang
35
21
0
06 Aug 2020
MED-TEX: Transferring and Explaining Knowledge with Less Data from Pretrained Medical Imaging Models
Thanh Nguyen-Duc
He Zhao
Jianfei Cai
Dinh Q. Phung
VLM
MedIm
33
4
0
06 Aug 2020
Prime-Aware Adaptive Distillation
Youcai Zhang
Zhonghao Lan
Yuchen Dai
Fangao Zeng
Yan Bai
Jie Chang
Yichen Wei
21
40
0
04 Aug 2020
Differentiable Feature Aggregation Search for Knowledge Distillation
Yushuo Guan
Pengyu Zhao
Bingxuan Wang
Yuanxing Zhang
Cong Yao
Kaigui Bian
Jian Tang
FedML
25
44
0
02 Aug 2020
Distilling Visual Priors from Self-Supervised Learning
Bingchen Zhao
Xin Wen
SSL
8
14
0
01 Aug 2020
Learning with Privileged Information for Efficient Image Super-Resolution
Wonkyung Lee
Junghyup Lee
Dohyung Kim
Bumsub Ham
33
134
0
15 Jul 2020
Learning to Learn Parameterized Classification Networks for Scalable Input Images
Duo Li
Anbang Yao
Qifeng Chen
22
11
0
13 Jul 2020
Data-Efficient Ranking Distillation for Image Retrieval
Zakaria Laskar
Arno Solin
VLM
8
4
0
10 Jul 2020
Distillation Guided Residual Learning for Binary Convolutional Neural Networks
Jianming Ye
Shiliang Zhang
Jingdong Wang
MQ
30
19
0
10 Jul 2020
Optical Flow Distillation: Towards Efficient and Stable Video Style Transfer
Xinghao Chen
Yiman Zhang
Yunhe Wang
Han Shu
Chunjing Xu
Chang Xu
VGen
21
54
0
10 Jul 2020
Evaluation for Weakly Supervised Object Localization: Protocol, Metrics, and Datasets
Junsuk Choe
Seong Joon Oh
Sanghyuk Chun
Seungho Lee
Zeynep Akata
Hyunjung Shim
WSOL
352
23
0
08 Jul 2020
Robust Re-Identification by Multiple Views Knowledge Distillation
Angelo Porrello
Luca Bergamini
Simone Calderara
32
65
0
08 Jul 2020
RIFLE: Backpropagation in Depth for Deep Transfer Learning through Re-Initializing the Fully-connected LayEr
Xingjian Li
Haoyi Xiong
Haozhe An
Chengzhong Xu
Dejing Dou
ODL
20
39
0
07 Jul 2020
Structured (De)composable Representations Trained with Neural Networks
Graham Spinks
Marie-Francine Moens
OCL
6
1
0
07 Jul 2020
LabelEnc: A New Intermediate Supervision Method for Object Detection
Miao Hao
Yitao Liu
Xinming Zhang
Jian Sun
22
25
0
07 Jul 2020
Improving Weakly Supervised Visual Grounding by Contrastive Knowledge Distillation
Liwei Wang
Jing-ling Huang
Yin Li
Kun Xu
Zhengyuan Yang
Dong Yu
ObjD
19
81
0
03 Jul 2020
Knowledge Distillation Beyond Model Compression
F. Sarfraz
Elahe Arani
Bahram Zonooz
35
40
0
03 Jul 2020
Domain Adaptation without Source Data
Youngeun Kim
Donghyeon Cho
Kyeongtak Han
Priyadarshini Panda
Sungeun Hong
TTA
11
174
0
03 Jul 2020
Interactive Knowledge Distillation
Shipeng Fu
Zhen Li
Jun Xu
Ming-Ming Cheng
Gwanggil Jeon
Xiaomin Yang
24
6
0
03 Jul 2020
Opportunities and Challenges in Deep Learning Adversarial Robustness: A Survey
S. Silva
Peyman Najafirad
AAML
OOD
31
131
0
01 Jul 2020
Attention-Guided Generative Adversarial Network to Address Atypical Anatomy in Modality Transfer
H. Emami
Ming Dong
C. Glide-Hurst
MedIm
GAN
34
5
0
27 Jun 2020
ULSAM: Ultra-Lightweight Subspace Attention Module for Compact Convolutional Neural Networks
Rajat Saini
N. Jha
B. K. Das
Sparsh Mittal
C.Krishna Mohan
29
83
0
26 Jun 2020
DanHAR: Dual Attention Network For Multimodal Human Activity Recognition Using Wearable Sensors
Wenbin Gao
Lei Zhang
Qi Teng
Jun He
Hao Wu
HAI
19
136
0
25 Jun 2020
Distilling Object Detectors with Task Adaptive Regularization
Ruoyu Sun
Fuhui Tang
Xiaopeng Zhang
H. Xiong
Qi Tian
ObjD
12
56
0
23 Jun 2020
Paying more attention to snapshots of Iterative Pruning: Improving Model Compression via Ensemble Distillation
Duong H. Le
Vo Trung Nhan
N. Thoai
VLM
27
7
0
20 Jun 2020
Video Playback Rate Perception for Self-supervisedSpatio-Temporal Representation Learning
Yuan Yao
Chang-rui Liu
Dezhao Luo
Yu Zhou
QiXiang Ye
29
169
0
20 Jun 2020
Optimizing Grouped Convolutions on Edge Devices
Perry Gibson
José Cano
Jack Turner
Elliot J. Crowley
Michael F. P. O'Boyle
Amos Storkey
16
25
0
17 Jun 2020
Multi-fidelity Neural Architecture Search with Knowledge Distillation
I. Trofimov
Nikita Klyuchnikov
Mikhail Salnikov
Alexander N. Filippov
Evgeny Burnaev
37
15
0
15 Jun 2020
Global Convergence of Sobolev Training for Overparameterized Neural Networks
Jorio Cocola
Paul Hand
11
6
0
14 Jun 2020
MetaPerturb: Transferable Regularizer for Heterogeneous Tasks and Architectures
Jeongun Ryu
Jaewoong Shin
Haebeom Lee
Sung Ju Hwang
AAML
OOD
22
8
0
13 Jun 2020
Ensemble Distillation for Robust Model Fusion in Federated Learning
Tao R. Lin
Lingjing Kong
Sebastian U. Stich
Martin Jaggi
FedML
19
1,017
0
12 Jun 2020
Knowledge Distillation Meets Self-Supervision
Guodong Xu
Ziwei Liu
Xiaoxiao Li
Chen Change Loy
FedML
44
280
0
12 Jun 2020
Knowledge Distillation: A Survey
Jianping Gou
B. Yu
Stephen J. Maybank
Dacheng Tao
VLM
28
2,857
0
09 Jun 2020
Fast Modeling and Understanding Fluid Dynamics Systems with Encoder-Decoder Networks
Rohan Thavarajah
X. Zhai
Zhe-Rui Ma
D. Castineira
PINN
AI4CE
21
8
0
09 Jun 2020
ResKD: Residual-Guided Knowledge Distillation
Xuewei Li
Songyuan Li
Bourahla Omar
Fei Wu
Xi Li
28
47
0
08 Jun 2020
Peer Collaborative Learning for Online Knowledge Distillation
Guile Wu
S. Gong
FedML
22
128
0
07 Jun 2020
Multi-view Contrastive Learning for Online Knowledge Distillation
Chuanguang Yang
Zhulin An
Yongjun Xu
24
23
0
07 Jun 2020
An Overview of Neural Network Compression
James OÑeill
AI4CE
47
98
0
05 Jun 2020
Channel Distillation: Channel-Wise Attention for Knowledge Distillation
Zaida Zhou
Chaoran Zhuge
Xinwei Guan
Wen Liu
19
49
0
02 Jun 2020
CHEER: Rich Model Helps Poor Model via Knowledge Infusion
Cao Xiao
T. Hoang
linda Qiao
Tengfei Ma
Jimeng Sun
21
3
0
21 May 2020
Weakly Supervised Representation Learning with Coarse Labels
Yuanhong Xu
Qi Qian
Hao Li
Rong Jin
Juhua Hu
SSL
28
8
0
19 May 2020
Learning from a Lightweight Teacher for Efficient Knowledge Distillation
Yuang Liu
Wei Zhang
Jun Wang
22
3
0
19 May 2020
Incremental Learning for End-to-End Automatic Speech Recognition
Li Fu
Xiaoxiao Li
Libo Zi
Zhengchen Zhang
Youzheng Wu
Xiaodong He
Bowen Zhou
CLL
42
23
0
11 May 2020
Data-Free Network Quantization With Adversarial Knowledge Distillation
Yoojin Choi
Jihwan P. Choi
Mostafa El-Khamy
Jungwon Lee
MQ
27
119
0
08 May 2020
MAZE: Data-Free Model Stealing Attack Using Zeroth-Order Gradient Estimation
Sanjay Kariyappa
A. Prakash
Moinuddin K. Qureshi
AAML
32
146
0
06 May 2020
Previous
1
2
3
...
18
19
20
...
22
23
24
Next