Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1612.03928
Cited By
Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer
12 December 2016
Sergey Zagoruyko
N. Komodakis
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer"
50 / 1,157 papers shown
Title
CNNs, LSTMs, and Attention Networks for Pathology Detection in Medical Data
N. Vogt
8
6
0
02 Dec 2019
Towards Oracle Knowledge Distillation with Neural Architecture Search
Minsoo Kang
Jonghwan Mun
Bohyung Han
FedML
41
44
0
29 Nov 2019
QKD: Quantization-aware Knowledge Distillation
Jangho Kim
Yash Bhalgat
Jinwon Lee
Chirag I. Patel
Nojun Kwak
MQ
24
64
0
28 Nov 2019
Region Normalization for Image Inpainting
Tao Yu
Zongyu Guo
Xin Jin
Shilin Wu
Zhibo Chen
Weiping Li
Zhizheng Zhang
Sen Liu
26
182
0
23 Nov 2019
MSD: Multi-Self-Distillation Learning via Multi-classifiers within Deep Neural Networks
Yunteng Luan
Hanyu Zhao
Zhi-Xin Yang
Yafei Dai
6
22
0
21 Nov 2019
Search to Distill: Pearls are Everywhere but not the Eyes
Yu Liu
Xuhui Jia
Mingxing Tan
Raviteja Vemulapalli
Yukun Zhu
Bradley Green
Xiaogang Wang
30
68
0
20 Nov 2019
Attention Guided Anomaly Localization in Images
Shashanka Venkataramanan
Kuan-Chuan Peng
Rajat Vikram Singh
Abhijit Mahalanobis
30
22
0
19 Nov 2019
Towards Making Deep Transfer Learning Never Hurt
Ruosi Wan
Haoyi Xiong
Xingjian Li
Zhanxing Zhu
Jun Huan
30
21
0
18 Nov 2019
Preparing Lessons: Improve Knowledge Distillation with Better Supervision
Tiancheng Wen
Shenqi Lai
Xueming Qian
25
68
0
18 Nov 2019
Label-similarity Curriculum Learning
Ürün Dogan
A. Deshmukh
Marcin Machura
Christian Igel
23
21
0
15 Nov 2019
Knowledge Representing: Efficient, Sparse Representation of Prior Knowledge for Knowledge Distillation
Junjie Liu
Dongchao Wen
Hongxing Gao
Wei Tao
Tse-Wei Chen
Kinya Osa
Masami Kato
30
21
0
13 Nov 2019
Collaborative Distillation for Top-N Recommendation
Jae-woong Lee
Minjin Choi
Jongwuk Lee
Hyunjung Shim
22
47
0
13 Nov 2019
Knowledge Distillation in Document Retrieval
Siamak Shakeri
A. Sethy
Cheng Cheng
FedML
14
5
0
11 Nov 2019
Distilling Pixel-Wise Feature Similarities for Semantic Segmentation
Yuhu Shan
19
7
0
31 Oct 2019
Adversarial Feature Alignment: Avoid Catastrophic Forgetting in Incremental Task Lifelong Learning
Hang Zhou
Ziwei Liu
Chenglei Wu
Ping Luo
Lifeng Sun
CLL
17
22
0
24 Oct 2019
Contrastive Representation Distillation
Yonglong Tian
Dilip Krishnan
Phillip Isola
47
1,034
0
23 Oct 2019
EdgeAI: A Vision for Deep Learning in IoT Era
Kartikeya Bhardwaj
Naveen Suda
R. Marculescu
19
12
0
23 Oct 2019
Torchreid: A Library for Deep Learning Person Re-Identification in Pytorch
Kaiyang Zhou
Tao Xiang
38
117
0
22 Oct 2019
Reduced-Order Modeling of Deep Neural Networks
Julia Gusak
Talgat Daulbaev
E. Ponomarev
A. Cichocki
Ivan Oseledets
BDL
AI4CE
22
8
0
15 Oct 2019
Learning Generalisable Omni-Scale Representations for Person Re-Identification
Kaiyang Zhou
Yongxin Yang
Andrea Cavallaro
Tao Xiang
30
217
0
15 Oct 2019
Orchestrating the Development Lifecycle of Machine Learning-Based IoT Applications: A Taxonomy and Survey
Bin Qian
Jie Su
Z. Wen
D. N. Jha
Yinhao Li
...
Albert Y. Zomaya
Omer F. Rana
Lizhe Wang
Maciej Koutny
R. Ranjan
28
4
0
11 Oct 2019
Noise as a Resource for Learning in Knowledge Distillation
Elahe Arani
F. Sarfraz
Bahram Zonooz
18
6
0
11 Oct 2019
On the Efficacy of Knowledge Distillation
Ligang He
Rui Mao
57
600
0
03 Oct 2019
Distillation
≈
\approx
≈
Early Stopping? Harvesting Dark Knowledge Utilizing Anisotropic Information Retrieval For Overparameterized Neural Network
Bin Dong
Jikai Hou
Yiping Lu
Zhihua Zhang
28
41
0
02 Oct 2019
AntMan: Sparse Low-Rank Compression to Accelerate RNN inference
Samyam Rajbhandari
H. Shrivastava
J. Rho
MQ
29
8
0
02 Oct 2019
Training convolutional neural networks with cheap convolutions and online distillation
Jiao Xie
Shaohui Lin
Yichen Zhang
Linkai Luo
27
12
0
28 Sep 2019
Attention Forcing for Sequence-to-sequence Model Training
Qingyun Dou
Yiting Lu
Joshua Efiong
Mark Gales
27
6
0
26 Sep 2019
Balanced Binary Neural Networks with Gated Residual
Mingzhu Shen
Xianglong Liu
Ruihao Gong
Kai Han
MQ
25
33
0
26 Sep 2019
Deep Model Transferability from Attribution Maps
Mingli Song
Yixin Chen
Xinchao Wang
Chengchao Shen
Xiuming Zhang
27
54
0
26 Sep 2019
Extremely Small BERT Models from Mixed-Vocabulary Training
Sanqiang Zhao
Raghav Gupta
Yang Song
Denny Zhou
VLM
14
53
0
25 Sep 2019
Attention Convolutional Binary Neural Tree for Fine-Grained Visual Categorization
Ruyi Ji
Longyin Wen
Libo Zhang
Dawei Du
Ynajun Wu
Chen Zhao
Xianglong Liu
Feiyue Huang
29
163
0
25 Sep 2019
FEED: Feature-level Ensemble for Knowledge Distillation
Seonguk Park
Nojun Kwak
FedML
31
41
0
24 Sep 2019
Controlling an Autonomous Vehicle with Deep Reinforcement Learning
A. Folkers
Matthias Rick
C. Büskens
22
67
0
24 Sep 2019
Class Feature Pyramids for Video Explanation
Alexandros Stergiou
G. Kapidis
Grigorios Kalliatakis
C. Chrysoulas
R. Poppe
R. Veltkamp
FAtt
33
18
0
18 Sep 2019
Transfer Learning with Dynamic Adversarial Adaptation Network
Chaohui Yu
Jindong Wang
Yiqiang Chen
Meiyu Huang
TTA
OOD
24
296
0
18 Sep 2019
Ensemble Knowledge Distillation for Learning Improved and Efficient Networks
Umar Asif
Jianbin Tang
S. Harrer
FedML
19
75
0
17 Sep 2019
Knowledge Transfer Graph for Deep Collaborative Learning
Soma Minami
Tsubasa Hirakawa
Takayoshi Yamashita
H. Fujiyoshi
30
9
0
10 Sep 2019
Extreme Low Resolution Activity Recognition with Confident Spatial-Temporal Attention Transfer
Yucai Bai
Qinglong Zou
Xieyuanli Chen
Lingxi Li
Zhengming Ding
Long Chen
20
3
0
09 Sep 2019
Auxiliary Learning for Deep Multi-task Learning
Yifan Liu
Bohan Zhuang
Chunhua Shen
Hao Chen
Wei Yin
MoE
33
10
0
05 Sep 2019
Attention-based Dropout Layer for Weakly Supervised Object Localization
Junsuk Choe
Hyunjung Shim
WSOL
22
365
0
27 Aug 2019
Improving Adversarial Robustness via Attention and Adversarial Logit Pairing
Dou Goodman
Xingjian Li
Ji Liu
Jun Huan
Tao Wei
AAML
16
7
0
23 Aug 2019
Customizing Student Networks From Heterogeneous Teachers via Adaptive Knowledge Amalgamation
Chengchao Shen
Mengqi Xue
Xinchao Wang
Mingli Song
Li Sun
Xiuming Zhang
MoMe
15
56
0
20 Aug 2019
SPA-GAN: Spatial Attention GAN for Image-to-Image Translation
H. Emami
Majid Moradi Aliabadi
Ming Dong
R. Chinnam
GAN
23
168
0
19 Aug 2019
MobileFAN: Transferring Deep Hidden Representation for Face Alignment
Yang Zhao
Yifan Liu
Chunhua Shen
Yongsheng Gao
Shengwu Xiong
CVBM
27
39
0
11 Aug 2019
Effective Training of Convolutional Neural Networks with Low-bitwidth Weights and Activations
Bohan Zhuang
Jing Liu
Mingkui Tan
Lingqiao Liu
Ian Reid
Chunhua Shen
MQ
29
45
0
10 Aug 2019
Learning Cross-Modal Deep Representations for Multi-Modal MR Image Segmentation
Cheng Li
Hui Sun
Zaiyi Liu
Meiyun Wang
Hairong Zheng
Shanshan Wang
27
44
0
06 Aug 2019
Learning Lightweight Lane Detection CNNs by Self Attention Distillation
Yuenan Hou
Zheng Ma
Chunxiao Liu
Chen Change Loy
15
550
0
02 Aug 2019
Memory- and Communication-Aware Model Compression for Distributed Deep Learning Inference on IoT
Kartikeya Bhardwaj
Chingyi Lin
A. L. Sartor
R. Marculescu
GNN
26
51
0
26 Jul 2019
Distilled Siamese Networks for Visual Tracking
Jianbing Shen
Yuanpei Liu
Xingping Dong
Xiankai Lu
Fahad Shahbaz Khan
Guosheng Lin
20
101
0
24 Jul 2019
Similarity-Preserving Knowledge Distillation
Frederick Tung
Greg Mori
45
961
0
23 Jul 2019
Previous
1
2
3
...
20
21
22
23
24
Next