ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1612.03928
  4. Cited By
Paying More Attention to Attention: Improving the Performance of
  Convolutional Neural Networks via Attention Transfer

Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer

12 December 2016
Sergey Zagoruyko
N. Komodakis
ArXivPDFHTML

Papers citing "Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer"

50 / 1,157 papers shown
Title
Knowledge Transfer via Dense Cross-Layer Mutual-Distillation
Knowledge Transfer via Dense Cross-Layer Mutual-Distillation
Anbang Yao
Dawei Sun
12
55
0
18 Aug 2020
Self-supervised Video Representation Learning by Pace Prediction
Self-supervised Video Representation Learning by Pace Prediction
Jiangliu Wang
Jianbo Jiao
Yunhui Liu
SSL
AI4TS
24
233
0
13 Aug 2020
Transferring Inter-Class Correlation
Transferring Inter-Class Correlation
Hui Wen
Yue-bo Wu
Chenming Yang
Jingjing Li
Yue Zhu
Xu Jiang
H. Duan
6
0
0
11 Aug 2020
Sharp Multiple Instance Learning for DeepFake Video Detection
Sharp Multiple Instance Learning for DeepFake Video Detection
Xiaodan Li
Yining Lang
YueFeng Chen
Xiaofeng Mao
Yuan He
Shuhui Wang
Hui Xue
Quan Lu
AAML
35
171
0
11 Aug 2020
HAPI: Hardware-Aware Progressive Inference
HAPI: Hardware-Aware Progressive Inference
Stefanos Laskaridis
Stylianos I. Venieris
Hyeji Kim
Nicholas D. Lane
25
45
0
10 Aug 2020
Exploring Relations in Untrimmed Videos for Self-Supervised Learning
Exploring Relations in Untrimmed Videos for Self-Supervised Learning
Dezhao Luo
Bo Fang
Yu Zhou
Yucan Zhou
Dayan Wu
Weiping Wang
35
21
0
06 Aug 2020
MED-TEX: Transferring and Explaining Knowledge with Less Data from
  Pretrained Medical Imaging Models
MED-TEX: Transferring and Explaining Knowledge with Less Data from Pretrained Medical Imaging Models
Thanh Nguyen-Duc
He Zhao
Jianfei Cai
Dinh Q. Phung
VLM
MedIm
33
4
0
06 Aug 2020
Prime-Aware Adaptive Distillation
Prime-Aware Adaptive Distillation
Youcai Zhang
Zhonghao Lan
Yuchen Dai
Fangao Zeng
Yan Bai
Jie Chang
Yichen Wei
21
40
0
04 Aug 2020
Differentiable Feature Aggregation Search for Knowledge Distillation
Differentiable Feature Aggregation Search for Knowledge Distillation
Yushuo Guan
Pengyu Zhao
Bingxuan Wang
Yuanxing Zhang
Cong Yao
Kaigui Bian
Jian Tang
FedML
25
44
0
02 Aug 2020
Distilling Visual Priors from Self-Supervised Learning
Distilling Visual Priors from Self-Supervised Learning
Bingchen Zhao
Xin Wen
SSL
8
14
0
01 Aug 2020
Learning with Privileged Information for Efficient Image
  Super-Resolution
Learning with Privileged Information for Efficient Image Super-Resolution
Wonkyung Lee
Junghyup Lee
Dohyung Kim
Bumsub Ham
33
134
0
15 Jul 2020
Learning to Learn Parameterized Classification Networks for Scalable
  Input Images
Learning to Learn Parameterized Classification Networks for Scalable Input Images
Duo Li
Anbang Yao
Qifeng Chen
22
11
0
13 Jul 2020
Data-Efficient Ranking Distillation for Image Retrieval
Data-Efficient Ranking Distillation for Image Retrieval
Zakaria Laskar
Arno Solin
VLM
8
4
0
10 Jul 2020
Distillation Guided Residual Learning for Binary Convolutional Neural
  Networks
Distillation Guided Residual Learning for Binary Convolutional Neural Networks
Jianming Ye
Shiliang Zhang
Jingdong Wang
MQ
30
19
0
10 Jul 2020
Optical Flow Distillation: Towards Efficient and Stable Video Style
  Transfer
Optical Flow Distillation: Towards Efficient and Stable Video Style Transfer
Xinghao Chen
Yiman Zhang
Yunhe Wang
Han Shu
Chunjing Xu
Chang Xu
VGen
21
54
0
10 Jul 2020
Evaluation for Weakly Supervised Object Localization: Protocol, Metrics,
  and Datasets
Evaluation for Weakly Supervised Object Localization: Protocol, Metrics, and Datasets
Junsuk Choe
Seong Joon Oh
Sanghyuk Chun
Seungho Lee
Zeynep Akata
Hyunjung Shim
WSOL
352
23
0
08 Jul 2020
Robust Re-Identification by Multiple Views Knowledge Distillation
Robust Re-Identification by Multiple Views Knowledge Distillation
Angelo Porrello
Luca Bergamini
Simone Calderara
32
65
0
08 Jul 2020
RIFLE: Backpropagation in Depth for Deep Transfer Learning through
  Re-Initializing the Fully-connected LayEr
RIFLE: Backpropagation in Depth for Deep Transfer Learning through Re-Initializing the Fully-connected LayEr
Xingjian Li
Haoyi Xiong
Haozhe An
Chengzhong Xu
Dejing Dou
ODL
20
39
0
07 Jul 2020
Structured (De)composable Representations Trained with Neural Networks
Structured (De)composable Representations Trained with Neural Networks
Graham Spinks
Marie-Francine Moens
OCL
6
1
0
07 Jul 2020
LabelEnc: A New Intermediate Supervision Method for Object Detection
LabelEnc: A New Intermediate Supervision Method for Object Detection
Miao Hao
Yitao Liu
Xinming Zhang
Jian Sun
22
25
0
07 Jul 2020
Improving Weakly Supervised Visual Grounding by Contrastive Knowledge
  Distillation
Improving Weakly Supervised Visual Grounding by Contrastive Knowledge Distillation
Liwei Wang
Jing-ling Huang
Yin Li
Kun Xu
Zhengyuan Yang
Dong Yu
ObjD
19
81
0
03 Jul 2020
Knowledge Distillation Beyond Model Compression
Knowledge Distillation Beyond Model Compression
F. Sarfraz
Elahe Arani
Bahram Zonooz
35
40
0
03 Jul 2020
Domain Adaptation without Source Data
Domain Adaptation without Source Data
Youngeun Kim
Donghyeon Cho
Kyeongtak Han
Priyadarshini Panda
Sungeun Hong
TTA
11
174
0
03 Jul 2020
Interactive Knowledge Distillation
Interactive Knowledge Distillation
Shipeng Fu
Zhen Li
Jun Xu
Ming-Ming Cheng
Gwanggil Jeon
Xiaomin Yang
24
6
0
03 Jul 2020
Opportunities and Challenges in Deep Learning Adversarial Robustness: A
  Survey
Opportunities and Challenges in Deep Learning Adversarial Robustness: A Survey
S. Silva
Peyman Najafirad
AAML
OOD
31
131
0
01 Jul 2020
Attention-Guided Generative Adversarial Network to Address Atypical
  Anatomy in Modality Transfer
Attention-Guided Generative Adversarial Network to Address Atypical Anatomy in Modality Transfer
H. Emami
Ming Dong
C. Glide-Hurst
MedIm
GAN
34
5
0
27 Jun 2020
ULSAM: Ultra-Lightweight Subspace Attention Module for Compact
  Convolutional Neural Networks
ULSAM: Ultra-Lightweight Subspace Attention Module for Compact Convolutional Neural Networks
Rajat Saini
N. Jha
B. K. Das
Sparsh Mittal
C.Krishna Mohan
29
83
0
26 Jun 2020
DanHAR: Dual Attention Network For Multimodal Human Activity Recognition
  Using Wearable Sensors
DanHAR: Dual Attention Network For Multimodal Human Activity Recognition Using Wearable Sensors
Wenbin Gao
Lei Zhang
Qi Teng
Jun He
Hao Wu
HAI
19
136
0
25 Jun 2020
Distilling Object Detectors with Task Adaptive Regularization
Distilling Object Detectors with Task Adaptive Regularization
Ruoyu Sun
Fuhui Tang
Xiaopeng Zhang
H. Xiong
Qi Tian
ObjD
12
56
0
23 Jun 2020
Paying more attention to snapshots of Iterative Pruning: Improving Model
  Compression via Ensemble Distillation
Paying more attention to snapshots of Iterative Pruning: Improving Model Compression via Ensemble Distillation
Duong H. Le
Vo Trung Nhan
N. Thoai
VLM
27
7
0
20 Jun 2020
Video Playback Rate Perception for Self-supervisedSpatio-Temporal
  Representation Learning
Video Playback Rate Perception for Self-supervisedSpatio-Temporal Representation Learning
Yuan Yao
Chang-rui Liu
Dezhao Luo
Yu Zhou
QiXiang Ye
29
169
0
20 Jun 2020
Optimizing Grouped Convolutions on Edge Devices
Optimizing Grouped Convolutions on Edge Devices
Perry Gibson
José Cano
Jack Turner
Elliot J. Crowley
Michael F. P. O'Boyle
Amos Storkey
16
25
0
17 Jun 2020
Multi-fidelity Neural Architecture Search with Knowledge Distillation
Multi-fidelity Neural Architecture Search with Knowledge Distillation
I. Trofimov
Nikita Klyuchnikov
Mikhail Salnikov
Alexander N. Filippov
Evgeny Burnaev
37
15
0
15 Jun 2020
Global Convergence of Sobolev Training for Overparameterized Neural
  Networks
Global Convergence of Sobolev Training for Overparameterized Neural Networks
Jorio Cocola
Paul Hand
11
6
0
14 Jun 2020
MetaPerturb: Transferable Regularizer for Heterogeneous Tasks and
  Architectures
MetaPerturb: Transferable Regularizer for Heterogeneous Tasks and Architectures
Jeongun Ryu
Jaewoong Shin
Haebeom Lee
Sung Ju Hwang
AAML
OOD
22
8
0
13 Jun 2020
Ensemble Distillation for Robust Model Fusion in Federated Learning
Ensemble Distillation for Robust Model Fusion in Federated Learning
Tao R. Lin
Lingjing Kong
Sebastian U. Stich
Martin Jaggi
FedML
19
1,017
0
12 Jun 2020
Knowledge Distillation Meets Self-Supervision
Knowledge Distillation Meets Self-Supervision
Guodong Xu
Ziwei Liu
Xiaoxiao Li
Chen Change Loy
FedML
44
280
0
12 Jun 2020
Knowledge Distillation: A Survey
Knowledge Distillation: A Survey
Jianping Gou
B. Yu
Stephen J. Maybank
Dacheng Tao
VLM
28
2,857
0
09 Jun 2020
Fast Modeling and Understanding Fluid Dynamics Systems with
  Encoder-Decoder Networks
Fast Modeling and Understanding Fluid Dynamics Systems with Encoder-Decoder Networks
Rohan Thavarajah
X. Zhai
Zhe-Rui Ma
D. Castineira
PINN
AI4CE
21
8
0
09 Jun 2020
ResKD: Residual-Guided Knowledge Distillation
ResKD: Residual-Guided Knowledge Distillation
Xuewei Li
Songyuan Li
Bourahla Omar
Fei Wu
Xi Li
28
47
0
08 Jun 2020
Peer Collaborative Learning for Online Knowledge Distillation
Peer Collaborative Learning for Online Knowledge Distillation
Guile Wu
S. Gong
FedML
22
128
0
07 Jun 2020
Multi-view Contrastive Learning for Online Knowledge Distillation
Multi-view Contrastive Learning for Online Knowledge Distillation
Chuanguang Yang
Zhulin An
Yongjun Xu
24
23
0
07 Jun 2020
An Overview of Neural Network Compression
An Overview of Neural Network Compression
James OÑeill
AI4CE
47
98
0
05 Jun 2020
Channel Distillation: Channel-Wise Attention for Knowledge Distillation
Channel Distillation: Channel-Wise Attention for Knowledge Distillation
Zaida Zhou
Chaoran Zhuge
Xinwei Guan
Wen Liu
19
49
0
02 Jun 2020
CHEER: Rich Model Helps Poor Model via Knowledge Infusion
CHEER: Rich Model Helps Poor Model via Knowledge Infusion
Cao Xiao
T. Hoang
linda Qiao
Tengfei Ma
Jimeng Sun
21
3
0
21 May 2020
Weakly Supervised Representation Learning with Coarse Labels
Weakly Supervised Representation Learning with Coarse Labels
Yuanhong Xu
Qi Qian
Hao Li
Rong Jin
Juhua Hu
SSL
28
8
0
19 May 2020
Learning from a Lightweight Teacher for Efficient Knowledge Distillation
Learning from a Lightweight Teacher for Efficient Knowledge Distillation
Yuang Liu
Wei Zhang
Jun Wang
22
3
0
19 May 2020
Incremental Learning for End-to-End Automatic Speech Recognition
Incremental Learning for End-to-End Automatic Speech Recognition
Li Fu
Xiaoxiao Li
Libo Zi
Zhengchen Zhang
Youzheng Wu
Xiaodong He
Bowen Zhou
CLL
42
23
0
11 May 2020
Data-Free Network Quantization With Adversarial Knowledge Distillation
Data-Free Network Quantization With Adversarial Knowledge Distillation
Yoojin Choi
Jihwan P. Choi
Mostafa El-Khamy
Jungwon Lee
MQ
27
119
0
08 May 2020
MAZE: Data-Free Model Stealing Attack Using Zeroth-Order Gradient
  Estimation
MAZE: Data-Free Model Stealing Attack Using Zeroth-Order Gradient Estimation
Sanjay Kariyappa
A. Prakash
Moinuddin K. Qureshi
AAML
32
146
0
06 May 2020
Previous
123...181920...222324
Next