ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1612.03928
  4. Cited By
Paying More Attention to Attention: Improving the Performance of
  Convolutional Neural Networks via Attention Transfer

Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer

12 December 2016
Sergey Zagoruyko
N. Komodakis
ArXivPDFHTML

Papers citing "Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer"

50 / 1,157 papers shown
Title
Highlight Every Step: Knowledge Distillation via Collaborative Teaching
Highlight Every Step: Knowledge Distillation via Collaborative Teaching
Haoran Zhao
Xin Sun
Junyu Dong
Changrui Chen
Zihe Dong
27
57
0
23 Jul 2019
Bringing Giant Neural Networks Down to Earth with Unlabeled Data
Bringing Giant Neural Networks Down to Earth with Unlabeled Data
Yehui Tang
Shan You
Chang Xu
Boxin Shi
Chao Xu
24
11
0
13 Jul 2019
Justifying Diagnosis Decisions by Deep Neural Networks
Justifying Diagnosis Decisions by Deep Neural Networks
Graham Spinks
Marie-Francine Moens
37
13
0
12 Jul 2019
Distill-2MD-MTL: Data Distillation based on Multi-Dataset Multi-Domain
  Multi-Task Frame Work to Solve Face Related Tasksks, Multi Task Learning,
  Semi-Supervised Learning
Distill-2MD-MTL: Data Distillation based on Multi-Dataset Multi-Domain Multi-Task Frame Work to Solve Face Related Tasksks, Multi Task Learning, Semi-Supervised Learning
Sepidehsadat Hosseini
M. Shabani
N. Cho
CVBM
36
3
0
08 Jul 2019
Understanding More about Human and Machine Attention in Deep Neural
  Networks
Understanding More about Human and Machine Attention in Deep Neural Networks
Qiuxia Lai
Salman Khan
Wenguan Wang
Jianbing Shen
Hanqiu Sun
Ling Shao
HAI
XAI
24
7
0
20 Jun 2019
GAN-Knowledge Distillation for one-stage Object Detection
GAN-Knowledge Distillation for one-stage Object Detection
Wanwei Wang
Jin ke Yu Fan Zong
ObjD
22
28
0
20 Jun 2019
A One-step Pruning-recovery Framework for Acceleration of Convolutional
  Neural Networks
A One-step Pruning-recovery Framework for Acceleration of Convolutional Neural Networks
Dong Wang
Lei Zhou
Xiao Bai
Jun Zhou
14
2
0
18 Jun 2019
BlockSwap: Fisher-guided Block Substitution for Network Compression on a
  Budget
BlockSwap: Fisher-guided Block Substitution for Network Compression on a Budget
Jack Turner
Elliot J. Crowley
Michael F. P. O'Boyle
Amos Storkey
Gavia Gray
22
37
0
10 Jun 2019
Distilling Object Detectors with Fine-grained Feature Imitation
Distilling Object Detectors with Fine-grained Feature Imitation
Tao Wang
Li-xin Yuan
Xiaopeng Zhang
Jiashi Feng
ObjD
13
378
0
09 Jun 2019
Separable Layers Enable Structured Efficient Linear Substitutions
Separable Layers Enable Structured Efficient Linear Substitutions
Gavia Gray
Elliot J. Crowley
Amos Storkey
30
3
0
03 Jun 2019
Deeply-supervised Knowledge Synergy
Deeply-supervised Knowledge Synergy
Dawei Sun
Anbang Yao
Aojun Zhou
Hao Zhao
12
63
0
03 Jun 2019
SCAN: A Scalable Neural Networks Framework Towards Compact and Efficient
  Models
SCAN: A Scalable Neural Networks Framework Towards Compact and Efficient Models
Linfeng Zhang
Zhanhong Tan
Jiebo Song
Jingwei Chen
Chenglong Bao
Kaisheng Ma
16
71
0
27 May 2019
ShrinkTeaNet: Million-scale Lightweight Face Recognition via Shrinking
  Teacher-Student Networks
ShrinkTeaNet: Million-scale Lightweight Face Recognition via Shrinking Teacher-Student Networks
C. Duong
Khoa Luu
Kha Gia Quach
Ngan Le
CVBM
23
39
0
25 May 2019
Zero-shot Knowledge Transfer via Adversarial Belief Matching
Zero-shot Knowledge Transfer via Adversarial Belief Matching
P. Micaelli
Amos Storkey
19
228
0
23 May 2019
Network Pruning via Transformable Architecture Search
Network Pruning via Transformable Architecture Search
Xuanyi Dong
Yi Yang
3DPC
23
240
0
23 May 2019
Be Your Own Teacher: Improve the Performance of Convolutional Neural
  Networks via Self Distillation
Be Your Own Teacher: Improve the Performance of Convolutional Neural Networks via Self Distillation
Linfeng Zhang
Jiebo Song
Anni Gao
Jingwei Chen
Chenglong Bao
Kaisheng Ma
FedML
27
846
0
17 May 2019
Dream Distillation: A Data-Independent Model Compression Framework
Dream Distillation: A Data-Independent Model Compression Framework
Kartikeya Bhardwaj
Naveen Suda
R. Marculescu
DD
27
54
0
17 May 2019
Learning What and Where to Transfer
Learning What and Where to Transfer
Yunhun Jang
Hankook Lee
Sung Ju Hwang
Jinwoo Shin
22
148
0
15 May 2019
Triplet Distillation for Deep Face Recognition
Triplet Distillation for Deep Face Recognition
Yushu Feng
Huan Wang
Daniel T. Yi
Roland Hu
CVBM
11
45
0
11 May 2019
Creating Lightweight Object Detectors with Model Compression for
  Deployment on Edge Devices
Creating Lightweight Object Detectors with Model Compression for Deployment on Edge Devices
Yiwu Yao
Weiqiang Yang
Haoqi Zhu
33
0
0
06 May 2019
Omni-Scale Feature Learning for Person Re-Identification
Omni-Scale Feature Learning for Person Re-Identification
Kaiyang Zhou
Yongxin Yang
Andrea Cavallaro
Tao Xiang
30
820
0
02 May 2019
Knowledge Distillation via Route Constrained Optimization
Knowledge Distillation via Route Constrained Optimization
Xiao Jin
Baoyun Peng
Yichao Wu
Yu Liu
Jiaheng Liu
Ding Liang
Junjie Yan
Xiaolin Hu
20
169
0
19 Apr 2019
Feature Fusion for Online Mutual Knowledge Distillation
Feature Fusion for Online Mutual Knowledge Distillation
Jangho Kim
Minsung Hyun
Inseop Chung
Nojun Kwak
FedML
26
91
0
19 Apr 2019
Biphasic Learning of GANs for High-Resolution Image-to-Image Translation
Biphasic Learning of GANs for High-Resolution Image-to-Image Translation
Jie Cao
Huaibo Huang
Yi Li
Jingtuo Liu
Ran He
Zhenan Sun
GAN
26
4
0
14 Apr 2019
Improved training of binary networks for human pose estimation and image
  recognition
Improved training of binary networks for human pose estimation and image recognition
Adrian Bulat
Georgios Tzimiropoulos
Jean Kossaifi
Maja Pantic
MQ
29
47
0
11 Apr 2019
Variational Information Distillation for Knowledge Transfer
Variational Information Distillation for Knowledge Transfer
Sungsoo Ahn
S. Hu
Andreas C. Damianou
Neil D. Lawrence
Zhenwen Dai
58
609
0
11 Apr 2019
Knowledge Squeezed Adversarial Network Compression
Knowledge Squeezed Adversarial Network Compression
Changyong Shu
Li Peng
Xie Yuan
Yanyun Qu
Longquan Dai
Lizhuang Ma
GAN
45
11
0
10 Apr 2019
Relational Knowledge Distillation
Relational Knowledge Distillation
Wonpyo Park
Dongju Kim
Yan Lu
Minsu Cho
16
1,387
0
10 Apr 2019
Learning Metrics from Teachers: Compact Networks for Image Embedding
Learning Metrics from Teachers: Compact Networks for Image Embedding
Lu Yu
V. O. Yazici
Xialei Liu
Joost van de Weijer
Yongmei Cheng
Arnau Ramisa
11
105
0
07 Apr 2019
Self-supervised Spatio-temporal Representation Learning for Videos by
  Predicting Motion and Appearance Statistics
Self-supervised Spatio-temporal Representation Learning for Videos by Predicting Motion and Appearance Statistics
Jiangliu Wang
Jianbo Jiao
Linchao Bao
Shengfeng He
Yunhui Liu
Wen Liu
SSL
24
204
0
07 Apr 2019
Attention Distillation for Learning Video Representations
Attention Distillation for Learning Video Representations
Miao Liu
Xin Chen
Yun C. Zhang
Yin Li
James M. Rehg
19
2
0
05 Apr 2019
A Comprehensive Overhaul of Feature Distillation
A Comprehensive Overhaul of Feature Distillation
Byeongho Heo
Jeesoo Kim
Sangdoo Yun
Hyojin Park
Nojun Kwak
J. Choi
19
569
0
03 Apr 2019
Correlation Congruence for Knowledge Distillation
Correlation Congruence for Knowledge Distillation
Baoyun Peng
Xiao Jin
Jiaheng Liu
Shunfeng Zhou
Yichao Wu
Yu Liu
Dongsheng Li
Zhaoning Zhang
63
507
0
03 Apr 2019
Improving Object Detection with Inverted Attention
Improving Object Detection with Inverted Attention
Zeyi Huang
Wei Ke
Dong Huang
29
32
0
28 Mar 2019
Training Quantized Neural Networks with a Full-precision Auxiliary
  Module
Training Quantized Neural Networks with a Full-precision Auxiliary Module
Bohan Zhuang
Lingqiao Liu
Mingkui Tan
Chunhua Shen
Ian Reid
MQ
32
62
0
27 Mar 2019
Towards Optimal Structured CNN Pruning via Generative Adversarial
  Learning
Towards Optimal Structured CNN Pruning via Generative Adversarial Learning
Shaohui Lin
Rongrong Ji
Chenqian Yan
Baochang Zhang
Liujuan Cao
QiXiang Ye
Feiyue Huang
David Doermann
CVBM
16
505
0
22 Mar 2019
Spiking-YOLO: Spiking Neural Network for Energy-Efficient Object
  Detection
Spiking-YOLO: Spiking Neural Network for Energy-Efficient Object Detection
Seijoon Kim
Seongsik Park
Byunggook Na
Sungroh Yoon
ObjD
17
410
0
12 Mar 2019
Knowledge Adaptation for Efficient Semantic Segmentation
Knowledge Adaptation for Efficient Semantic Segmentation
Tong He
Chunhua Shen
Zhi Tian
Dong Gong
Changming Sun
Youliang Yan
SSeg
8
225
0
12 Mar 2019
Structured Knowledge Distillation for Dense Prediction
Structured Knowledge Distillation for Dense Prediction
Yifan Liu
Chris Liu
Jingdong Wang
Zhenbo Luo
27
576
0
11 Mar 2019
Graphical Contrastive Losses for Scene Graph Parsing
Graphical Contrastive Losses for Scene Graph Parsing
Ji Zhang
Kevin J. Shih
Ahmed Elgammal
Andrew Tao
Bryan Catanzaro
21
227
0
07 Mar 2019
Multi-loss-aware Channel Pruning of Deep Networks
Multi-loss-aware Channel Pruning of Deep Networks
Yiming Hu
Siyang Sun
Jianquan Li
Jiagang Zhu
Xingang Wang
Qingyi Gu
20
8
0
27 Feb 2019
Mockingbird: Defending Against Deep-Learning-Based Website
  Fingerprinting Attacks with Adversarial Traces
Mockingbird: Defending Against Deep-Learning-Based Website Fingerprinting Attacks with Adversarial Traces
Mohammad Saidur Rahman
Mohsen Imani
Nate Mathews
M. Wright
AAML
14
80
0
18 Feb 2019
Improved Knowledge Distillation via Teacher Assistant
Improved Knowledge Distillation via Teacher Assistant
Seyed Iman Mirzadeh
Mehrdad Farajtabar
Ang Li
Nir Levine
Akihiro Matsukawa
H. Ghasemzadeh
53
1,067
0
09 Feb 2019
DELTA: DEep Learning Transfer using Feature Map with Attention for
  Convolutional Networks
DELTA: DEep Learning Transfer using Feature Map with Attention for Convolutional Networks
Xingjian Li
Haoyi Xiong
Hanchao Wang
Yuxuan Rao
Liping Liu
Zeyu Chen
Jun Huan
21
171
0
26 Jan 2019
Attending Category Disentangled Global Context for Image Classification
Keke Tang
Guodong Wei
Runnan Chen
Jie Zhu
Zhaoquan Gu
Wenping Wang
17
0
0
17 Dec 2018
Learning Student Networks via Feature Embedding
Learning Student Networks via Feature Embedding
Hanting Chen
Yunhe Wang
Chang Xu
Chao Xu
Dacheng Tao
19
94
0
17 Dec 2018
Optimizing speed/accuracy trade-off for person re-identification via
  knowledge distillation
Optimizing speed/accuracy trade-off for person re-identification via knowledge distillation
Idoia Ruiz
Bogdan Raducanu
Rakesh Mehta
J. Amores
19
12
0
07 Dec 2018
An Embarrassingly Simple Approach for Knowledge Distillation
An Embarrassingly Simple Approach for Knowledge Distillation
Mengya Gao
Yujun Shen
Quanquan Li
Junjie Yan
Liang Wan
Dahua Lin
Chen Change Loy
Xiaoou Tang
12
4
0
05 Dec 2018
Accelerating Large Scale Knowledge Distillation via Dynamic Importance
  Sampling
Accelerating Large Scale Knowledge Distillation via Dynamic Importance Sampling
Minghan Li
Tanli Zuo
Ruicheng Li
Martha White
Weishi Zheng
29
3
0
03 Dec 2018
Unsupervised Domain Adaptation using Generative Models and
  Self-ensembling
Unsupervised Domain Adaptation using Generative Models and Self-ensembling
Eman T. Hassan
Xin Chen
David J. Crandall
OOD
20
7
0
02 Dec 2018
Previous
123...21222324
Next