ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1612.03928
  4. Cited By
Paying More Attention to Attention: Improving the Performance of
  Convolutional Neural Networks via Attention Transfer

Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer

12 December 2016
Sergey Zagoruyko
N. Komodakis
ArXivPDFHTML

Papers citing "Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer"

50 / 1,157 papers shown
Title
Group channel pruning and spatial attention distilling for object
  detection
Group channel pruning and spatial attention distilling for object detection
Yun Chu
Pu Li
Yong Bai
Zhuhua Hu
Yongqing Chen
Jiafeng Lu
VLM
24
13
0
02 Jun 2023
Symmetric Uncertainty-Aware Feature Transmission for Depth
  Super-Resolution
Symmetric Uncertainty-Aware Feature Transmission for Depth Super-Resolution
Wuxuan Shi
Mang Ye
Bo Du
25
17
0
01 Jun 2023
Are Large Kernels Better Teachers than Transformers for ConvNets?
Are Large Kernels Better Teachers than Transformers for ConvNets?
Tianjin Huang
Lu Yin
Zhenyu Zhang
Lijuan Shen
Meng Fang
Mykola Pechenizkiy
Zhangyang Wang
Shiwei Liu
38
13
0
30 May 2023
Budget-Aware Graph Convolutional Network Design using Probabilistic
  Magnitude Pruning
Budget-Aware Graph Convolutional Network Design using Probabilistic Magnitude Pruning
H. Sahbi
23
0
0
30 May 2023
Improving Knowledge Distillation via Regularizing Feature Norm and
  Direction
Improving Knowledge Distillation via Regularizing Feature Norm and Direction
Yuzhu Wang
Lechao Cheng
Manni Duan
Yongheng Wang
Zunlei Feng
Shu Kong
47
20
0
26 May 2023
Triplet Knowledge Distillation
Triplet Knowledge Distillation
Xijun Wang
Dongyang Liu
Meina Kan
Chunrui Han
Zhongqin Wu
Shiguang Shan
42
3
0
25 May 2023
VanillaKD: Revisit the Power of Vanilla Knowledge Distillation from
  Small Scale to Large Scale
VanillaKD: Revisit the Power of Vanilla Knowledge Distillation from Small Scale to Large Scale
Zhiwei Hao
Jianyuan Guo
Kai Han
Han Hu
Chang Xu
Yunhe Wang
38
16
0
25 May 2023
On the Impact of Knowledge Distillation for Model Interpretability
On the Impact of Knowledge Distillation for Model Interpretability
Hyeongrok Han
Siwon Kim
Hyun-Soo Choi
Sungroh Yoon
29
4
0
25 May 2023
HARD: Hard Augmentations for Robust Distillation
HARD: Hard Augmentations for Robust Distillation
Arne F. Nix
Max F. Burg
Fabian H. Sinz
AAML
44
1
0
24 May 2023
AdvFunMatch: When Consistent Teaching Meets Adversarial Robustness
AdvFunMatch: When Consistent Teaching Meets Adversarial Robustness
Ziuhi Wu
Haichang Gao
Bingqian Zhou
Ping Wang
AAML
24
0
0
24 May 2023
Decoupled Kullback-Leibler Divergence Loss
Decoupled Kullback-Leibler Divergence Loss
Jiequan Cui
Zhuotao Tian
Zhisheng Zhong
Xiaojuan Qi
Bei Yu
Hanwang Zhang
39
38
0
23 May 2023
NORM: Knowledge Distillation via N-to-One Representation Matching
NORM: Knowledge Distillation via N-to-One Representation Matching
Xiaolong Liu
Lujun Li
Chao Li
Anbang Yao
55
68
0
23 May 2023
Revisiting Data Augmentation in Model Compression: An Empirical and
  Comprehensive Study
Revisiting Data Augmentation in Model Compression: An Empirical and Comprehensive Study
Muzhou Yu
Linfeng Zhang
Kaisheng Ma
28
2
0
22 May 2023
Distilling ChatGPT for Explainable Automated Student Answer Assessment
Distilling ChatGPT for Explainable Automated Student Answer Assessment
Jiazheng Li
Lin Gui
Yuxiang Zhou
David West
Cesare Aloisi
Yulan He
28
28
0
22 May 2023
Is Synthetic Data From Diffusion Models Ready for Knowledge
  Distillation?
Is Synthetic Data From Diffusion Models Ready for Knowledge Distillation?
Zheng Li
Yuxuan Li
Penghai Zhao
Renjie Song
Xiang Li
Jian Yang
34
19
0
22 May 2023
Cross-modality Data Augmentation for End-to-End Sign Language
  Translation
Cross-modality Data Augmentation for End-to-End Sign Language Translation
Jinhui Ye
Wenxiang Jiao
Xing Wang
Zhaopeng Tu
Hui Xiong
SLR
21
21
0
18 May 2023
Student-friendly Knowledge Distillation
Student-friendly Knowledge Distillation
Mengyang Yuan
Bo Lang
Fengnan Quan
20
17
0
18 May 2023
Cross-Modality Time-Variant Relation Learning for Generating Dynamic
  Scene Graphs
Cross-Modality Time-Variant Relation Learning for Generating Dynamic Scene Graphs
Jingyi Wang
Jinfa Huang
Can Zhang
Zhidong Deng
49
7
0
15 May 2023
On enhancing the robustness of Vision Transformers: Defensive Diffusion
On enhancing the robustness of Vision Transformers: Defensive Diffusion
Raza Imam
Muhammad Huzaifa
Mohammed El-Amine Azz
MedIm
DiffM
46
5
0
14 May 2023
DynamicKD: An Effective Knowledge Distillation via Dynamic Entropy
  Correction-Based Distillation for Gap Optimizing
DynamicKD: An Effective Knowledge Distillation via Dynamic Entropy Correction-Based Distillation for Gap Optimizing
Songling Zhu
Ronghua Shang
Bo Yuan
Weitong Zhang
Yangyang Li
Licheng Jiao
35
7
0
09 May 2023
Leveraging Synthetic Targets for Machine Translation
Leveraging Synthetic Targets for Machine Translation
Sarthak Mittal
Oleksii Hrinchuk
Oleksii Kuchaiev
35
2
0
07 May 2023
Structural and Statistical Texture Knowledge Distillation for Semantic Segmentation
Structural and Statistical Texture Knowledge Distillation for Semantic Segmentation
Deyi Ji
Haoran Wang
Mingyuan Tao
Jianqiang Huang
Xiansheng Hua
Hongtao Lu
35
61
0
06 May 2023
Avatar Knowledge Distillation: Self-ensemble Teacher Paradigm with
  Uncertainty
Avatar Knowledge Distillation: Self-ensemble Teacher Paradigm with Uncertainty
Yuan Zhang
Weihua Chen
Yichen Lu
Tao Huang
Xiuyu Sun
Jian Cao
52
8
0
04 May 2023
Learngene: Inheriting Condensed Knowledge from the Ancestry Model to
  Descendant Models
Learngene: Inheriting Condensed Knowledge from the Ancestry Model to Descendant Models
Qiufeng Wang
Xu Yang
Shuxia Lin
Jing Wang
Xin Geng
43
10
0
03 May 2023
CORSD: Class-Oriented Relational Self Distillation
CORSD: Class-Oriented Relational Self Distillation
Muzhou Yu
S. Tan
Kailu Wu
Runpei Dong
Linfeng Zhang
Kaisheng Ma
24
0
0
28 Apr 2023
Class Attention Transfer Based Knowledge Distillation
Class Attention Transfer Based Knowledge Distillation
Ziyao Guo
Haonan Yan
Hui Li
Xiao-La Lin
18
64
0
25 Apr 2023
Bayesian Optimization Meets Self-Distillation
Bayesian Optimization Meets Self-Distillation
HyunJae Lee
Heon Song
Hyeonsoo Lee
Gi-hyeon Lee
Suyeong Park
Donggeun Yoo
UQCV
BDL
41
1
0
25 Apr 2023
Improving Knowledge Distillation via Transferring Learning Ability
Improving Knowledge Distillation via Transferring Learning Ability
Long Liu
Tong Li
Hui Cheng
13
1
0
24 Apr 2023
Function-Consistent Feature Distillation
Function-Consistent Feature Distillation
Dongyang Liu
Meina Kan
Shiguang Shan
Xilin Chen
52
18
0
24 Apr 2023
Knowledge Distillation Under Ideal Joint Classifier Assumption
Knowledge Distillation Under Ideal Joint Classifier Assumption
Huayu Li
Xiwen Chen
G. Ditzler
Janet Roveda
Ao Li
18
1
0
19 Apr 2023
Deep Collective Knowledge Distillation
Deep Collective Knowledge Distillation
Jihyeon Seo
Kyusam Oh
Chanho Min
Yongkeun Yun
Sungwoo Cho
19
0
0
18 Apr 2023
MMANet: Margin-aware Distillation and Modality-aware Regularization for
  Incomplete Multimodal Learning
MMANet: Margin-aware Distillation and Modality-aware Regularization for Incomplete Multimodal Learning
Shicai Wei
Yang Luo
Chunbo Luo
26
29
0
17 Apr 2023
LaSNN: Layer-wise ANN-to-SNN Distillation for Effective and Efficient
  Training in Deep Spiking Neural Networks
LaSNN: Layer-wise ANN-to-SNN Distillation for Effective and Efficient Training in Deep Spiking Neural Networks
Di Hong
Jiangrong Shen
Yu Qi
Yueming Wang
25
5
0
17 Apr 2023
SFT-KD-Recon: Learning a Student-friendly Teacher for Knowledge
  Distillation in Magnetic Resonance Image Reconstruction
SFT-KD-Recon: Learning a Student-friendly Teacher for Knowledge Distillation in Magnetic Resonance Image Reconstruction
NagaGayathri Matcha
Sriprabha Ramanarayanan
Mohammad Al Fahim
S. RahulG
Keerthi Ram
M. Sivaprakasam
21
2
0
11 Apr 2023
Towards Efficient Task-Driven Model Reprogramming with Foundation Models
Towards Efficient Task-Driven Model Reprogramming with Foundation Models
Shoukai Xu
Jiangchao Yao
Ran Luo
Shuhai Zhang
Zihao Lian
Mingkui Tan
Bo Han
Yaowei Wang
29
6
0
05 Apr 2023
Cross-Class Feature Augmentation for Class Incremental Learning
Cross-Class Feature Augmentation for Class Incremental Learning
Taehoon Kim
Jaeyoo Park
Bohyung Han
CLL
32
5
0
04 Apr 2023
Label-guided Attention Distillation for Lane Segmentation
Label-guided Attention Distillation for Lane Segmentation
Zhikang Liu
Lanyun Zhu
24
15
0
04 Apr 2023
Long-Tailed Visual Recognition via Self-Heterogeneous Integration with
  Knowledge Excavation
Long-Tailed Visual Recognition via Self-Heterogeneous Integration with Knowledge Excavation
Yang Jin
Mengke Li
Yang Lu
Y. Cheung
Hanzi Wang
43
21
0
03 Apr 2023
DIME-FM: DIstilling Multimodal and Efficient Foundation Models
DIME-FM: DIstilling Multimodal and Efficient Foundation Models
Ximeng Sun
Pengchuan Zhang
Peizhao Zhang
Hardik Shah
Kate Saenko
Xide Xia
VLM
30
20
0
31 Mar 2023
CAMEL: Communicative Agents for "Mind" Exploration of Large Language
  Model Society
CAMEL: Communicative Agents for "Mind" Exploration of Large Language Model Society
Ge Li
Hasan Hammoud
Hani Itani
Dmitrii Khizbullin
Guohao Li
SyDa
ALM
52
413
0
31 Mar 2023
Decomposed Cross-modal Distillation for RGB-based Temporal Action
  Detection
Decomposed Cross-modal Distillation for RGB-based Temporal Action Detection
Pilhyeon Lee
Taeoh Kim
Minho Shim
Dongyoon Wee
H. Byun
41
11
0
30 Mar 2023
Information-Theoretic GAN Compression with Variational Energy-based
  Model
Information-Theoretic GAN Compression with Variational Energy-based Model
Minsoo Kang
Hyewon Yoo
Eunhee Kang
Sehwan Ki
Hyong-Euk Lee
Bohyung Han
GAN
31
3
0
28 Mar 2023
DisWOT: Student Architecture Search for Distillation WithOut Training
DisWOT: Student Architecture Search for Distillation WithOut Training
Peijie Dong
Lujun Li
Zimian Wei
46
57
0
28 Mar 2023
Hard Sample Matters a Lot in Zero-Shot Quantization
Hard Sample Matters a Lot in Zero-Shot Quantization
Huantong Li
Xiangmiao Wu
Fanbing Lv
Daihai Liao
Thomas H. Li
Yonggang Zhang
Bo Han
Mingkui Tan
MQ
24
20
0
24 Mar 2023
Exploiting Unlabelled Photos for Stronger Fine-Grained SBIR
Exploiting Unlabelled Photos for Stronger Fine-Grained SBIR
Aneeshan Sain
A. Bhunia
Subhadeep Koley
Pinaki Nath Chowdhury
Soumitri Chattopadhyay
Tao Xiang
Yi-Zhe Song
30
18
0
24 Mar 2023
A Simple and Generic Framework for Feature Distillation via Channel-wise
  Transformation
A Simple and Generic Framework for Feature Distillation via Channel-wise Transformation
Ziwei Liu
Yongtao Wang
Xiaojie Chu
32
5
0
23 Mar 2023
From Knowledge Distillation to Self-Knowledge Distillation: A Unified
  Approach with Normalized Loss and Customized Soft Labels
From Knowledge Distillation to Self-Knowledge Distillation: A Unified Approach with Normalized Loss and Customized Soft Labels
Zhendong Yang
Ailing Zeng
Zhe Li
Tianke Zhang
Chun Yuan
Yu Li
31
74
0
23 Mar 2023
Understanding the Role of the Projector in Knowledge Distillation
Understanding the Role of the Projector in Knowledge Distillation
Roy Miles
K. Mikolajczyk
35
21
0
20 Mar 2023
Channel-Aware Distillation Transformer for Depth Estimation on Nano
  Drones
Channel-Aware Distillation Transformer for Depth Estimation on Nano Drones
Ning Zhang
F. Nex
G. Vosselman
N. Kerle
34
1
0
18 Mar 2023
Neural Architecture Search for Effective Teacher-Student Knowledge
  Transfer in Language Models
Neural Architecture Search for Effective Teacher-Student Knowledge Transfer in Language Models
Aashka Trivedi
Takuma Udagawa
Michele Merler
Yikang Shen
Yousef El-Kurdi
Bishwaranjan Bhattacharjee
40
7
0
16 Mar 2023
Previous
123...678...222324
Next