ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1612.03928
  4. Cited By
Paying More Attention to Attention: Improving the Performance of
  Convolutional Neural Networks via Attention Transfer

Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer

12 December 2016
Sergey Zagoruyko
N. Komodakis
ArXivPDFHTML

Papers citing "Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer"

50 / 1,157 papers shown
Title
Diffusion Models and Representation Learning: A Survey
Diffusion Models and Representation Learning: A Survey
Michael Fuest
Pingchuan Ma
Ming Gui
Johannes S. Fischer
Vincent Tao Hu
Bjorn Ommer
DiffM
43
20
0
30 Jun 2024
MuGSI: Distilling GNNs with Multi-Granularity Structural Information for
  Graph Classification
MuGSI: Distilling GNNs with Multi-Granularity Structural Information for Graph Classification
Tianjun Yao
Jiaqi Sun
Defu Cao
Kun Zhang
Guangyi Chen
43
5
0
28 Jun 2024
CSAKD: Knowledge Distillation with Cross Self-Attention for
  Hyperspectral and Multispectral Image Fusion
CSAKD: Knowledge Distillation with Cross Self-Attention for Hyperspectral and Multispectral Image Fusion
Chih-Chung Hsu
Chih-Chien Ni
Chia-Ming Lee
Li-Wei Kang
24
3
0
28 Jun 2024
MAGIC: Meta-Ability Guided Interactive Chain-of-Distillation for
  Effective-and-Efficient Vision-and-Language Navigation
MAGIC: Meta-Ability Guided Interactive Chain-of-Distillation for Effective-and-Efficient Vision-and-Language Navigation
Liuyi Wang
Zongtao He
Mengjiao Shen
Jingwei Yang
Chengju Liu
Qijun Chen
VLM
36
2
0
25 Jun 2024
InFiConD: Interactive No-code Fine-tuning with Concept-based Knowledge
  Distillation
InFiConD: Interactive No-code Fine-tuning with Concept-based Knowledge Distillation
Jinbin Huang
Wenbin He
Liang Gou
Liu Ren
Chris Bryan
55
0
0
25 Jun 2024
AND: Audio Network Dissection for Interpreting Deep Acoustic Models
AND: Audio Network Dissection for Interpreting Deep Acoustic Models
Tung-Yu Wu
Yu-Xiang Lin
Tsui-Wei Weng
54
1
0
24 Jun 2024
SelfReg-UNet: Self-Regularized UNet for Medical Image Segmentation
SelfReg-UNet: Self-Regularized UNet for Medical Image Segmentation
Wenhui Zhu
Xiwen Chen
Peijie Qiu
Mohammad Farazi
Aristeidis Sotiras
Abolfazl Razi
Yalin Wang
SSeg
SSL
46
11
0
21 Jun 2024
Lightweight Model Pre-training via Language Guided Knowledge
  Distillation
Lightweight Model Pre-training via Language Guided Knowledge Distillation
Mingsheng Li
Lin Zhang
Mingzhen Zhu
Zilong Huang
Gang Yu
Jiayuan Fan
Tao Chen
41
1
0
17 Jun 2024
NBA: defensive distillation for backdoor removal via neural behavior
  alignment
NBA: defensive distillation for backdoor removal via neural behavior alignment
Zonghao Ying
Bin Wu
AAML
26
6
0
16 Jun 2024
Depth Anything V2
Depth Anything V2
Lihe Yang
Bingyi Kang
Zilong Huang
Zhen Zhao
Xiaogang Xu
Jiashi Feng
Hengshuang Zhao
DiffM
VLM
MDE
59
337
0
13 Jun 2024
PC-LoRA: Low-Rank Adaptation for Progressive Model Compression with
  Knowledge Distillation
PC-LoRA: Low-Rank Adaptation for Progressive Model Compression with Knowledge Distillation
Injoon Hwang
Haewon Park
Youngwan Lee
Jooyoung Yang
SunJae Maeng
AI4CE
27
0
0
13 Jun 2024
Adaptive Teaching with Shared Classifier for Knowledge Distillation
Adaptive Teaching with Shared Classifier for Knowledge Distillation
Jaeyeon Jang
Young-Ik Kim
Jisu Lim
Hyeonseong Lee
21
0
0
12 Jun 2024
DistilDoc: Knowledge Distillation for Visually-Rich Document Applications
DistilDoc: Knowledge Distillation for Visually-Rich Document Applications
Jordy Van Landeghem
Subhajit Maity
Ayan Banerjee
Matthew Blaschko
Marie-Francine Moens
Josep Lladós
Sanket Biswas
52
2
0
12 Jun 2024
ReDistill: Residual Encoded Distillation for Peak Memory Reduction of CNNs
ReDistill: Residual Encoded Distillation for Peak Memory Reduction of CNNs
Fang Chen
Gourav Datta
Mujahid Al Rafi
Hyeran Jeon
Meng Tang
93
1
0
06 Jun 2024
Mutual Information Guided Backdoor Mitigation for Pre-trained Encoders
Mutual Information Guided Backdoor Mitigation for Pre-trained Encoders
Tingxu Han
Weisong Sun
Ziqi Ding
Chunrong Fang
Hanwei Qian
Jiaxun Li
Zhenyu Chen
Xiangyu Zhang
AAML
38
7
0
05 Jun 2024
Distilling Aggregated Knowledge for Weakly-Supervised Video Anomaly Detection
Distilling Aggregated Knowledge for Weakly-Supervised Video Anomaly Detection
Jash Dalvi
Ali Dabouei
Gunjan Dhanuka
Min Xu
28
0
0
05 Jun 2024
Robust Knowledge Distillation Based on Feature Variance Against
  Backdoored Teacher Model
Robust Knowledge Distillation Based on Feature Variance Against Backdoored Teacher Model
Jinyin Chen
Xiaoming Zhao
Haibin Zheng
Xiao Li
Sheng Xiang
Haifeng Guo
AAML
30
3
0
01 Jun 2024
Provable Contrastive Continual Learning
Provable Contrastive Continual Learning
Yichen Wen
Zhiquan Tan
Kaipeng Zheng
Chuanlong Xie
Weiran Huang
CLL
34
4
0
29 May 2024
Relational Self-supervised Distillation with Compact Descriptors for
  Image Copy Detection
Relational Self-supervised Distillation with Compact Descriptors for Image Copy Detection
Juntae Kim
Sungwon Woo
Jongho Nang
42
1
0
28 May 2024
LoReTrack: Efficient and Accurate Low-Resolution Transformer Tracking
LoReTrack: Efficient and Accurate Low-Resolution Transformer Tracking
Shaohua Dong
Yunhe Feng
Qing Yang
Yuewei Lin
Heng Fan
58
1
0
27 May 2024
Low-Resolution Chest X-ray Classification via Knowledge Distillation and
  Multi-task Learning
Low-Resolution Chest X-ray Classification via Knowledge Distillation and Multi-task Learning
Yasmeena Akhter
Rishabh Ranjan
Richa Singh
Mayank Vatsa
16
0
0
22 May 2024
AMFD: Distillation via Adaptive Multimodal Fusion for Multispectral
  Pedestrian Detection
AMFD: Distillation via Adaptive Multimodal Fusion for Multispectral Pedestrian Detection
Zizhao Chen
Yeqiang Qian
Xiaoxiao Yang
Chunxiang Wang
Ming Yang
29
1
0
21 May 2024
CLRKDNet: Speeding up Lane Detection with Knowledge Distillation
CLRKDNet: Speeding up Lane Detection with Knowledge Distillation
Weiqing Qi
Guoyang Zhao
Fulong Ma
Linwei Zheng
Ming Liu
53
1
0
21 May 2024
Exploring Dark Knowledge under Various Teacher Capacities and Addressing
  Capacity Mismatch
Exploring Dark Knowledge under Various Teacher Capacities and Addressing Capacity Mismatch
Xin-Chun Li
Wen-Shu Fan
Bowen Tao
Le Gan
De-Chuan Zhan
37
2
0
21 May 2024
Stereo-Knowledge Distillation from dpMV to Dual Pixels for Light Field
  Video Reconstruction
Stereo-Knowledge Distillation from dpMV to Dual Pixels for Light Field Video Reconstruction
Aryan Garg
Raghav Mallampali
Akshat Joshi
Shrisudhan Govindarajan
Kaushik Mitra
39
0
0
20 May 2024
Fully Exploiting Every Real Sample: SuperPixel Sample Gradient Model
  Stealing
Fully Exploiting Every Real Sample: SuperPixel Sample Gradient Model Stealing
Yunlong Zhao
Xiaoheng Deng
Yijing Liu
Xin-jun Pei
Jiazhi Xia
Wei Chen
AAML
37
3
0
18 May 2024
Exploring Graph-based Knowledge: Multi-Level Feature Distillation via
  Channels Relational Graph
Exploring Graph-based Knowledge: Multi-Level Feature Distillation via Channels Relational Graph
Zhiwei Wang
Jun Huang
Longhua Ma
Chengyu Wu
Hongyu Ma
35
0
0
14 May 2024
GLiRA: Black-Box Membership Inference Attack via Knowledge Distillation
GLiRA: Black-Box Membership Inference Attack via Knowledge Distillation
Andrey V. Galichin
Mikhail Aleksandrovich Pautov
Alexey Zhavoronkin
Oleg Y. Rogov
Ivan Oseledets
AAML
34
1
0
13 May 2024
From Algorithm to Hardware: A Survey on Efficient and Safe Deployment of
  Deep Neural Networks
From Algorithm to Hardware: A Survey on Efficient and Safe Deployment of Deep Neural Networks
Xue Geng
Zhe Wang
Chunyun Chen
Qing Xu
Kaixin Xu
...
Zhenghua Chen
M. Aly
Jie Lin
Min-man Wu
Xiaoli Li
33
1
0
09 May 2024
DVMSR: Distillated Vision Mamba for Efficient Super-Resolution
DVMSR: Distillated Vision Mamba for Efficient Super-Resolution
Xiaoyan Lei
Wenlong Zhang
Weifeng Cao
29
11
0
05 May 2024
Multi-view Action Recognition via Directed Gromov-Wasserstein
  Discrepancy
Multi-view Action Recognition via Directed Gromov-Wasserstein Discrepancy
Hoang-Quan Nguyen
Thanh-Dat Truong
Khoa Luu
36
1
0
02 May 2024
ATOM: Attention Mixer for Efficient Dataset Distillation
ATOM: Attention Mixer for Efficient Dataset Distillation
Samir Khaki
A. Sajedi
Kai Wang
Lucy Z. Liu
Y. Lawryshyn
Konstantinos N. Plataniotis
52
3
0
02 May 2024
Deep Boosting Learning: A Brand-new Cooperative Approach for Image-Text
  Matching
Deep Boosting Learning: A Brand-new Cooperative Approach for Image-Text Matching
Haiwen Diao
Ying Zhang
Shang Gao
Xiang Ruan
Huchuan Lu
36
3
0
28 Apr 2024
Low-Rank Knowledge Decomposition for Medical Foundation Models
Low-Rank Knowledge Decomposition for Medical Foundation Models
Yuhang Zhou
Haolin Li
Siyuan Du
Jiangchao Yao
Ya Zhang
Yanfeng Wang
28
3
0
26 Apr 2024
CNN2GNN: How to Bridge CNN with GNN
CNN2GNN: How to Bridge CNN with GNN
Ziheng Jiao
Hongyuan Zhang
Xuelong Li
26
1
0
23 Apr 2024
A Multimodal Feature Distillation with CNN-Transformer Network for Brain
  Tumor Segmentation with Incomplete Modalities
A Multimodal Feature Distillation with CNN-Transformer Network for Brain Tumor Segmentation with Incomplete Modalities
Ming Kang
F. F. Ting
Raphaël C.-W. Phan
Zongyuan Ge
Chee-Ming Ting
54
2
0
22 Apr 2024
CKD: Contrastive Knowledge Distillation from A Sample-wise Perspective
CKD: Contrastive Knowledge Distillation from A Sample-wise Perspective
Wencheng Zhu
Xin Zhou
Pengfei Zhu
Yu Wang
Qinghua Hu
VLM
64
1
0
22 Apr 2024
MergeNet: Knowledge Migration across Heterogeneous Models, Tasks, and
  Modalities
MergeNet: Knowledge Migration across Heterogeneous Models, Tasks, and Modalities
Kunxi Li
Tianyu Zhan
Kairui Fu
Shengyu Zhang
Kun Kuang
Jiwei Li
Zhou Zhao
Fei Wu
MoMe
24
0
0
20 Apr 2024
Dynamic Temperature Knowledge Distillation
Dynamic Temperature Knowledge Distillation
Yukang Wei
Yu Bai
38
4
0
19 Apr 2024
An Experimental Study on Exploring Strong Lightweight Vision
  Transformers via Masked Image Modeling Pre-Training
An Experimental Study on Exploring Strong Lightweight Vision Transformers via Masked Image Modeling Pre-Training
Jin Gao
Shubo Lin
Shaoru Wang
Yutong Kou
Zeming Li
Liang Li
Congxuan Zhang
Xiaoqin Zhang
Yizheng Wang
Weiming Hu
47
1
0
18 Apr 2024
Dynamic Self-adaptive Multiscale Distillation from Pre-trained
  Multimodal Large Model for Efficient Cross-modal Representation Learning
Dynamic Self-adaptive Multiscale Distillation from Pre-trained Multimodal Large Model for Efficient Cross-modal Representation Learning
Zhengyang Liang
Meiyu Liang
Wei Huang
Yawen Li
Zhe Xue
55
1
0
16 Apr 2024
ReffAKD: Resource-efficient Autoencoder-based Knowledge Distillation
ReffAKD: Resource-efficient Autoencoder-based Knowledge Distillation
Divyang Doshi
Jung-Eun Kim
31
1
0
15 Apr 2024
MTKD: Multi-Teacher Knowledge Distillation for Image Super-Resolution
MTKD: Multi-Teacher Knowledge Distillation for Image Super-Resolution
Yuxuan Jiang
Chen Feng
Fan Zhang
David Bull
SupR
53
11
0
15 Apr 2024
Bootstrapping Chest CT Image Understanding by Distilling Knowledge from
  X-ray Expert Models
Bootstrapping Chest CT Image Understanding by Distilling Knowledge from X-ray Expert Models
Weiwei Cao
Jianpeng Zhang
Yingda Xia
Tony C. W. Mok
Zi Li
X. Ye
Le Lu
Jian Zheng
Yuxing Tang
Ling Zhang
31
1
0
07 Apr 2024
On the Surprising Efficacy of Distillation as an Alternative to
  Pre-Training Small Models
On the Surprising Efficacy of Distillation as an Alternative to Pre-Training Small Models
Sean Farhat
Deming Chen
47
0
0
04 Apr 2024
Knowledge Distillation with Multi-granularity Mixture of Priors for
  Image Super-Resolution
Knowledge Distillation with Multi-granularity Mixture of Priors for Image Super-Resolution
Simiao Li
Yun-feng Zhang
Wei Li
Hanting Chen
Wenjia Wang
Bingyi Jing
Shaohui Lin
Jie Hu
SupR
40
1
0
03 Apr 2024
Improve Knowledge Distillation via Label Revision and Data Selection
Improve Knowledge Distillation via Label Revision and Data Selection
Weichao Lan
Yiu-ming Cheung
Qing Xu
Buhua Liu
Zhikai Hu
Mengke Li
Zhenghua Chen
37
2
0
03 Apr 2024
Adaptive Feature Fusion Neural Network for Glaucoma Segmentation on
  Unseen Fundus Images
Adaptive Feature Fusion Neural Network for Glaucoma Segmentation on Unseen Fundus Images
Jiyuan Zhong
Hu Ke
Ming Yan
MedIm
26
1
0
02 Apr 2024
Task Integration Distillation for Object Detectors
Task Integration Distillation for Object Detectors
Hai Su
ZhenWen Jian
Songsen Yu
46
1
0
02 Apr 2024
A Comprehensive Review of Knowledge Distillation in Computer Vision
A Comprehensive Review of Knowledge Distillation in Computer Vision
Sheikh Musa Kaleem
Tufail Rouf
Gousia Habib
Tausifa Jan Saleem
Brejesh Lall
VLM
35
13
0
01 Apr 2024
Previous
123456...222324
Next