Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1612.03928
Cited By
Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer
12 December 2016
Sergey Zagoruyko
N. Komodakis
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer"
50 / 1,157 papers shown
Title
Diffusion Models and Representation Learning: A Survey
Michael Fuest
Pingchuan Ma
Ming Gui
Johannes S. Fischer
Vincent Tao Hu
Bjorn Ommer
DiffM
43
20
0
30 Jun 2024
MuGSI: Distilling GNNs with Multi-Granularity Structural Information for Graph Classification
Tianjun Yao
Jiaqi Sun
Defu Cao
Kun Zhang
Guangyi Chen
43
5
0
28 Jun 2024
CSAKD: Knowledge Distillation with Cross Self-Attention for Hyperspectral and Multispectral Image Fusion
Chih-Chung Hsu
Chih-Chien Ni
Chia-Ming Lee
Li-Wei Kang
24
3
0
28 Jun 2024
MAGIC: Meta-Ability Guided Interactive Chain-of-Distillation for Effective-and-Efficient Vision-and-Language Navigation
Liuyi Wang
Zongtao He
Mengjiao Shen
Jingwei Yang
Chengju Liu
Qijun Chen
VLM
36
2
0
25 Jun 2024
InFiConD: Interactive No-code Fine-tuning with Concept-based Knowledge Distillation
Jinbin Huang
Wenbin He
Liang Gou
Liu Ren
Chris Bryan
55
0
0
25 Jun 2024
AND: Audio Network Dissection for Interpreting Deep Acoustic Models
Tung-Yu Wu
Yu-Xiang Lin
Tsui-Wei Weng
54
1
0
24 Jun 2024
SelfReg-UNet: Self-Regularized UNet for Medical Image Segmentation
Wenhui Zhu
Xiwen Chen
Peijie Qiu
Mohammad Farazi
Aristeidis Sotiras
Abolfazl Razi
Yalin Wang
SSeg
SSL
46
11
0
21 Jun 2024
Lightweight Model Pre-training via Language Guided Knowledge Distillation
Mingsheng Li
Lin Zhang
Mingzhen Zhu
Zilong Huang
Gang Yu
Jiayuan Fan
Tao Chen
41
1
0
17 Jun 2024
NBA: defensive distillation for backdoor removal via neural behavior alignment
Zonghao Ying
Bin Wu
AAML
26
6
0
16 Jun 2024
Depth Anything V2
Lihe Yang
Bingyi Kang
Zilong Huang
Zhen Zhao
Xiaogang Xu
Jiashi Feng
Hengshuang Zhao
DiffM
VLM
MDE
59
337
0
13 Jun 2024
PC-LoRA: Low-Rank Adaptation for Progressive Model Compression with Knowledge Distillation
Injoon Hwang
Haewon Park
Youngwan Lee
Jooyoung Yang
SunJae Maeng
AI4CE
27
0
0
13 Jun 2024
Adaptive Teaching with Shared Classifier for Knowledge Distillation
Jaeyeon Jang
Young-Ik Kim
Jisu Lim
Hyeonseong Lee
21
0
0
12 Jun 2024
DistilDoc: Knowledge Distillation for Visually-Rich Document Applications
Jordy Van Landeghem
Subhajit Maity
Ayan Banerjee
Matthew Blaschko
Marie-Francine Moens
Josep Lladós
Sanket Biswas
52
2
0
12 Jun 2024
ReDistill: Residual Encoded Distillation for Peak Memory Reduction of CNNs
Fang Chen
Gourav Datta
Mujahid Al Rafi
Hyeran Jeon
Meng Tang
93
1
0
06 Jun 2024
Mutual Information Guided Backdoor Mitigation for Pre-trained Encoders
Tingxu Han
Weisong Sun
Ziqi Ding
Chunrong Fang
Hanwei Qian
Jiaxun Li
Zhenyu Chen
Xiangyu Zhang
AAML
38
7
0
05 Jun 2024
Distilling Aggregated Knowledge for Weakly-Supervised Video Anomaly Detection
Jash Dalvi
Ali Dabouei
Gunjan Dhanuka
Min Xu
28
0
0
05 Jun 2024
Robust Knowledge Distillation Based on Feature Variance Against Backdoored Teacher Model
Jinyin Chen
Xiaoming Zhao
Haibin Zheng
Xiao Li
Sheng Xiang
Haifeng Guo
AAML
30
3
0
01 Jun 2024
Provable Contrastive Continual Learning
Yichen Wen
Zhiquan Tan
Kaipeng Zheng
Chuanlong Xie
Weiran Huang
CLL
34
4
0
29 May 2024
Relational Self-supervised Distillation with Compact Descriptors for Image Copy Detection
Juntae Kim
Sungwon Woo
Jongho Nang
42
1
0
28 May 2024
LoReTrack: Efficient and Accurate Low-Resolution Transformer Tracking
Shaohua Dong
Yunhe Feng
Qing Yang
Yuewei Lin
Heng Fan
58
1
0
27 May 2024
Low-Resolution Chest X-ray Classification via Knowledge Distillation and Multi-task Learning
Yasmeena Akhter
Rishabh Ranjan
Richa Singh
Mayank Vatsa
16
0
0
22 May 2024
AMFD: Distillation via Adaptive Multimodal Fusion for Multispectral Pedestrian Detection
Zizhao Chen
Yeqiang Qian
Xiaoxiao Yang
Chunxiang Wang
Ming Yang
29
1
0
21 May 2024
CLRKDNet: Speeding up Lane Detection with Knowledge Distillation
Weiqing Qi
Guoyang Zhao
Fulong Ma
Linwei Zheng
Ming Liu
53
1
0
21 May 2024
Exploring Dark Knowledge under Various Teacher Capacities and Addressing Capacity Mismatch
Xin-Chun Li
Wen-Shu Fan
Bowen Tao
Le Gan
De-Chuan Zhan
37
2
0
21 May 2024
Stereo-Knowledge Distillation from dpMV to Dual Pixels for Light Field Video Reconstruction
Aryan Garg
Raghav Mallampali
Akshat Joshi
Shrisudhan Govindarajan
Kaushik Mitra
39
0
0
20 May 2024
Fully Exploiting Every Real Sample: SuperPixel Sample Gradient Model Stealing
Yunlong Zhao
Xiaoheng Deng
Yijing Liu
Xin-jun Pei
Jiazhi Xia
Wei Chen
AAML
37
3
0
18 May 2024
Exploring Graph-based Knowledge: Multi-Level Feature Distillation via Channels Relational Graph
Zhiwei Wang
Jun Huang
Longhua Ma
Chengyu Wu
Hongyu Ma
35
0
0
14 May 2024
GLiRA: Black-Box Membership Inference Attack via Knowledge Distillation
Andrey V. Galichin
Mikhail Aleksandrovich Pautov
Alexey Zhavoronkin
Oleg Y. Rogov
Ivan Oseledets
AAML
34
1
0
13 May 2024
From Algorithm to Hardware: A Survey on Efficient and Safe Deployment of Deep Neural Networks
Xue Geng
Zhe Wang
Chunyun Chen
Qing Xu
Kaixin Xu
...
Zhenghua Chen
M. Aly
Jie Lin
Min-man Wu
Xiaoli Li
33
1
0
09 May 2024
DVMSR: Distillated Vision Mamba for Efficient Super-Resolution
Xiaoyan Lei
Wenlong Zhang
Weifeng Cao
29
11
0
05 May 2024
Multi-view Action Recognition via Directed Gromov-Wasserstein Discrepancy
Hoang-Quan Nguyen
Thanh-Dat Truong
Khoa Luu
36
1
0
02 May 2024
ATOM: Attention Mixer for Efficient Dataset Distillation
Samir Khaki
A. Sajedi
Kai Wang
Lucy Z. Liu
Y. Lawryshyn
Konstantinos N. Plataniotis
52
3
0
02 May 2024
Deep Boosting Learning: A Brand-new Cooperative Approach for Image-Text Matching
Haiwen Diao
Ying Zhang
Shang Gao
Xiang Ruan
Huchuan Lu
36
3
0
28 Apr 2024
Low-Rank Knowledge Decomposition for Medical Foundation Models
Yuhang Zhou
Haolin Li
Siyuan Du
Jiangchao Yao
Ya Zhang
Yanfeng Wang
28
3
0
26 Apr 2024
CNN2GNN: How to Bridge CNN with GNN
Ziheng Jiao
Hongyuan Zhang
Xuelong Li
26
1
0
23 Apr 2024
A Multimodal Feature Distillation with CNN-Transformer Network for Brain Tumor Segmentation with Incomplete Modalities
Ming Kang
F. F. Ting
Raphaël C.-W. Phan
Zongyuan Ge
Chee-Ming Ting
54
2
0
22 Apr 2024
CKD: Contrastive Knowledge Distillation from A Sample-wise Perspective
Wencheng Zhu
Xin Zhou
Pengfei Zhu
Yu Wang
Qinghua Hu
VLM
64
1
0
22 Apr 2024
MergeNet: Knowledge Migration across Heterogeneous Models, Tasks, and Modalities
Kunxi Li
Tianyu Zhan
Kairui Fu
Shengyu Zhang
Kun Kuang
Jiwei Li
Zhou Zhao
Fei Wu
MoMe
24
0
0
20 Apr 2024
Dynamic Temperature Knowledge Distillation
Yukang Wei
Yu Bai
38
4
0
19 Apr 2024
An Experimental Study on Exploring Strong Lightweight Vision Transformers via Masked Image Modeling Pre-Training
Jin Gao
Shubo Lin
Shaoru Wang
Yutong Kou
Zeming Li
Liang Li
Congxuan Zhang
Xiaoqin Zhang
Yizheng Wang
Weiming Hu
47
1
0
18 Apr 2024
Dynamic Self-adaptive Multiscale Distillation from Pre-trained Multimodal Large Model for Efficient Cross-modal Representation Learning
Zhengyang Liang
Meiyu Liang
Wei Huang
Yawen Li
Zhe Xue
55
1
0
16 Apr 2024
ReffAKD: Resource-efficient Autoencoder-based Knowledge Distillation
Divyang Doshi
Jung-Eun Kim
31
1
0
15 Apr 2024
MTKD: Multi-Teacher Knowledge Distillation for Image Super-Resolution
Yuxuan Jiang
Chen Feng
Fan Zhang
David Bull
SupR
53
11
0
15 Apr 2024
Bootstrapping Chest CT Image Understanding by Distilling Knowledge from X-ray Expert Models
Weiwei Cao
Jianpeng Zhang
Yingda Xia
Tony C. W. Mok
Zi Li
X. Ye
Le Lu
Jian Zheng
Yuxing Tang
Ling Zhang
31
1
0
07 Apr 2024
On the Surprising Efficacy of Distillation as an Alternative to Pre-Training Small Models
Sean Farhat
Deming Chen
47
0
0
04 Apr 2024
Knowledge Distillation with Multi-granularity Mixture of Priors for Image Super-Resolution
Simiao Li
Yun-feng Zhang
Wei Li
Hanting Chen
Wenjia Wang
Bingyi Jing
Shaohui Lin
Jie Hu
SupR
40
1
0
03 Apr 2024
Improve Knowledge Distillation via Label Revision and Data Selection
Weichao Lan
Yiu-ming Cheung
Qing Xu
Buhua Liu
Zhikai Hu
Mengke Li
Zhenghua Chen
37
2
0
03 Apr 2024
Adaptive Feature Fusion Neural Network for Glaucoma Segmentation on Unseen Fundus Images
Jiyuan Zhong
Hu Ke
Ming Yan
MedIm
26
1
0
02 Apr 2024
Task Integration Distillation for Object Detectors
Hai Su
ZhenWen Jian
Songsen Yu
46
1
0
02 Apr 2024
A Comprehensive Review of Knowledge Distillation in Computer Vision
Sheikh Musa Kaleem
Tufail Rouf
Gousia Habib
Tausifa Jan Saleem
Brejesh Lall
VLM
35
13
0
01 Apr 2024
Previous
1
2
3
4
5
6
...
22
23
24
Next