ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1706.00384
  4. Cited By
Deep Mutual Learning

Deep Mutual Learning

1 June 2017
Ying Zhang
Tao Xiang
Timothy M. Hospedales
Huchuan Lu
    FedML
ArXivPDFHTML

Papers citing "Deep Mutual Learning"

50 / 710 papers shown
Title
What is Next when Sequential Prediction Meets Implicitly Hard
  Interaction?
What is Next when Sequential Prediction Meets Implicitly Hard Interaction?
Kaixi Hu
Lin Li
Qing Xie
Jianquan Liu
Xiaohui Tao
27
19
0
14 Feb 2022
Real World Large Scale Recommendation Systems Reproducibility and Smooth
  Activations
Real World Large Scale Recommendation Systems Reproducibility and Smooth Activations
G. Shamir
Dong Lin
HAI
OffRL
25
6
0
14 Feb 2022
Open-set Adversarial Defense with Clean-Adversarial Mutual Learning
Open-set Adversarial Defense with Clean-Adversarial Mutual Learning
Rui Shao
Pramuditha Perera
Pong C. Yuen
Vishal M. Patel
AAML
28
32
0
12 Feb 2022
Personalized Federated Learning via Convex Clustering
Personalized Federated Learning via Convex Clustering
Aleksandar Armacki
Dragana Bajović
D. Jakovetić
S. Kar
FedML
21
16
0
01 Feb 2022
Learning Fast, Learning Slow: A General Continual Learning Method based
  on Complementary Learning System
Learning Fast, Learning Slow: A General Continual Learning Method based on Complementary Learning System
Elahe Arani
F. Sarfraz
Bahram Zonooz
CLL
93
125
0
29 Jan 2022
PCL: Peer-Contrastive Learning with Diverse Augmentations for
  Unsupervised Sentence Embeddings
PCL: Peer-Contrastive Learning with Diverse Augmentations for Unsupervised Sentence Embeddings
Qiyu Wu
Chongyang Tao
Tao Shen
Can Xu
Xiubo Geng
Daxin Jiang
SSL
27
34
0
28 Jan 2022
Image-to-Video Re-Identification via Mutual Discriminative Knowledge
  Transfer
Image-to-Video Re-Identification via Mutual Discriminative Knowledge Transfer
Pichao Wang
Fan Wang
Hao Li
24
0
0
21 Jan 2022
STURE: Spatial-Temporal Mutual Representation Learning for Robust Data
  Association in Online Multi-Object Tracking
STURE: Spatial-Temporal Mutual Representation Learning for Robust Data Association in Online Multi-Object Tracking
Haidong Wang
Zhiyong Li
Yaping Li
Ke Nai
Ming Wen
VOT
30
7
0
18 Jan 2022
GearNet: Stepwise Dual Learning for Weakly Supervised Domain Adaptation
GearNet: Stepwise Dual Learning for Weakly Supervised Domain Adaptation
Renchunzi Xie
Hongxin Wei
Lei Feng
Bo An
26
11
0
16 Jan 2022
FedDTG:Federated Data-Free Knowledge Distillation via Three-Player
  Generative Adversarial Networks
FedDTG:Federated Data-Free Knowledge Distillation via Three-Player Generative Adversarial Networks
Zhenyuan Zhang
Tao Shen
Jie M. Zhang
Chao-Xiang Wu
FedML
20
13
0
10 Jan 2022
Role of Data Augmentation Strategies in Knowledge Distillation for
  Wearable Sensor Data
Role of Data Augmentation Strategies in Knowledge Distillation for Wearable Sensor Data
Eunyeong Jeon
Anirudh Som
Ankita Shukla
Kristina Hasanaj
M. Buman
Pavan Turaga
40
11
0
01 Jan 2022
Data-Free Knowledge Transfer: A Survey
Data-Free Knowledge Transfer: A Survey
Yuang Liu
Wei Zhang
Jun Wang
Jianyong Wang
40
48
0
31 Dec 2021
A formal approach to good practices in Pseudo-Labeling for Unsupervised
  Domain Adaptive Re-Identification
A formal approach to good practices in Pseudo-Labeling for Unsupervised Domain Adaptive Re-Identification
Fabian Dubourvieux
Romaric Audigier
Angélique Loesch
Samia Ainouz
S. Canu
29
7
0
24 Dec 2021
A Label Dependence-aware Sequence Generation Model for Multi-level
  Implicit Discourse Relation Recognition
A Label Dependence-aware Sequence Generation Model for Multi-level Implicit Discourse Relation Recognition
Changxing Wu
Liuwen Cao
Yubin Ge
Yang Liu
Min Zhang
Jinsong Su
34
32
0
22 Dec 2021
Incremental Cross-view Mutual Distillation for Self-supervised Medical
  CT Synthesis
Incremental Cross-view Mutual Distillation for Self-supervised Medical CT Synthesis
Chaowei Fang
Liang Wang
Dingwen Zhang
Jun Xu
Yixuan Yuan
Junwei Han
OOD
37
13
0
20 Dec 2021
Anomaly Discovery in Semantic Segmentation via Distillation Comparison
  Networks
Anomaly Discovery in Semantic Segmentation via Distillation Comparison Networks
Huan Zhou
Shi Gong
Yu Zhou
Zengqiang Zheng
Ronghua Liu
Xiang Bai
31
1
0
18 Dec 2021
Distillation of Human-Object Interaction Contexts for Action Recognition
Distillation of Human-Object Interaction Contexts for Action Recognition
Muna Almushyti
Frederick W. Li
39
3
0
17 Dec 2021
TransZero++: Cross Attribute-Guided Transformer for Zero-Shot Learning
TransZero++: Cross Attribute-Guided Transformer for Zero-Shot Learning
Shiming Chen
Zi-Quan Hong
Wenjin Hou
Guosen Xie
Yibing Song
Jian-jun Zhao
Xinge You
Shuicheng Yan
Ling Shao
ViT
22
44
0
16 Dec 2021
Mutual Adversarial Training: Learning together is better than going
  alone
Mutual Adversarial Training: Learning together is better than going alone
Jiang-Long Liu
Chun Pong Lau
Hossein Souri
S. Feizi
Ramalingam Chellappa
OOD
AAML
48
24
0
09 Dec 2021
Boosting Contrastive Learning with Relation Knowledge Distillation
Boosting Contrastive Learning with Relation Knowledge Distillation
Kai Zheng
Yuanjiang Wang
Ye Yuan
SSL
19
13
0
08 Dec 2021
Handwritten Mathematical Expression Recognition via Attention
  Aggregation based Bi-directional Mutual Learning
Handwritten Mathematical Expression Recognition via Attention Aggregation based Bi-directional Mutual Learning
Xiaohang Bian
Bo Qin
Xiaozhe Xin
Jianwu Li
Xuefeng Su
Yanfeng Wang
40
49
0
07 Dec 2021
Bootstrapping ViTs: Towards Liberating Vision Transformers from
  Pre-training
Bootstrapping ViTs: Towards Liberating Vision Transformers from Pre-training
Haofei Zhang
Jiarui Duan
Mengqi Xue
Mingli Song
Li Sun
Xiuming Zhang
ViT
AI4CE
30
16
0
07 Dec 2021
Extracting knowledge from features with multilevel abstraction
Extracting knowledge from features with multilevel abstraction
Jin-Siang Lin
Zhaoyang Li
24
0
0
04 Dec 2021
From Consensus to Disagreement: Multi-Teacher Distillation for
  Semi-Supervised Relation Extraction
From Consensus to Disagreement: Multi-Teacher Distillation for Semi-Supervised Relation Extraction
Wanli Li
T. Qian
25
2
0
02 Dec 2021
Vision Pair Learning: An Efficient Training Framework for Image
  Classification
Vision Pair Learning: An Efficient Training Framework for Image Classification
Bei Tong
Xiaoyuan Yu
ViT
25
0
0
02 Dec 2021
Improved Knowledge Distillation via Adversarial Collaboration
Improved Knowledge Distillation via Adversarial Collaboration
Zhiqiang Liu
Chengkai Huang
Yanxia Liu
31
2
0
29 Nov 2021
Self-Distilled Self-Supervised Representation Learning
Self-Distilled Self-Supervised Representation Learning
Jiho Jang
Seonhoon Kim
Kiyoon Yoo
Chaerin Kong
Jang-Hyun Kim
Nojun Kwak
SSL
28
14
0
25 Nov 2021
Altering Backward Pass Gradients improves Convergence
Altering Backward Pass Gradients improves Convergence
Bishshoy Das
M. Mondal
Brejesh Lall
S. Joshi
Sumantra Dutta Roy
24
0
0
24 Nov 2021
EvDistill: Asynchronous Events to End-task Learning via Bidirectional
  Reconstruction-guided Cross-modal Knowledge Distillation
EvDistill: Asynchronous Events to End-task Learning via Bidirectional Reconstruction-guided Cross-modal Knowledge Distillation
Lin Wang
Yujeong Chae
Sung-Hoon Yoon
Tae-Kyun Kim
Kuk-Jin Yoon
47
64
0
24 Nov 2021
Semi-Online Knowledge Distillation
Semi-Online Knowledge Distillation
Zhiqiang Liu
Yanxia Liu
Chengkai Huang
27
5
0
23 Nov 2021
Decentralized Federated Learning through Proxy Model Sharing
Decentralized Federated Learning through Proxy Model Sharing
Shivam Kalra
Junfeng Wen
Jesse C. Cresswell
M. Volkovs
Hamid R. Tizhoosh
FedML
21
94
0
22 Nov 2021
Towards Comprehensive Monocular Depth Estimation: Multiple Heads Are
  Better Than One
Towards Comprehensive Monocular Depth Estimation: Multiple Heads Are Better Than One
Shuwei Shao
Ran Li
Z. Pei
Zhong Liu
Weihai Chen
Wentao Zhu
Xingming Wu
Baochang Zhang
ViT
MDE
36
11
0
16 Nov 2021
Edge-Cloud Polarization and Collaboration: A Comprehensive Survey for AI
Edge-Cloud Polarization and Collaboration: A Comprehensive Survey for AI
Jiangchao Yao
Shengyu Zhang
Yang Yao
Feng Wang
Jianxin Ma
...
Kun Kuang
Chao-Xiang Wu
Fei Wu
Jingren Zhou
Hongxia Yang
28
91
0
11 Nov 2021
Learning to Disentangle Scenes for Person Re-identification
Learning to Disentangle Scenes for Person Re-identification
Xianghao Zang
Gezhong Li
Wei-Nan Gao
Xiujun Shu
24
33
0
10 Nov 2021
Does Thermal data make the detection systems more reliable?
Does Thermal data make the detection systems more reliable?
Shruthi Gowda
Bahram Zonooz
Elahe Arani
14
1
0
09 Nov 2021
Exploiting Robust Unsupervised Video Person Re-identification
Exploiting Robust Unsupervised Video Person Re-identification
Xianghao Zang
Gezhong Li
Wei-Nan Gao
Xiujun Shu
VOS
31
11
0
09 Nov 2021
A Survey on Green Deep Learning
A Survey on Green Deep Learning
Jingjing Xu
Wangchunshu Zhou
Zhiyi Fu
Hao Zhou
Lei Li
VLM
86
83
0
08 Nov 2021
Arch-Net: Model Distillation for Architecture Agnostic Model Deployment
Arch-Net: Model Distillation for Architecture Agnostic Model Deployment
Weixin Xu
Zipeng Feng
Shuangkang Fang
Song Yuan
Yi Yang
Shuchang Zhou
MQ
30
1
0
01 Nov 2021
PP-ShiTu: A Practical Lightweight Image Recognition System
PP-ShiTu: A Practical Lightweight Image Recognition System
Shengyun Wei
Ruoyu Guo
Cheng Cui
Bin Lu
Shuilong Dong
...
Xueying Lyu
Qiwen Liu
Xiaoguang Hu
Dianhai Yu
Yanjun Ma
CVBM
26
6
0
01 Nov 2021
Towards Model Agnostic Federated Learning Using Knowledge Distillation
Towards Model Agnostic Federated Learning Using Knowledge Distillation
A. Afonin
Sai Praneeth Karimireddy
FedML
30
45
0
28 Oct 2021
GenURL: A General Framework for Unsupervised Representation Learning
GenURL: A General Framework for Unsupervised Representation Learning
Siyuan Li
Zicheng Liu
Z. Zang
Di Wu
Zhiyuan Chen
Stan Z. Li
OOD
3DGS
OffRL
39
9
0
27 Oct 2021
Self-Supervised Knowledge Transfer via Loosely Supervised Auxiliary
  Tasks
Self-Supervised Knowledge Transfer via Loosely Supervised Auxiliary Tasks
Seungbum Hong
Jihun Yoon
Junmo Kim
Min-Kook Choi
SSL
16
1
0
25 Oct 2021
MUSE: Feature Self-Distillation with Mutual Information and
  Self-Information
MUSE: Feature Self-Distillation with Mutual Information and Self-Information
Yunpeng Gong
Ye Yu
Gaurav Mittal
Greg Mori
Mei Chen
SSL
32
2
0
25 Oct 2021
An attention-driven hierarchical multi-scale representation for visual
  recognition
An attention-driven hierarchical multi-scale representation for visual recognition
Zachary Wharton
Chenglong Bao
Asish Bera
BDL
20
1
0
23 Oct 2021
Augmenting Knowledge Distillation With Peer-To-Peer Mutual Learning For
  Model Compression
Augmenting Knowledge Distillation With Peer-To-Peer Mutual Learning For Model Compression
Usma Niyaz
Deepti R. Bathula
26
8
0
21 Oct 2021
Deep Learning for HDR Imaging: State-of-the-Art and Future Trends
Deep Learning for HDR Imaging: State-of-the-Art and Future Trends
Lin Wang
Kuk-Jin Yoon
21
87
0
20 Oct 2021
Adaptive Distillation: Aggregating Knowledge from Multiple Paths for
  Efficient Distillation
Adaptive Distillation: Aggregating Knowledge from Multiple Paths for Efficient Distillation
Sumanth Chennupati
Mohammad Mahdi Kamani
Zhongwei Cheng
Lin Chen
35
4
0
19 Oct 2021
Mitigating Memorization of Noisy Labels via Regularization between
  Representations
Mitigating Memorization of Noisy Labels via Regularization between Representations
Hao Cheng
Zhaowei Zhu
Xing Sun
Yang Liu
NoLa
43
28
0
18 Oct 2021
FrugalScore: Learning Cheaper, Lighter and Faster Evaluation Metricsfor
  Automatic Text Generation
FrugalScore: Learning Cheaper, Lighter and Faster Evaluation Metricsfor Automatic Text Generation
Moussa Kamal Eddine
Guokan Shang
A. Tixier
Michalis Vazirgiannis
26
25
0
16 Oct 2021
FedMe: Federated Learning via Model Exchange
FedMe: Federated Learning via Model Exchange
Koji Matsuda
Yuya Sasaki
Chuan Xiao
Makoto Onizuka
FedML
50
19
0
15 Oct 2021
Previous
123...789...131415
Next