MLKD-BERT: Multi-level Knowledge Distillation for Pre-trained Language
  Models

MLKD-BERT: Multi-level Knowledge Distillation for Pre-trained Language Models

    KELM

Papers citing "MLKD-BERT: Multi-level Knowledge Distillation for Pre-trained Language Models"

17 / 17 papers shown
Title
ERNIE: Enhanced Language Representation with Informative Entities
ERNIE: Enhanced Language Representation with Informative Entities
Zhengyan Zhang
Xu Han
Zhiyuan Liu
Xin Jiang
Maosong Sun
Qun Liu
113
1,402
0
17 May 2019

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.