Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1901.11504
Cited By
Multi-Task Deep Neural Networks for Natural Language Understanding
31 January 2019
Xiaodong Liu
Pengcheng He
Weizhu Chen
Jianfeng Gao
AI4CE
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Multi-Task Deep Neural Networks for Natural Language Understanding"
50 / 541 papers shown
Title
GradTS: A Gradient-Based Automatic Auxiliary Task Selection Method Based on Transformer Networks
Weicheng Ma
Renze Lou
Kai Zhang
Lili Wang
Soroush Vosoughi
23
8
0
13 Sep 2021
CPT: A Pre-Trained Unbalanced Transformer for Both Chinese Language Understanding and Generation
Yunfan Shao
Zhichao Geng
Yitao Liu
Junqi Dai
Hang Yan
Fei Yang
Li Zhe
Hujun Bao
Xipeng Qiu
MedIm
70
149
0
13 Sep 2021
Leveraging Table Content for Zero-shot Text-to-SQL with Meta-Learning
Yongrui Chen
Xinnan Guo
Chaojie Wang
Jian Qiu
Guilin Qi
Meng Wang
Huiying Li
LMTD
VLM
26
10
0
12 Sep 2021
ReasonBERT: Pre-trained to Reason with Distant Supervision
Xiang Deng
Yu-Chuan Su
Alyssa Lees
You Wu
Cong Yu
Huan Sun
ReLM
RALM
OffRL
LRM
19
31
0
10 Sep 2021
Multitask Balanced and Recalibrated Network for Medical Code Prediction
Wei Sun
Shaoxiong Ji
Min Zhang
Pekka Marttinen
33
15
0
06 Sep 2021
Finetuned Language Models Are Zero-Shot Learners
Jason W. Wei
Maarten Bosma
Vincent Zhao
Kelvin Guu
Adams Wei Yu
Brian Lester
Nan Du
Andrew M. Dai
Quoc V. Le
ALM
UQCV
40
3,600
0
03 Sep 2021
CodeT5: Identifier-aware Unified Pre-trained Encoder-Decoder Models for Code Understanding and Generation
Yue Wang
Weishi Wang
Shafiq Joty
Guosheng Lin
252
1,512
0
02 Sep 2021
How Does Adversarial Fine-Tuning Benefit BERT?
J. Ebrahimi
Hao Yang
Wei Zhang
AAML
26
4
0
31 Aug 2021
Span Fine-tuning for Pre-trained Language Models
Rongzhou Bao
Zhuosheng Zhang
Hai Zhao
14
2
0
29 Aug 2021
SMedBERT: A Knowledge-Enhanced Pre-trained Language Model with Structured Semantics for Medical Text Mining
Taolin Zhang
Zerui Cai
Chengyu Wang
Minghui Qiu
Bite Yang
Xiaofeng He
AI4MH
28
52
0
20 Aug 2021
Contextualizing Variation in Text Style Transfer Datasets
S. Schoch
Wanyu Du
Yangfeng Ji
27
5
0
17 Aug 2021
Accurate, yet inconsistent? Consistency Analysis on Language Understanding Models
Myeongjun Jang
D. Kwon
Thomas Lukasiewicz
38
13
0
15 Aug 2021
AMMUS : A Survey of Transformer-based Pretrained Models in Natural Language Processing
Katikapalli Subramanyam Kalyan
A. Rajasekharan
S. Sangeetha
VLM
LM&MA
31
261
0
12 Aug 2021
Leveraging Commonsense Knowledge on Classifying False News and Determining Checkworthiness of Claims
Ipek Baris Schlicht
Erhan Sezerer
Selma Tekir
Oul Han
Zeyd Boukhers
24
0
0
08 Aug 2021
ExBERT: An External Knowledge Enhanced BERT for Natural Language Inference
Amit Gajbhiye
Noura Al Moubayed
S. Bradley
14
9
0
03 Aug 2021
Exceeding the Limits of Visual-Linguistic Multi-Task Learning
Cameron R. Wolfe
Keld T. Lundgaard
VLM
45
2
0
27 Jul 2021
Adaptive Transfer Learning on Graph Neural Networks
Xueting Han
Zhenhuan Huang
Bang An
Jing Bai
30
57
0
19 Jul 2021
FLEX: Unifying Evaluation for Few-Shot NLP
Jonathan Bragg
Arman Cohan
Kyle Lo
Iz Beltagy
208
104
0
15 Jul 2021
A Flexible Multi-Task Model for BERT Serving
Tianwen Wei
Jianwei Qi
Shenghuang He
34
7
0
12 Jul 2021
ChineseBERT: Chinese Pretraining Enhanced by Glyph and Pinyin Information
Zijun Sun
Xiaoya Li
Xiaofei Sun
Yuxian Meng
Xiang Ao
Qing He
Fei Wu
Jiwei Li
SSeg
57
184
0
30 Jun 2021
Specializing Multilingual Language Models: An Empirical Study
Ethan C. Chau
Noah A. Smith
32
27
0
16 Jun 2021
A Semi-supervised Multi-task Learning Approach to Classify Customer Contact Intents
Li Dong
Matthew C. Spencer
Amir Biagi
27
3
0
10 Jun 2021
Parameter-efficient Multi-task Fine-tuning for Transformers via Shared Hypernetworks
Rabeeh Karimi Mahabadi
Sebastian Ruder
Mostafa Dehghani
James Henderson
MoE
39
296
0
08 Jun 2021
Hierarchical Task Learning from Language Instructions with Unified Transformers and Self-Monitoring
Yichi Zhang
J. Chai
25
78
0
07 Jun 2021
MergeDistill: Merging Pre-trained Language Models using Distillation
Simran Khanuja
Melvin Johnson
Partha P. Talukdar
35
16
0
05 Jun 2021
Bi-Granularity Contrastive Learning for Post-Training in Few-Shot Scene
Ruikun Luo
Guanhuan Huang
Xiaojun Quan
CLL
8
10
0
04 Jun 2021
Adjacency List Oriented Relational Fact Extraction via Adaptive Multi-task Learning
Fubang Zhao
Zhuoren Jiang
Yangyang Kang
Changlong Sun
Xiaozhong Liu
14
9
0
03 Jun 2021
Training ELECTRA Augmented with Multi-word Selection
Jiaming Shen
Jialu Liu
Tianqi Liu
Cong Yu
Jiawei Han
31
9
0
31 May 2021
Weighted Training for Cross-Task Learning
Shuxiao Chen
K. Crammer
Han He
Dan Roth
Weijie J. Su
22
28
0
28 May 2021
TreeBERT: A Tree-Based Pre-Trained Model for Programming Language
Xue Jiang
Zhuoran Zheng
Chen Lyu
Liang Li
Lei Lyu
27
90
0
26 May 2021
Super Tickets in Pre-Trained Language Models: From Model Compression to Improving Generalization
Chen Liang
Simiao Zuo
Minshuo Chen
Haoming Jiang
Xiaodong Liu
Pengcheng He
T. Zhao
Weizhu Chen
20
68
0
25 May 2021
Multi-Task Learning of Generation and Classification for Emotion-Aware Dialogue Response Generation
Tatsuya Ide
Daisuke Kawahara
13
26
0
25 May 2021
True Few-Shot Learning with Language Models
Ethan Perez
Douwe Kiela
Kyunghyun Cho
21
428
0
24 May 2021
Killing One Bird with Two Stones: Model Extraction and Attribute Inference Attacks against BERT-based APIs
Chen Chen
Xuanli He
Lingjuan Lyu
Fangzhao Wu
SILM
MIACV
65
7
0
23 May 2021
Training Bi-Encoders for Word Sense Disambiguation
Harsh Kohli
51
4
0
21 May 2021
Out-of-Manifold Regularization in Contextual Embedding Space for Text Classification
Seonghyeon Lee
Dongha Lee
Hwanjo Yu
24
4
0
14 May 2021
Go Beyond Plain Fine-tuning: Improving Pretrained Models for Social Commonsense
Ting-Yun Chang
Yang Liu
Karthik Gopalakrishnan
Behnam Hedayatnia
Pei Zhou
Dilek Z. Hakkani-Tür
ReLM
VLM
AI4MH
LRM
16
1
0
12 May 2021
OCHADAI-KYOTO at SemEval-2021 Task 1: Enhancing Model Generalization and Robustness for Lexical Complexity Prediction
Y. Taya
L. Pereira
Fei Cheng
Ichiro Kobayashi
36
0
0
12 May 2021
WASSA@IITK at WASSA 2021: Multi-task Learning and Transformer Finetuning for Emotion Classification and Empathy Prediction
Jay Mundra
Rohan Gupta
Sagnik Mukherjee
11
14
0
20 Apr 2021
LaTeX-Numeric: Language-agnostic Text attribute eXtraction for E-commerce Numeric Attributes
Kartik Mehta
I. Oprea
Nikhil Rasiwasia
28
6
0
19 Apr 2021
BigGreen at SemEval-2021 Task 1: Lexical Complexity Prediction with Assembly Models
A. Islam
Weicheng Ma
Soroush Vosoughi
30
4
0
19 Apr 2021
Multilingual and Cross-Lingual Intent Detection from Spoken Data
D. Gerz
Pei-hao Su
Razvan Kusztos
Avishek Mondal
M. Lis
Eshan Singhal
N. Mrksic
Tsung-Hsien Wen
Ivan Vulić
24
35
0
17 Apr 2021
AMMU : A Survey of Transformer-based Biomedical Pretrained Language Models
Katikapalli Subramanyam Kalyan
A. Rajasekharan
S. Sangeetha
LM&MA
MedIm
31
164
0
16 Apr 2021
MT-Opt: Continuous Multi-Task Robotic Reinforcement Learning at Scale
Dmitry Kalashnikov
Jacob Varley
Yevgen Chebotar
Benjamin Swanson
Rico Jonschkowski
Chelsea Finn
Sergey Levine
Karol Hausman
OffRL
47
272
0
16 Apr 2021
What's in your Head? Emergent Behaviour in Multi-Task Transformer Models
Mor Geva
Uri Katz
Aviv Ben-Arie
Jonathan Berant
LRM
48
11
0
13 Apr 2021
Structural analysis of an all-purpose question answering model
Vincent Micheli
Quentin Heinrich
Franccois Fleuret
Wacim Belblidia
26
3
0
13 Apr 2021
On Unifying Misinformation Detection
Nayeon Lee
Belinda Z. Li
Sinong Wang
Pascale Fung
Hao Ma
Wen-tau Yih
Madian Khabsa
14
24
0
12 Apr 2021
Adversarial Regularization as Stackelberg Game: An Unrolled Optimization Approach
Simiao Zuo
Chen Liang
Haoming Jiang
Xiaodong Liu
Pengcheng He
Jianfeng Gao
Weizhu Chen
T. Zhao
66
9
0
11 Apr 2021
Lone Pine at SemEval-2021 Task 5: Fine-Grained Detection of Hate Speech Using BERToxic
Yakoob Khan
Weicheng Ma
Soroush Vosoughi
18
7
0
08 Apr 2021
CodeTrans: Towards Cracking the Language of Silicon's Code Through Self-Supervised Deep Learning and High Performance Computing
Ahmed Elnaggar
Wei Ding
Llion Jones
Tom Gibbs
Tamas B. Fehér
Christoph Angerer
Silvia Severini
Florian Matthes
B. Rost
28
72
0
06 Apr 2021
Previous
1
2
3
...
5
6
7
...
9
10
11
Next