Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1902.00751
Cited By
v1
v2 (latest)
Parameter-Efficient Transfer Learning for NLP
2 February 2019
N. Houlsby
A. Giurgiu
Stanislaw Jastrzebski
Bruna Morrone
Quentin de Laroussilhe
Andrea Gesmundo
Mona Attariyan
Sylvain Gelly
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Parameter-Efficient Transfer Learning for NLP"
50 / 2,860 papers shown
Title
MultiCQA: Zero-Shot Transfer of Self-Supervised Text Matching Models on a Massive Scale
Andreas Rucklé
Jonas Pfeiffer
Iryna Gurevych
83
38
0
02 Oct 2020
Scalable Transfer Learning with Expert Models
J. Puigcerver
C. Riquelme
Basil Mustafa
Cédric Renggli
André Susano Pinto
Sylvain Gelly
Daniel Keysers
N. Houlsby
138
65
0
28 Sep 2020
Deep Transformers with Latent Depth
Xian Li
Asa Cooper Stickland
Yuqing Tang
X. Kong
73
23
0
28 Sep 2020
Conditionally Adaptive Multi-Task Learning: Improving Transfer Learning in NLP Using Fewer Parameters & Less Data
Jonathan Pilault
Amine Elhattami
C. Pal
CLL
MoE
94
92
0
19 Sep 2020
Reusing a Pretrained Language Model on Languages with Limited Corpora for Unsupervised NMT
Alexandra Chronopoulou
Dario Stojanovski
Alexander Fraser
113
33
0
16 Sep 2020
Efficient Transformers: A Survey
Yi Tay
Mostafa Dehghani
Dara Bahri
Donald Metzler
VLM
264
1,137
0
14 Sep 2020
On Transfer Learning of Traditional Frequency and Time Domain Features in Turning
Melih C. Yesilli
Firas A. Khasawneh
36
8
0
28 Aug 2020
The Adapter-Bot: All-In-One Controllable Conversational Model
Andrea Madotto
Zhaojiang Lin
Yejin Bang
Pascale Fung
104
63
0
28 Aug 2020
Improvement of a dedicated model for open domain persona-aware dialogue generation
Qiang Han
54
0
0
27 Aug 2020
Is Supervised Syntactic Parsing Beneficial for Language Understanding? An Empirical Investigation
Goran Glavaš
Ivan Vulić
100
69
0
15 Aug 2020
Adaptation Algorithms for Neural Network-Based Speech Recognition: An Overview
P. Bell
Joachim Fainberg
Ondˇrej Klejch
Jinyu Li
Steve Renals
P. Swietojanski
129
78
0
14 Aug 2020
Adaptable Multi-Domain Language Model for Transformer ASR
Taewoo Lee
Min-Joong Lee
Tae Gyoon Kang
Seokyeong Jung
Minseok Kwon
...
Ho-Gyeong Kim
Jiseung Jeong
Jihyun Lee
Hosik Lee
Y. S. Choi
56
18
0
14 Aug 2020
Detecting Transaction-based Tax Evasion Activities on Social Media Platforms Using Multi-modal Deep Neural Networks
Lelin Zhang
Xi Nan
Eva Huang
Sidong Liu
23
3
0
27 Jul 2020
Drinking from a Firehose: Continual Learning with Web-scale Natural Language
Hexiang Hu
Ozan Sener
Fei Sha
V. Koltun
CLL
62
27
0
18 Jul 2020
Continual BERT: Continual Learning for Adaptive Extractive Summarization of COVID-19 Literature
Jongjin Park
CLL
68
16
0
07 Jul 2020
Composed Fine-Tuning: Freezing Pre-Trained Denoising Autoencoders for Improved Generalization
Sang Michael Xie
Tengyu Ma
Percy Liang
143
15
0
29 Jun 2020
Does Non-COVID19 Lung Lesion Help? Investigating Transferability in COVID-19 CT Image Segmentation
Yixin Wang
Yao Zhang
Yang Liu
Jiang Tian
Cheng Zhong
Zhongchao Shi
Yang Zhang
Zhiqiang He
78
62
0
23 Jun 2020
Revisiting Few-sample BERT Fine-tuning
Tianyi Zhang
Felix Wu
Arzoo Katiyar
Kilian Q. Weinberger
Yoav Artzi
184
446
0
10 Jun 2020
A Survey on Transfer Learning in Natural Language Processing
Zaid Alyafeai
Maged S. Alshaibani
Irfan Ahmad
91
75
0
31 May 2020
Common Sense or World Knowledge? Investigating Adapter-Based Knowledge Injection into Pretrained Transformers
Anne Lauscher
Olga Majewska
Leonardo F. R. Ribeiro
Iryna Gurevych
Nikolai Rozanov
Goran Glavaš
KELM
84
81
0
24 May 2020
Low Resource Multi-Task Sequence Tagging -- Revisiting Dynamic Conditional Random Fields
Jonas Pfeiffer
Edwin Simpson
Iryna Gurevych
67
5
0
01 May 2020
AdapterFusion: Non-Destructive Task Composition for Transfer Learning
Jonas Pfeiffer
Aishwarya Kamath
Andreas Rucklé
Kyunghyun Cho
Iryna Gurevych
CLL
MoMe
269
861
0
01 May 2020
MAD-X: An Adapter-Based Framework for Multi-Task Cross-Lingual Transfer
Jonas Pfeiffer
Ivan Vulić
Iryna Gurevych
Sebastian Ruder
180
631
0
30 Apr 2020
Recipes for Adapting Pre-trained Monolingual and Multilingual Models to Machine Translation
Asa Cooper Stickland
Xian Li
Marjan Ghazvininejad
105
46
0
30 Apr 2020
UDapter: Language Adaptation for Truly Universal Dependency Parsing
Ahmet Üstün
Arianna Bisazza
G. Bouma
Gertjan van Noord
76
117
0
29 Apr 2020
General Purpose Text Embeddings from Pre-trained Language Models for Scalable Inference
Jingfei Du
Myle Ott
Haoran Li
Xing Zhou
Veselin Stoyanov
AI4CE
68
10
0
29 Apr 2020
Masking as an Efficient Alternative to Finetuning for Pretrained Language Models
Mengjie Zhao
Tao R. Lin
Fei Mi
Martin Jaggi
Hinrich Schütze
77
121
0
26 Apr 2020
How fine can fine-tuning be? Learning efficient language models
Evani Radiya-Dixit
Xin Wang
53
66
0
24 Apr 2020
Fine-tuning Multi-hop Question Answering with Hierarchical Graph Network
Guanming Xiong
134
0
0
20 Apr 2020
Beyond Fine-tuning: Few-Sample Sentence Embedding Transfer
Siddhant Garg
Rohit Kumar Sharma
Yingyu Liang
82
4
0
10 Apr 2020
Exploring Versatile Generative Language Model Via Parameter-Efficient Transfer Learning
Zhaojiang Lin
Andrea Madotto
Pascale Fung
121
164
0
08 Apr 2020
Meta-Learning for Few-Shot NMT Adaptation
Amr Sharaf
Hany Hassan
Hal Daumé
66
35
0
06 Apr 2020
A Survey of Deep Learning for Scientific Discovery
M. Raghu
Erica Schmidt
OOD
AI4CE
182
123
0
26 Mar 2020
Pre-trained Models for Natural Language Processing: A Survey
Xipeng Qiu
Tianxiang Sun
Yige Xu
Yunfan Shao
Ning Dai
Xuanjing Huang
LM&MA
VLM
453
1,500
0
18 Mar 2020
A Primer in BERTology: What we know about how BERT works
Anna Rogers
Olga Kovaleva
Anna Rumshisky
OffRL
170
1,511
0
27 Feb 2020
Exploring BERT Parameter Efficiency on the Stanford Question Answering Dataset v2.0
Eric Hulburd
53
5
0
25 Feb 2020
Improving BERT Fine-Tuning via Self-Ensemble and Self-Distillation
Yige Xu
Xipeng Qiu
L. Zhou
Xuanjing Huang
83
68
0
24 Feb 2020
Contextual Lensing of Universal Sentence Representations
J. Kiros
57
5
0
20 Feb 2020
K-Adapter: Infusing Knowledge into Pre-Trained Models with Adapters
Ruize Wang
Duyu Tang
Nan Duan
Zhongyu Wei
Xuanjing Huang
Jianshu Ji
Guihong Cao
Daxin Jiang
Ming Zhou
KELM
177
557
0
05 Feb 2020
Parameter-Efficient Transfer from Sequential Behaviors for User Modeling and Recommendation
Fajie Yuan
Xiangnan He
Alexandros Karatzoglou
Liguang Zhang
59
0
0
13 Jan 2020
Neural Network Surgery with Sets
Jonathan Raiman
Susan Zhang
Christy Dennison
58
4
0
13 Dec 2019
Pre-train and Plug-in: Flexible Conditional Text Generation with Variational Auto-Encoders
Yu Duan
Canwen Xu
Jiaxin Pei
Jialong Han
Chenliang Li
150
43
0
10 Nov 2019
Learning to Few-Shot Learn Across Diverse Natural Language Classification Tasks
Trapit Bansal
Rishikesh Jha
Andrew McCallum
SSL
118
121
0
10 Nov 2019
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization
Haoming Jiang
Pengcheng He
Weizhu Chen
Xiaodong Liu
Jianfeng Gao
T. Zhao
143
563
0
08 Nov 2019
What Would Elsa Do? Freezing Layers During Transformer Fine-Tuning
Jaejun Lee
Raphael Tang
Jimmy J. Lin
69
127
0
08 Nov 2019
Towards Domain Adaptation from Limited Data for Question Answering Using Deep Neural Networks
Timothy J. Hazen
Shehzaad Dhuliawala
Daniel Boies
OOD
68
19
0
06 Nov 2019
On the Cross-lingual Transferability of Monolingual Representations
Mikel Artetxe
Sebastian Ruder
Dani Yogatama
289
801
0
25 Oct 2019
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
Colin Raffel
Noam M. Shazeer
Adam Roberts
Katherine Lee
Sharan Narang
Michael Matena
Yanqi Zhou
Wei Li
Peter J. Liu
AIMat
1.0K
20,462
0
23 Oct 2019
You May Not Need Order in Time Series Forecasting
Yunkai Zhang
Qiao Jiang
Shurui Li
Xiaoyong Jin
Xueying Ma
Xifeng Yan
AI4TS
28
3
0
21 Oct 2019
SesameBERT: Attention for Anywhere
Ta-Chun Su
Hsiang-Chih Cheng
66
7
0
08 Oct 2019
Previous
1
2
3
...
56
57
58
Next