ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1902.00751
  4. Cited By
Parameter-Efficient Transfer Learning for NLP
v1v2 (latest)

Parameter-Efficient Transfer Learning for NLP

2 February 2019
N. Houlsby
A. Giurgiu
Stanislaw Jastrzebski
Bruna Morrone
Quentin de Laroussilhe
Andrea Gesmundo
Mona Attariyan
Sylvain Gelly
ArXiv (abs)PDFHTML

Papers citing "Parameter-Efficient Transfer Learning for NLP"

50 / 2,860 papers shown
Title
MultiCQA: Zero-Shot Transfer of Self-Supervised Text Matching Models on
  a Massive Scale
MultiCQA: Zero-Shot Transfer of Self-Supervised Text Matching Models on a Massive Scale
Andreas Rucklé
Jonas Pfeiffer
Iryna Gurevych
83
38
0
02 Oct 2020
Scalable Transfer Learning with Expert Models
Scalable Transfer Learning with Expert Models
J. Puigcerver
C. Riquelme
Basil Mustafa
Cédric Renggli
André Susano Pinto
Sylvain Gelly
Daniel Keysers
N. Houlsby
138
65
0
28 Sep 2020
Deep Transformers with Latent Depth
Deep Transformers with Latent Depth
Xian Li
Asa Cooper Stickland
Yuqing Tang
X. Kong
73
23
0
28 Sep 2020
Conditionally Adaptive Multi-Task Learning: Improving Transfer Learning
  in NLP Using Fewer Parameters & Less Data
Conditionally Adaptive Multi-Task Learning: Improving Transfer Learning in NLP Using Fewer Parameters & Less Data
Jonathan Pilault
Amine Elhattami
C. Pal
CLLMoE
94
92
0
19 Sep 2020
Reusing a Pretrained Language Model on Languages with Limited Corpora
  for Unsupervised NMT
Reusing a Pretrained Language Model on Languages with Limited Corpora for Unsupervised NMT
Alexandra Chronopoulou
Dario Stojanovski
Alexander Fraser
113
33
0
16 Sep 2020
Efficient Transformers: A Survey
Efficient Transformers: A Survey
Yi Tay
Mostafa Dehghani
Dara Bahri
Donald Metzler
VLM
264
1,137
0
14 Sep 2020
On Transfer Learning of Traditional Frequency and Time Domain Features
  in Turning
On Transfer Learning of Traditional Frequency and Time Domain Features in Turning
Melih C. Yesilli
Firas A. Khasawneh
36
8
0
28 Aug 2020
The Adapter-Bot: All-In-One Controllable Conversational Model
The Adapter-Bot: All-In-One Controllable Conversational Model
Andrea Madotto
Zhaojiang Lin
Yejin Bang
Pascale Fung
104
63
0
28 Aug 2020
Improvement of a dedicated model for open domain persona-aware dialogue
  generation
Improvement of a dedicated model for open domain persona-aware dialogue generation
Qiang Han
54
0
0
27 Aug 2020
Is Supervised Syntactic Parsing Beneficial for Language Understanding?
  An Empirical Investigation
Is Supervised Syntactic Parsing Beneficial for Language Understanding? An Empirical Investigation
Goran Glavaš
Ivan Vulić
100
69
0
15 Aug 2020
Adaptation Algorithms for Neural Network-Based Speech Recognition: An
  Overview
Adaptation Algorithms for Neural Network-Based Speech Recognition: An Overview
P. Bell
Joachim Fainberg
Ondˇrej Klejch
Jinyu Li
Steve Renals
P. Swietojanski
129
78
0
14 Aug 2020
Adaptable Multi-Domain Language Model for Transformer ASR
Adaptable Multi-Domain Language Model for Transformer ASR
Taewoo Lee
Min-Joong Lee
Tae Gyoon Kang
Seokyeong Jung
Minseok Kwon
...
Ho-Gyeong Kim
Jiseung Jeong
Jihyun Lee
Hosik Lee
Y. S. Choi
56
18
0
14 Aug 2020
Detecting Transaction-based Tax Evasion Activities on Social Media
  Platforms Using Multi-modal Deep Neural Networks
Detecting Transaction-based Tax Evasion Activities on Social Media Platforms Using Multi-modal Deep Neural Networks
Lelin Zhang
Xi Nan
Eva Huang
Sidong Liu
23
3
0
27 Jul 2020
Drinking from a Firehose: Continual Learning with Web-scale Natural
  Language
Drinking from a Firehose: Continual Learning with Web-scale Natural Language
Hexiang Hu
Ozan Sener
Fei Sha
V. Koltun
CLL
62
27
0
18 Jul 2020
Continual BERT: Continual Learning for Adaptive Extractive Summarization
  of COVID-19 Literature
Continual BERT: Continual Learning for Adaptive Extractive Summarization of COVID-19 Literature
Jongjin Park
CLL
68
16
0
07 Jul 2020
Composed Fine-Tuning: Freezing Pre-Trained Denoising Autoencoders for
  Improved Generalization
Composed Fine-Tuning: Freezing Pre-Trained Denoising Autoencoders for Improved Generalization
Sang Michael Xie
Tengyu Ma
Percy Liang
143
15
0
29 Jun 2020
Does Non-COVID19 Lung Lesion Help? Investigating Transferability in
  COVID-19 CT Image Segmentation
Does Non-COVID19 Lung Lesion Help? Investigating Transferability in COVID-19 CT Image Segmentation
Yixin Wang
Yao Zhang
Yang Liu
Jiang Tian
Cheng Zhong
Zhongchao Shi
Yang Zhang
Zhiqiang He
78
62
0
23 Jun 2020
Revisiting Few-sample BERT Fine-tuning
Revisiting Few-sample BERT Fine-tuning
Tianyi Zhang
Felix Wu
Arzoo Katiyar
Kilian Q. Weinberger
Yoav Artzi
184
446
0
10 Jun 2020
A Survey on Transfer Learning in Natural Language Processing
A Survey on Transfer Learning in Natural Language Processing
Zaid Alyafeai
Maged S. Alshaibani
Irfan Ahmad
91
75
0
31 May 2020
Common Sense or World Knowledge? Investigating Adapter-Based Knowledge
  Injection into Pretrained Transformers
Common Sense or World Knowledge? Investigating Adapter-Based Knowledge Injection into Pretrained Transformers
Anne Lauscher
Olga Majewska
Leonardo F. R. Ribeiro
Iryna Gurevych
Nikolai Rozanov
Goran Glavaš
KELM
84
81
0
24 May 2020
Low Resource Multi-Task Sequence Tagging -- Revisiting Dynamic
  Conditional Random Fields
Low Resource Multi-Task Sequence Tagging -- Revisiting Dynamic Conditional Random Fields
Jonas Pfeiffer
Edwin Simpson
Iryna Gurevych
67
5
0
01 May 2020
AdapterFusion: Non-Destructive Task Composition for Transfer Learning
AdapterFusion: Non-Destructive Task Composition for Transfer Learning
Jonas Pfeiffer
Aishwarya Kamath
Andreas Rucklé
Kyunghyun Cho
Iryna Gurevych
CLLMoMe
269
861
0
01 May 2020
MAD-X: An Adapter-Based Framework for Multi-Task Cross-Lingual Transfer
MAD-X: An Adapter-Based Framework for Multi-Task Cross-Lingual Transfer
Jonas Pfeiffer
Ivan Vulić
Iryna Gurevych
Sebastian Ruder
180
631
0
30 Apr 2020
Recipes for Adapting Pre-trained Monolingual and Multilingual Models to
  Machine Translation
Recipes for Adapting Pre-trained Monolingual and Multilingual Models to Machine Translation
Asa Cooper Stickland
Xian Li
Marjan Ghazvininejad
105
46
0
30 Apr 2020
UDapter: Language Adaptation for Truly Universal Dependency Parsing
UDapter: Language Adaptation for Truly Universal Dependency Parsing
Ahmet Üstün
Arianna Bisazza
G. Bouma
Gertjan van Noord
76
117
0
29 Apr 2020
General Purpose Text Embeddings from Pre-trained Language Models for
  Scalable Inference
General Purpose Text Embeddings from Pre-trained Language Models for Scalable Inference
Jingfei Du
Myle Ott
Haoran Li
Xing Zhou
Veselin Stoyanov
AI4CE
68
10
0
29 Apr 2020
Masking as an Efficient Alternative to Finetuning for Pretrained
  Language Models
Masking as an Efficient Alternative to Finetuning for Pretrained Language Models
Mengjie Zhao
Tao R. Lin
Fei Mi
Martin Jaggi
Hinrich Schütze
77
121
0
26 Apr 2020
How fine can fine-tuning be? Learning efficient language models
How fine can fine-tuning be? Learning efficient language models
Evani Radiya-Dixit
Xin Wang
53
66
0
24 Apr 2020
Fine-tuning Multi-hop Question Answering with Hierarchical Graph Network
Guanming Xiong
134
0
0
20 Apr 2020
Beyond Fine-tuning: Few-Sample Sentence Embedding Transfer
Beyond Fine-tuning: Few-Sample Sentence Embedding Transfer
Siddhant Garg
Rohit Kumar Sharma
Yingyu Liang
82
4
0
10 Apr 2020
Exploring Versatile Generative Language Model Via Parameter-Efficient
  Transfer Learning
Exploring Versatile Generative Language Model Via Parameter-Efficient Transfer Learning
Zhaojiang Lin
Andrea Madotto
Pascale Fung
121
164
0
08 Apr 2020
Meta-Learning for Few-Shot NMT Adaptation
Meta-Learning for Few-Shot NMT Adaptation
Amr Sharaf
Hany Hassan
Hal Daumé
66
35
0
06 Apr 2020
A Survey of Deep Learning for Scientific Discovery
A Survey of Deep Learning for Scientific Discovery
M. Raghu
Erica Schmidt
OODAI4CE
182
123
0
26 Mar 2020
Pre-trained Models for Natural Language Processing: A Survey
Pre-trained Models for Natural Language Processing: A Survey
Xipeng Qiu
Tianxiang Sun
Yige Xu
Yunfan Shao
Ning Dai
Xuanjing Huang
LM&MAVLM
453
1,500
0
18 Mar 2020
A Primer in BERTology: What we know about how BERT works
A Primer in BERTology: What we know about how BERT works
Anna Rogers
Olga Kovaleva
Anna Rumshisky
OffRL
170
1,511
0
27 Feb 2020
Exploring BERT Parameter Efficiency on the Stanford Question Answering
  Dataset v2.0
Exploring BERT Parameter Efficiency on the Stanford Question Answering Dataset v2.0
Eric Hulburd
53
5
0
25 Feb 2020
Improving BERT Fine-Tuning via Self-Ensemble and Self-Distillation
Improving BERT Fine-Tuning via Self-Ensemble and Self-Distillation
Yige Xu
Xipeng Qiu
L. Zhou
Xuanjing Huang
83
68
0
24 Feb 2020
Contextual Lensing of Universal Sentence Representations
Contextual Lensing of Universal Sentence Representations
J. Kiros
57
5
0
20 Feb 2020
K-Adapter: Infusing Knowledge into Pre-Trained Models with Adapters
K-Adapter: Infusing Knowledge into Pre-Trained Models with Adapters
Ruize Wang
Duyu Tang
Nan Duan
Zhongyu Wei
Xuanjing Huang
Jianshu Ji
Guihong Cao
Daxin Jiang
Ming Zhou
KELM
177
557
0
05 Feb 2020
Parameter-Efficient Transfer from Sequential Behaviors for User Modeling
  and Recommendation
Parameter-Efficient Transfer from Sequential Behaviors for User Modeling and Recommendation
Fajie Yuan
Xiangnan He
Alexandros Karatzoglou
Liguang Zhang
59
0
0
13 Jan 2020
Neural Network Surgery with Sets
Neural Network Surgery with Sets
Jonathan Raiman
Susan Zhang
Christy Dennison
58
4
0
13 Dec 2019
Pre-train and Plug-in: Flexible Conditional Text Generation with
  Variational Auto-Encoders
Pre-train and Plug-in: Flexible Conditional Text Generation with Variational Auto-Encoders
Yu Duan
Canwen Xu
Jiaxin Pei
Jialong Han
Chenliang Li
150
43
0
10 Nov 2019
Learning to Few-Shot Learn Across Diverse Natural Language
  Classification Tasks
Learning to Few-Shot Learn Across Diverse Natural Language Classification Tasks
Trapit Bansal
Rishikesh Jha
Andrew McCallum
SSL
118
121
0
10 Nov 2019
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language
  Models through Principled Regularized Optimization
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization
Haoming Jiang
Pengcheng He
Weizhu Chen
Xiaodong Liu
Jianfeng Gao
T. Zhao
143
563
0
08 Nov 2019
What Would Elsa Do? Freezing Layers During Transformer Fine-Tuning
What Would Elsa Do? Freezing Layers During Transformer Fine-Tuning
Jaejun Lee
Raphael Tang
Jimmy J. Lin
69
127
0
08 Nov 2019
Towards Domain Adaptation from Limited Data for Question Answering Using
  Deep Neural Networks
Towards Domain Adaptation from Limited Data for Question Answering Using Deep Neural Networks
Timothy J. Hazen
Shehzaad Dhuliawala
Daniel Boies
OOD
68
19
0
06 Nov 2019
On the Cross-lingual Transferability of Monolingual Representations
On the Cross-lingual Transferability of Monolingual Representations
Mikel Artetxe
Sebastian Ruder
Dani Yogatama
289
801
0
25 Oct 2019
Exploring the Limits of Transfer Learning with a Unified Text-to-Text
  Transformer
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
Colin Raffel
Noam M. Shazeer
Adam Roberts
Katherine Lee
Sharan Narang
Michael Matena
Yanqi Zhou
Wei Li
Peter J. Liu
AIMat
1.0K
20,462
0
23 Oct 2019
You May Not Need Order in Time Series Forecasting
You May Not Need Order in Time Series Forecasting
Yunkai Zhang
Qiao Jiang
Shurui Li
Xiaoyong Jin
Xueying Ma
Xifeng Yan
AI4TS
28
3
0
21 Oct 2019
SesameBERT: Attention for Anywhere
SesameBERT: Attention for Anywhere
Ta-Chun Su
Hsiang-Chih Cheng
66
7
0
08 Oct 2019
Previous
123...565758
Next