Papers
Communities
Organizations
Events
Blog
Pricing
Search
Open menu
Home
Papers
1902.00751
Cited By
v1
v2 (latest)
Parameter-Efficient Transfer Learning for NLP
2 February 2019
N. Houlsby
A. Giurgiu
Stanislaw Jastrzebski
Bruna Morrone
Quentin de Laroussilhe
Andrea Gesmundo
Mona Attariyan
Sylvain Gelly
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Parameter-Efficient Transfer Learning for NLP"
50 / 2,860 papers shown
Title
Residual Adapters for Few-Shot Text-to-Speech Speaker Adaptation
Nobuyuki Morioka
Heiga Zen
Nanxin Chen
Yu Zhang
Yifan Ding
103
16
0
28 Oct 2022
Weight Averaging: A Simple Yet Effective Method to Overcome Catastrophic Forgetting in Automatic Speech Recognition
Steven Vander Eeckt
Hugo Van hamme
CLL
MoMe
126
15
0
27 Oct 2022
Robust Domain Adaptation for Pre-trained Multilingual Neural Machine Translation Models
Mathieu Grosso
Pirashanth Ratnamogan
Alexis Mathey
William Vanhuffel
Michael Fotso Fotso
58
3
0
26 Oct 2022
Beyond English-Centric Bitexts for Better Multilingual Language Representation Learning
Barun Patra
Saksham Singhal
Shaohan Huang
Zewen Chi
Li Dong
Furu Wei
Vishrav Chaudhary
Xia Song
127
24
0
26 Oct 2022
Inducer-tuning: Connecting Prefix-tuning and Adapter-tuning
Yifan Chen
Devamanyu Hazarika
Mahdi Namazifar
Yang Liu
Di Jin
Dilek Z. Hakkani-Tür
73
4
0
26 Oct 2022
Learning Better Intent Representations for Financial Open Intent Classification
Xianzhi Li
Will Aitken
Xiao-Dan Zhu
Stephen W. Thomas
AIFin
72
8
0
25 Oct 2022
Leveraging Open Data and Task Augmentation to Automated Behavioral Coding of Psychotherapy Conversations in Low-Resource Scenarios
Zhuohao Chen
Nikolaos Flemotomos
Zac E. Imel
David C. Atkins
Shrikanth Narayanan
72
4
0
25 Oct 2022
Exploring Mode Connectivity for Pre-trained Language Models
Yujia Qin
Cheng Qian
Jing Yi
Weize Chen
Yankai Lin
Xu Han
Zhiyuan Liu
Maosong Sun
Jie Zhou
99
21
0
25 Oct 2022
PALT: Parameter-Lite Transfer of Language Models for Knowledge Graph Completion
Jianhao Shen
Chenguang Wang
Ye Yuan
Jiawei Han
Heng Ji
Koushik Sen
Ming Zhang
Dawn Song
KELM
ALM
VPVLM
95
8
0
25 Oct 2022
Parameter-Efficient Legal Domain Adaptation
Jonathan Li
R. Bhambhoria
Xiao-Dan Zhu
ELM
AILaw
ALM
106
14
0
25 Oct 2022
Evaluating Parameter Efficient Learning for Generation
Peng Xu
M. Patwary
Shrimai Prabhumoye
Virginia Adams
R. Prenger
Ming-Yu Liu
Nayeon Lee
Mohammad Shoeybi
Bryan Catanzaro
MoE
72
3
0
25 Oct 2022
Adapters for Enhanced Modeling of Multilingual Knowledge and Text
Buse Giledereli
Wenxiang Jiao
Mei-Jun Liu
Carl Allen
Zhaopeng Tu
Mrinmaya Sachan
99
11
0
24 Oct 2022
Different Tunes Played with Equal Skill: Exploring a Unified Optimization Subspace for Delta Tuning
Jing Yi
Weize Chen
Yujia Qin
Yankai Lin
Ning Ding
Xu Han
Zhiyuan Liu
Maosong Sun
Jie Zhou
118
2
0
24 Oct 2022
NVIDIA FLARE: Federated Learning from Simulation to Real-World
H. Roth
Yan Cheng
Yuhong Wen
Isaac Yang
Ziyue Xu
...
Daguang Xu
Nic Ma
Prerna Dogra
Mona G. Flores
Andrew Feng
FedML
AI4CE
139
101
0
24 Oct 2022
Specializing Multi-domain NMT via Penalizing Low Mutual Information
Jiyoung Lee
H. Kim
Hyun Chang Cho
Edward Choi
Cheonbok Park
72
4
0
24 Oct 2022
Unsupervised Non-transferable Text Classification
Guangtao Zeng
Wei Lu
98
6
0
23 Oct 2022
Model ensemble instead of prompt fusion: a sample-specific knowledge transfer method for few-shot prompt tuning
Xiangyu Peng
Chen Xing
Prafulla Kumar Choubey
Chien-Sheng Wu
Caiming Xiong
VLM
137
12
0
23 Oct 2022
Exploring The Landscape of Distributional Robustness for Question Answering Models
Anas Awadalla
Mitchell Wortsman
Gabriel Ilharco
Sewon Min
Ian H. Magnusson
Hannaneh Hajishirzi
Ludwig Schmidt
ELM
OOD
KELM
124
21
0
22 Oct 2022
PATS: Sensitivity-aware Noisy Learning for Pretrained Language Models
Yupeng Zhang
Hongzhi Zhang
Sirui Wang
Wei Wu
Zhoujun Li
AAML
104
1
0
22 Oct 2022
Clip-Tuning: Towards Derivative-free Prompt Learning with a Mixture of Rewards
Yekun Chai
Shuohuan Wang
Yu Sun
Hao Tian
Hua Wu
Haifeng Wang
VLM
96
17
0
21 Oct 2022
m
4
A
d
a
p
t
e
r
m^4Adapter
m
4
A
d
a
pt
er
: Multilingual Multi-Domain Adaptation for Machine Translation with a Meta-Adapter
Wen Lai
Alexandra Chronopoulou
Alexander Fraser
87
3
0
21 Oct 2022
Efficiently Tuned Parameters are Task Embeddings
Wangchunshu Zhou
Canwen Xu
Julian McAuley
58
8
0
21 Oct 2022
Late Prompt Tuning: A Late Prompt Could Be Better Than Many Prompts
Xiangyang Liu
Tianxiang Sun
Xuanjing Huang
Xipeng Qiu
VLM
105
29
0
20 Oct 2022
Disentangling Reasoning Capabilities from Language Models with Compositional Reasoning Transformers
Wanjun Zhong
Tingting Ma
Jiahai Wang
Jian Yin
Tiejun Zhao
Chin-Yew Lin
Nan Duan
LRM
CoGe
85
2
0
20 Oct 2022
Towards Sustainable Self-supervised Learning
Shanghua Gao
Pan Zhou
Mingg-Ming Cheng
Shuicheng Yan
CLL
137
7
0
20 Oct 2022
Prompting through Prototype: A Prototype-based Prompt Learning on Pretrained Vision-Language Models
Yue Zhang
Hongliang Fei
Dingcheng Li
Tan Yu
Ping Li
VPVLM
VLM
71
9
0
19 Oct 2022
Incorporating Relevance Feedback for Information-Seeking Retrieval using Few-Shot Document Re-Ranking
Tim Baumgärtner
Leonardo F. R. Ribeiro
Nils Reimers
Iryna Gurevych
86
6
0
19 Oct 2022
Revision Transformers: Instructing Language Models to Change their Values
Felix Friedrich
Wolfgang Stammer
P. Schramowski
Kristian Kersting
KELM
77
8
0
19 Oct 2022
Continued Pretraining for Better Zero- and Few-Shot Promptability
Zhaofeng Wu
IV RobertL.Logan
Pete Walsh
Akshita Bhagia
Dirk Groeneveld
Sameer Singh
Iz Beltagy
VLM
108
12
0
19 Oct 2022
Exclusive Supermask Subnetwork Training for Continual Learning
Prateek Yadav
Joey Tianyi Zhou
CLL
100
6
0
18 Oct 2022
Hidden State Variability of Pretrained Language Models Can Guide Computation Reduction for Transfer Learning
Shuo Xie
Jiahao Qiu
Ankita Pasad
Li Du
Qing Qu
Hongyuan Mei
92
16
0
18 Oct 2022
Tiny-Attention Adapter: Contexts Are More Important Than the Number of Parameters
Hongyu Zhao
Hao Tan
Hongyuan Mei
MoE
87
18
0
18 Oct 2022
Using Bottleneck Adapters to Identify Cancer in Clinical Notes under Low-Resource Constraints
Omid Rohanian
Hannah Jauncey
Mohammadmahdi Nouriborji
Vinod Kumar Chauhan
Bronner P. Gonccalves
Christiana Kartsonaki
Isaric Clinical Characterisation Group
L. Merson
David Clifton
77
7
0
17 Oct 2022
Scaling & Shifting Your Features: A New Baseline for Efficient Model Tuning
Dongze Lian
Daquan Zhou
Jiashi Feng
Xinchao Wang
124
265
0
17 Oct 2022
SUPERB @ SLT 2022: Challenge on Generalization and Efficiency of Self-Supervised Speech Representation Learning
Tzu-hsun Feng
Annie Dong
Ching-Feng Yeh
Shu-Wen Yang
Tzu-Quan Lin
...
Xuankai Chang
Shinji Watanabe
Abdel-rahman Mohamed
Shang-Wen Li
Hung-yi Lee
ELM
SSL
102
35
0
16 Oct 2022
RedApt: An Adaptor for wav2vec 2 Encoding \\ Faster and Smaller Speech Translation without Quality Compromise
Jinming Zhao
Haomiao Yang
Gholamreza Haffari
Ehsan Shareghi
VLM
70
2
0
16 Oct 2022
Prompt Conditioned VAE: Enhancing Generative Replay for Lifelong Learning in Task-Oriented Dialogue
Ying Zhao
Yinhe Zheng
Zhiliang Tian
Chang Gao
Yu Bowen
Haiyang Yu
Yongbin Li
Jianguo Sun
N. Zhang
CLL
OffRL
74
13
0
14 Oct 2022
BERTScore is Unfair: On Social Bias in Language Model-Based Metrics for Text Generation
Tianxiang Sun
Junliang He
Xipeng Qiu
Xuanjing Huang
94
47
0
14 Oct 2022
Multitask Pre-training of Modular Prompt for Chinese Few-Shot Learning
Tianxiang Sun
Zhengfu He
Qinen Zhu
Xipeng Qiu
Xuanjing Huang
VLM
VPVLM
53
21
0
14 Oct 2022
DyLoRA: Parameter Efficient Tuning of Pre-trained Models using Dynamic Search-Free Low-Rank Adaptation
Mojtaba Valipour
Mehdi Rezagholizadeh
I. Kobyzev
A. Ghodsi
173
185
0
14 Oct 2022
M2D2: A Massively Multi-domain Language Modeling Dataset
Machel Reid
Victor Zhong
Suchin Gururangan
Luke Zettlemoyer
64
22
0
13 Oct 2022
Predicting Fine-Tuning Performance with Probing
Zining Zhu
Soroosh Shahtalebi
Frank Rudzicz
72
10
0
13 Oct 2022
Feature-Proxy Transformer for Few-Shot Segmentation
Jianwei Zhang
Yifan Sun
Yi Yang
Wei Chen
ViT
84
63
0
13 Oct 2022
Are Sample-Efficient NLP Models More Robust?
Nelson F. Liu
Ananya Kumar
Percy Liang
Robin Jia
VLM
OOD
76
6
0
12 Oct 2022
Can Pretrained Language Models (Yet) Reason Deductively?
Moy Yuan
Songbo Hu
Ivan Vulić
Anna Korhonen
Zaiqiao Meng
ReLM
ELM
LRM
83
8
0
12 Oct 2022
MiniALBERT: Model Distillation via Parameter-Efficient Recursive Transformers
Mohammadmahdi Nouriborji
Omid Rohanian
Samaneh Kouchaki
David Clifton
92
8
0
12 Oct 2022
Back to the Future: On Potential Histories in NLP
Zeerak Talat
Anne Lauscher
AI4TS
78
4
0
12 Oct 2022
Continual Training of Language Models for Few-Shot Learning
Zixuan Ke
Haowei Lin
Yijia Shao
Hu Xu
Lei Shu
Bin Liu
KELM
BDL
CLL
145
36
0
11 Oct 2022
Model Cascading: Towards Jointly Improving Efficiency and Accuracy of NLP Systems
Neeraj Varshney
Chitta Baral
92
28
0
11 Oct 2022
Multi-CLS BERT: An Efficient Alternative to Traditional Ensembling
Haw-Shiuan Chang
Ruei-Yao Sun
Kathryn Ricci
Andrew McCallum
112
15
0
10 Oct 2022
Previous
1
2
3
...
46
47
48
...
56
57
58
Next