ResearchTrend.AI
  • Papers
  • Communities
  • Organizations
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1902.00751
  4. Cited By
Parameter-Efficient Transfer Learning for NLP
v1v2 (latest)

Parameter-Efficient Transfer Learning for NLP

2 February 2019
N. Houlsby
A. Giurgiu
Stanislaw Jastrzebski
Bruna Morrone
Quentin de Laroussilhe
Andrea Gesmundo
Mona Attariyan
Sylvain Gelly
ArXiv (abs)PDFHTML

Papers citing "Parameter-Efficient Transfer Learning for NLP"

50 / 2,860 papers shown
Title
Residual Adapters for Few-Shot Text-to-Speech Speaker Adaptation
Residual Adapters for Few-Shot Text-to-Speech Speaker Adaptation
Nobuyuki Morioka
Heiga Zen
Nanxin Chen
Yu Zhang
Yifan Ding
103
16
0
28 Oct 2022
Weight Averaging: A Simple Yet Effective Method to Overcome Catastrophic
  Forgetting in Automatic Speech Recognition
Weight Averaging: A Simple Yet Effective Method to Overcome Catastrophic Forgetting in Automatic Speech Recognition
Steven Vander Eeckt
Hugo Van hamme
CLLMoMe
126
15
0
27 Oct 2022
Robust Domain Adaptation for Pre-trained Multilingual Neural Machine
  Translation Models
Robust Domain Adaptation for Pre-trained Multilingual Neural Machine Translation Models
Mathieu Grosso
Pirashanth Ratnamogan
Alexis Mathey
William Vanhuffel
Michael Fotso Fotso
58
3
0
26 Oct 2022
Beyond English-Centric Bitexts for Better Multilingual Language
  Representation Learning
Beyond English-Centric Bitexts for Better Multilingual Language Representation Learning
Barun Patra
Saksham Singhal
Shaohan Huang
Zewen Chi
Li Dong
Furu Wei
Vishrav Chaudhary
Xia Song
127
24
0
26 Oct 2022
Inducer-tuning: Connecting Prefix-tuning and Adapter-tuning
Inducer-tuning: Connecting Prefix-tuning and Adapter-tuning
Yifan Chen
Devamanyu Hazarika
Mahdi Namazifar
Yang Liu
Di Jin
Dilek Z. Hakkani-Tür
73
4
0
26 Oct 2022
Learning Better Intent Representations for Financial Open Intent
  Classification
Learning Better Intent Representations for Financial Open Intent Classification
Xianzhi Li
Will Aitken
Xiao-Dan Zhu
Stephen W. Thomas
AIFin
72
8
0
25 Oct 2022
Leveraging Open Data and Task Augmentation to Automated Behavioral
  Coding of Psychotherapy Conversations in Low-Resource Scenarios
Leveraging Open Data and Task Augmentation to Automated Behavioral Coding of Psychotherapy Conversations in Low-Resource Scenarios
Zhuohao Chen
Nikolaos Flemotomos
Zac E. Imel
David C. Atkins
Shrikanth Narayanan
72
4
0
25 Oct 2022
Exploring Mode Connectivity for Pre-trained Language Models
Exploring Mode Connectivity for Pre-trained Language Models
Yujia Qin
Cheng Qian
Jing Yi
Weize Chen
Yankai Lin
Xu Han
Zhiyuan Liu
Maosong Sun
Jie Zhou
99
21
0
25 Oct 2022
PALT: Parameter-Lite Transfer of Language Models for Knowledge Graph
  Completion
PALT: Parameter-Lite Transfer of Language Models for Knowledge Graph Completion
Jianhao Shen
Chenguang Wang
Ye Yuan
Jiawei Han
Heng Ji
Koushik Sen
Ming Zhang
Dawn Song
KELMALMVPVLM
95
8
0
25 Oct 2022
Parameter-Efficient Legal Domain Adaptation
Parameter-Efficient Legal Domain Adaptation
Jonathan Li
R. Bhambhoria
Xiao-Dan Zhu
ELMAILawALM
106
14
0
25 Oct 2022
Evaluating Parameter Efficient Learning for Generation
Evaluating Parameter Efficient Learning for Generation
Peng Xu
M. Patwary
Shrimai Prabhumoye
Virginia Adams
R. Prenger
Ming-Yu Liu
Nayeon Lee
Mohammad Shoeybi
Bryan Catanzaro
MoE
72
3
0
25 Oct 2022
Adapters for Enhanced Modeling of Multilingual Knowledge and Text
Adapters for Enhanced Modeling of Multilingual Knowledge and Text
Buse Giledereli
Wenxiang Jiao
Mei-Jun Liu
Carl Allen
Zhaopeng Tu
Mrinmaya Sachan
99
11
0
24 Oct 2022
Different Tunes Played with Equal Skill: Exploring a Unified
  Optimization Subspace for Delta Tuning
Different Tunes Played with Equal Skill: Exploring a Unified Optimization Subspace for Delta Tuning
Jing Yi
Weize Chen
Yujia Qin
Yankai Lin
Ning Ding
Xu Han
Zhiyuan Liu
Maosong Sun
Jie Zhou
118
2
0
24 Oct 2022
NVIDIA FLARE: Federated Learning from Simulation to Real-World
NVIDIA FLARE: Federated Learning from Simulation to Real-World
H. Roth
Yan Cheng
Yuhong Wen
Isaac Yang
Ziyue Xu
...
Daguang Xu
Nic Ma
Prerna Dogra
Mona G. Flores
Andrew Feng
FedMLAI4CE
139
101
0
24 Oct 2022
Specializing Multi-domain NMT via Penalizing Low Mutual Information
Specializing Multi-domain NMT via Penalizing Low Mutual Information
Jiyoung Lee
H. Kim
Hyun Chang Cho
Edward Choi
Cheonbok Park
72
4
0
24 Oct 2022
Unsupervised Non-transferable Text Classification
Unsupervised Non-transferable Text Classification
Guangtao Zeng
Wei Lu
98
6
0
23 Oct 2022
Model ensemble instead of prompt fusion: a sample-specific knowledge
  transfer method for few-shot prompt tuning
Model ensemble instead of prompt fusion: a sample-specific knowledge transfer method for few-shot prompt tuning
Xiangyu Peng
Chen Xing
Prafulla Kumar Choubey
Chien-Sheng Wu
Caiming Xiong
VLM
137
12
0
23 Oct 2022
Exploring The Landscape of Distributional Robustness for Question
  Answering Models
Exploring The Landscape of Distributional Robustness for Question Answering Models
Anas Awadalla
Mitchell Wortsman
Gabriel Ilharco
Sewon Min
Ian H. Magnusson
Hannaneh Hajishirzi
Ludwig Schmidt
ELMOODKELM
124
21
0
22 Oct 2022
PATS: Sensitivity-aware Noisy Learning for Pretrained Language Models
PATS: Sensitivity-aware Noisy Learning for Pretrained Language Models
Yupeng Zhang
Hongzhi Zhang
Sirui Wang
Wei Wu
Zhoujun Li
AAML
104
1
0
22 Oct 2022
Clip-Tuning: Towards Derivative-free Prompt Learning with a Mixture of
  Rewards
Clip-Tuning: Towards Derivative-free Prompt Learning with a Mixture of Rewards
Yekun Chai
Shuohuan Wang
Yu Sun
Hao Tian
Hua Wu
Haifeng Wang
VLM
96
17
0
21 Oct 2022
$m^4Adapter$: Multilingual Multi-Domain Adaptation for Machine
  Translation with a Meta-Adapter
m4Adapterm^4Adapterm4Adapter: Multilingual Multi-Domain Adaptation for Machine Translation with a Meta-Adapter
Wen Lai
Alexandra Chronopoulou
Alexander Fraser
87
3
0
21 Oct 2022
Efficiently Tuned Parameters are Task Embeddings
Efficiently Tuned Parameters are Task Embeddings
Wangchunshu Zhou
Canwen Xu
Julian McAuley
58
8
0
21 Oct 2022
Late Prompt Tuning: A Late Prompt Could Be Better Than Many Prompts
Late Prompt Tuning: A Late Prompt Could Be Better Than Many Prompts
Xiangyang Liu
Tianxiang Sun
Xuanjing Huang
Xipeng Qiu
VLM
105
29
0
20 Oct 2022
Disentangling Reasoning Capabilities from Language Models with
  Compositional Reasoning Transformers
Disentangling Reasoning Capabilities from Language Models with Compositional Reasoning Transformers
Wanjun Zhong
Tingting Ma
Jiahai Wang
Jian Yin
Tiejun Zhao
Chin-Yew Lin
Nan Duan
LRMCoGe
85
2
0
20 Oct 2022
Towards Sustainable Self-supervised Learning
Towards Sustainable Self-supervised Learning
Shanghua Gao
Pan Zhou
Mingg-Ming Cheng
Shuicheng Yan
CLL
137
7
0
20 Oct 2022
Prompting through Prototype: A Prototype-based Prompt Learning on
  Pretrained Vision-Language Models
Prompting through Prototype: A Prototype-based Prompt Learning on Pretrained Vision-Language Models
Yue Zhang
Hongliang Fei
Dingcheng Li
Tan Yu
Ping Li
VPVLMVLM
71
9
0
19 Oct 2022
Incorporating Relevance Feedback for Information-Seeking Retrieval using
  Few-Shot Document Re-Ranking
Incorporating Relevance Feedback for Information-Seeking Retrieval using Few-Shot Document Re-Ranking
Tim Baumgärtner
Leonardo F. R. Ribeiro
Nils Reimers
Iryna Gurevych
86
6
0
19 Oct 2022
Revision Transformers: Instructing Language Models to Change their
  Values
Revision Transformers: Instructing Language Models to Change their Values
Felix Friedrich
Wolfgang Stammer
P. Schramowski
Kristian Kersting
KELM
77
8
0
19 Oct 2022
Continued Pretraining for Better Zero- and Few-Shot Promptability
Continued Pretraining for Better Zero- and Few-Shot Promptability
Zhaofeng Wu
IV RobertL.Logan
Pete Walsh
Akshita Bhagia
Dirk Groeneveld
Sameer Singh
Iz Beltagy
VLM
108
12
0
19 Oct 2022
Exclusive Supermask Subnetwork Training for Continual Learning
Exclusive Supermask Subnetwork Training for Continual Learning
Prateek Yadav
Joey Tianyi Zhou
CLL
100
6
0
18 Oct 2022
Hidden State Variability of Pretrained Language Models Can Guide
  Computation Reduction for Transfer Learning
Hidden State Variability of Pretrained Language Models Can Guide Computation Reduction for Transfer Learning
Shuo Xie
Jiahao Qiu
Ankita Pasad
Li Du
Qing Qu
Hongyuan Mei
92
16
0
18 Oct 2022
Tiny-Attention Adapter: Contexts Are More Important Than the Number of
  Parameters
Tiny-Attention Adapter: Contexts Are More Important Than the Number of Parameters
Hongyu Zhao
Hao Tan
Hongyuan Mei
MoE
87
18
0
18 Oct 2022
Using Bottleneck Adapters to Identify Cancer in Clinical Notes under
  Low-Resource Constraints
Using Bottleneck Adapters to Identify Cancer in Clinical Notes under Low-Resource Constraints
Omid Rohanian
Hannah Jauncey
Mohammadmahdi Nouriborji
Vinod Kumar Chauhan
Bronner P. Gonccalves
Christiana Kartsonaki
Isaric Clinical Characterisation Group
L. Merson
David Clifton
77
7
0
17 Oct 2022
Scaling & Shifting Your Features: A New Baseline for Efficient Model
  Tuning
Scaling & Shifting Your Features: A New Baseline for Efficient Model Tuning
Dongze Lian
Daquan Zhou
Jiashi Feng
Xinchao Wang
124
265
0
17 Oct 2022
SUPERB @ SLT 2022: Challenge on Generalization and Efficiency of
  Self-Supervised Speech Representation Learning
SUPERB @ SLT 2022: Challenge on Generalization and Efficiency of Self-Supervised Speech Representation Learning
Tzu-hsun Feng
Annie Dong
Ching-Feng Yeh
Shu-Wen Yang
Tzu-Quan Lin
...
Xuankai Chang
Shinji Watanabe
Abdel-rahman Mohamed
Shang-Wen Li
Hung-yi Lee
ELMSSL
102
35
0
16 Oct 2022
RedApt: An Adaptor for wav2vec 2 Encoding \\ Faster and Smaller Speech
  Translation without Quality Compromise
RedApt: An Adaptor for wav2vec 2 Encoding \\ Faster and Smaller Speech Translation without Quality Compromise
Jinming Zhao
Haomiao Yang
Gholamreza Haffari
Ehsan Shareghi
VLM
70
2
0
16 Oct 2022
Prompt Conditioned VAE: Enhancing Generative Replay for Lifelong
  Learning in Task-Oriented Dialogue
Prompt Conditioned VAE: Enhancing Generative Replay for Lifelong Learning in Task-Oriented Dialogue
Ying Zhao
Yinhe Zheng
Zhiliang Tian
Chang Gao
Yu Bowen
Haiyang Yu
Yongbin Li
Jianguo Sun
N. Zhang
CLLOffRL
74
13
0
14 Oct 2022
BERTScore is Unfair: On Social Bias in Language Model-Based Metrics for
  Text Generation
BERTScore is Unfair: On Social Bias in Language Model-Based Metrics for Text Generation
Tianxiang Sun
Junliang He
Xipeng Qiu
Xuanjing Huang
94
47
0
14 Oct 2022
Multitask Pre-training of Modular Prompt for Chinese Few-Shot Learning
Multitask Pre-training of Modular Prompt for Chinese Few-Shot Learning
Tianxiang Sun
Zhengfu He
Qinen Zhu
Xipeng Qiu
Xuanjing Huang
VLMVPVLM
53
21
0
14 Oct 2022
DyLoRA: Parameter Efficient Tuning of Pre-trained Models using Dynamic
  Search-Free Low-Rank Adaptation
DyLoRA: Parameter Efficient Tuning of Pre-trained Models using Dynamic Search-Free Low-Rank Adaptation
Mojtaba Valipour
Mehdi Rezagholizadeh
I. Kobyzev
A. Ghodsi
173
185
0
14 Oct 2022
M2D2: A Massively Multi-domain Language Modeling Dataset
M2D2: A Massively Multi-domain Language Modeling Dataset
Machel Reid
Victor Zhong
Suchin Gururangan
Luke Zettlemoyer
64
22
0
13 Oct 2022
Predicting Fine-Tuning Performance with Probing
Predicting Fine-Tuning Performance with Probing
Zining Zhu
Soroosh Shahtalebi
Frank Rudzicz
72
10
0
13 Oct 2022
Feature-Proxy Transformer for Few-Shot Segmentation
Feature-Proxy Transformer for Few-Shot Segmentation
Jianwei Zhang
Yifan Sun
Yi Yang
Wei Chen
ViT
84
63
0
13 Oct 2022
Are Sample-Efficient NLP Models More Robust?
Are Sample-Efficient NLP Models More Robust?
Nelson F. Liu
Ananya Kumar
Percy Liang
Robin Jia
VLMOOD
76
6
0
12 Oct 2022
Can Pretrained Language Models (Yet) Reason Deductively?
Can Pretrained Language Models (Yet) Reason Deductively?
Moy Yuan
Songbo Hu
Ivan Vulić
Anna Korhonen
Zaiqiao Meng
ReLMELMLRM
83
8
0
12 Oct 2022
MiniALBERT: Model Distillation via Parameter-Efficient Recursive
  Transformers
MiniALBERT: Model Distillation via Parameter-Efficient Recursive Transformers
Mohammadmahdi Nouriborji
Omid Rohanian
Samaneh Kouchaki
David Clifton
92
8
0
12 Oct 2022
Back to the Future: On Potential Histories in NLP
Back to the Future: On Potential Histories in NLP
Zeerak Talat
Anne Lauscher
AI4TS
78
4
0
12 Oct 2022
Continual Training of Language Models for Few-Shot Learning
Continual Training of Language Models for Few-Shot Learning
Zixuan Ke
Haowei Lin
Yijia Shao
Hu Xu
Lei Shu
Bin Liu
KELMBDLCLL
145
36
0
11 Oct 2022
Model Cascading: Towards Jointly Improving Efficiency and Accuracy of
  NLP Systems
Model Cascading: Towards Jointly Improving Efficiency and Accuracy of NLP Systems
Neeraj Varshney
Chitta Baral
92
28
0
11 Oct 2022
Multi-CLS BERT: An Efficient Alternative to Traditional Ensembling
Multi-CLS BERT: An Efficient Alternative to Traditional Ensembling
Haw-Shiuan Chang
Ruei-Yao Sun
Kathryn Ricci
Andrew McCallum
112
15
0
10 Oct 2022
Previous
123...464748...565758
Next