ResearchTrend.AI
  • Papers
  • Communities
  • Organizations
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1902.00751
  4. Cited By
Parameter-Efficient Transfer Learning for NLP
v1v2 (latest)

Parameter-Efficient Transfer Learning for NLP

2 February 2019
N. Houlsby
A. Giurgiu
Stanislaw Jastrzebski
Bruna Morrone
Quentin de Laroussilhe
Andrea Gesmundo
Mona Attariyan
Sylvain Gelly
ArXiv (abs)PDFHTML

Papers citing "Parameter-Efficient Transfer Learning for NLP"

50 / 2,860 papers shown
Title
Modulating Language Models with Emotions
Modulating Language Models with Emotions
Ruibo Liu
Jason W. Wei
Chenyan Jia
Soroush Vosoughi
65
23
0
17 Aug 2021
AMMUS : A Survey of Transformer-based Pretrained Models in Natural
  Language Processing
AMMUS : A Survey of Transformer-based Pretrained Models in Natural Language Processing
Katikapalli Subramanyam Kalyan
A. Rajasekharan
S. Sangeetha
VLMLM&MA
116
270
0
12 Aug 2021
DEMix Layers: Disentangling Domains for Modular Language Modeling
DEMix Layers: Disentangling Domains for Modular Language Modeling
Suchin Gururangan
Michael Lewis
Ari Holtzman
Noah A. Smith
Luke Zettlemoyer
KELMMoE
125
138
0
11 Aug 2021
Noisy Channel Language Model Prompting for Few-Shot Text Classification
Noisy Channel Language Model Prompting for Few-Shot Text Classification
Sewon Min
Michael Lewis
Hannaneh Hajishirzi
Luke Zettlemoyer
VLM
120
220
0
09 Aug 2021
Robust Transfer Learning with Pretrained Language Models through
  Adapters
Robust Transfer Learning with Pretrained Language Models through Adapters
Wenjuan Han
Bo Pang
Ying Nian Wu
74
56
0
05 Aug 2021
Domain Adaptor Networks for Hyperspectral Image Recognition
Domain Adaptor Networks for Hyperspectral Image Recognition
Gustavo Pérez
Subhransu Maji
31
0
0
03 Aug 2021
Exceeding the Limits of Visual-Linguistic Multi-Task Learning
Exceeding the Limits of Visual-Linguistic Multi-Task Learning
Cameron R. Wolfe
Keld T. Lundgaard
VLM
80
2
0
27 Jul 2021
Cross-lingual Transferring of Pre-trained Contextualized Language Models
Cross-lingual Transferring of Pre-trained Contextualized Language Models
Zuchao Li
Kevin Parnow
Hai Zhao
Zhuosheng Zhang
Rui Wang
Masao Utiyama
Eiichiro Sumita
57
8
0
27 Jul 2021
An Adapter Based Pre-Training for Efficient and Scalable Self-Supervised
  Speech Representation Learning
An Adapter Based Pre-Training for Efficient and Scalable Self-Supervised Speech Representation Learning
Samuel Kessler
Bethan Thomas
S. Karout
SSL
81
30
0
26 Jul 2021
ProtoTransformer: A Meta-Learning Approach to Providing Student Feedback
ProtoTransformer: A Meta-Learning Approach to Providing Student Feedback
Mike Wu
Noah D. Goodman
Chris Piech
Chelsea Finn
89
19
0
23 Jul 2021
A Flexible Multi-Task Model for BERT Serving
A Flexible Multi-Task Model for BERT Serving
Tianwen Wei
Jianwei Qi
Shenghuang He
48
7
0
12 Jul 2021
A Primer on Pretrained Multilingual Language Models
A Primer on Pretrained Multilingual Language Models
Sumanth Doddapaneni
Gowtham Ramesh
Mitesh M. Khapra
Anoop Kunchukuttan
Pratyush Kumar
LRM
123
76
0
01 Jul 2021
Scientia Potentia Est -- On the Role of Knowledge in Computational
  Argumentation
Scientia Potentia Est -- On the Role of Knowledge in Computational Argumentation
Anne Lauscher
Henning Wachsmuth
Iryna Gurevych
Goran Glavaš
104
34
0
01 Jul 2021
Cutting Down on Prompts and Parameters: Simple Few-Shot Learning with
  Language Models
Cutting Down on Prompts and Parameters: Simple Few-Shot Learning with Language Models
Robert L Logan IV
Ivana Balavzević
Eric Wallace
Fabio Petroni
Sameer Singh
Sebastian Riedel
VPVLM
113
213
0
24 Jun 2021
Learning Multiple Stock Trading Patterns with Temporal Routing Adaptor
  and Optimal Transport
Learning Multiple Stock Trading Patterns with Temporal Routing Adaptor and Optimal Transport
Hengxu Lin
Dong Zhou
Weiqing Liu
Jiang Bian
AIFinAI4TS
88
67
0
24 Jun 2021
Do Language Models Perform Generalizable Commonsense Inference?
Do Language Models Perform Generalizable Commonsense Inference?
Peifeng Wang
Filip Ilievski
Muhao Chen
Xiang Ren
ReLMLRM
53
19
0
22 Jun 2021
BitFit: Simple Parameter-efficient Fine-tuning for Transformer-based
  Masked Language-models
BitFit: Simple Parameter-efficient Fine-tuning for Transformer-based Masked Language-models
Elad Ben-Zaken
Shauli Ravfogel
Yoav Goldberg
316
1,250
0
18 Jun 2021
Adversarial Training Helps Transfer Learning via Better Representations
Adversarial Training Helps Transfer Learning via Better Representations
Zhun Deng
Linjun Zhang
Kailas Vodrahalli
Kenji Kawaguchi
James Zou
GAN
91
54
0
18 Jun 2021
LoRA: Low-Rank Adaptation of Large Language Models
LoRA: Low-Rank Adaptation of Large Language Models
J. E. Hu
Yelong Shen
Phillip Wallis
Zeyuan Allen-Zhu
Yuanzhi Li
Shean Wang
Lu Wang
Weizhu Chen
OffRLAI4TSAI4CEALMAIMat
890
10,661
0
17 Jun 2021
Specializing Multilingual Language Models: An Empirical Study
Specializing Multilingual Language Models: An Empirical Study
Ethan C. Chau
Noah A. Smith
139
27
0
16 Jun 2021
Neural Supervised Domain Adaptation by Augmenting Pre-trained Models
  with Random Units
Neural Supervised Domain Adaptation by Augmenting Pre-trained Models with Random Units
Sara Meftah
N. Semmar
Y. Tamaazousti
H. Essafi
F. Sadat
67
3
0
09 Jun 2021
Compacter: Efficient Low-Rank Hypercomplex Adapter Layers
Compacter: Efficient Low-Rank Hypercomplex Adapter Layers
Rabeeh Karimi Mahabadi
James Henderson
Sebastian Ruder
MoE
163
495
0
08 Jun 2021
Parameter-efficient Multi-task Fine-tuning for Transformers via Shared
  Hypernetworks
Parameter-efficient Multi-task Fine-tuning for Transformers via Shared Hypernetworks
Rabeeh Karimi Mahabadi
Sebastian Ruder
Mostafa Dehghani
James Henderson
MoE
86
314
0
08 Jun 2021
On the Effectiveness of Adapter-based Tuning for Pretrained Language
  Model Adaptation
On the Effectiveness of Adapter-based Tuning for Pretrained Language Model Adaptation
Ruidan He
Linlin Liu
Hai Ye
Qingyu Tan
Bosheng Ding
Liying Cheng
Jia-Wei Low
Lidong Bing
Luo Si
65
205
0
06 Jun 2021
Signal Transformer: Complex-valued Attention and Meta-Learning for
  Signal Recognition
Signal Transformer: Complex-valued Attention and Meta-Learning for Signal Recognition
Yihong Dong
Ying Peng
Muqiao Yang
Songtao Lu
Qingjiang Shi
105
9
0
05 Jun 2021
Language Scaling for Universal Suggested Replies Model
Language Scaling for Universal Suggested Replies Model
Qianlan Ying
Payal Bajaj
Budhaditya Deb
Yu Yang
Wei Wang
Bojia Lin
Milad Shokouhi
Xia Song
Yang Yang
Daxin Jiang
LRM
61
2
0
04 Jun 2021
Enabling Lightweight Fine-tuning for Pre-trained Language Model
  Compression based on Matrix Product Operators
Enabling Lightweight Fine-tuning for Pre-trained Language Model Compression based on Matrix Product Operators
Peiyu Liu
Ze-Feng Gao
Wayne Xin Zhao
Z. Xie
Zhong-Yi Lu
Ji-Rong Wen
48
30
0
04 Jun 2021
How to Adapt Your Pretrained Multilingual Model to 1600 Languages
How to Adapt Your Pretrained Multilingual Model to 1600 Languages
Abteen Ebrahimi
Katharina Kann
LRMVLM
100
70
0
03 Jun 2021
Lightweight Adapter Tuning for Multilingual Speech Translation
Lightweight Adapter Tuning for Multilingual Speech Translation
Hang Le
J. Pino
Changhan Wang
Jiatao Gu
D. Schwab
Laurent Besacier
147
90
0
02 Jun 2021
Crowdsourcing Learning as Domain Adaptation: A Case Study on Named
  Entity Recognition
Crowdsourcing Learning as Domain Adaptation: A Case Study on Named Entity Recognition
Xin Zhang
Guangwei Xu
Yueheng Sun
Meishan Zhang
Pengjun Xie
84
22
0
31 May 2021
Exploiting Adapters for Cross-lingual Low-resource Speech Recognition
Exploiting Adapters for Cross-lingual Low-resource Speech Recognition
Wenxin Hou
Hanlin Zhu
Yidong Wang
Jindong Wang
Tao Qin
Renjun Xu
T. Shinozaki
71
65
0
18 May 2021
Lexicon Enhanced Chinese Sequence Labeling Using BERT Adapter
Lexicon Enhanced Chinese Sequence Labeling Using BERT Adapter
Wei Liu
Xiyan Fu
Yueqian Zhang
Wenming Xiao
66
163
0
15 May 2021
Aggregating From Multiple Target-Shifted Sources
Aggregating From Multiple Target-Shifted Sources
Changjian Shui
Zijian Li
Jiaqi Li
Christian Gagné
Charles Ling
Boyu Wang
95
31
0
09 May 2021
Adapting by Pruning: A Case Study on BERT
Adapting by Pruning: A Case Study on BERT
Yang Gao
Nicolo Colombo
Wen Wang
64
17
0
07 May 2021
MineGAN++: Mining Generative Models for Efficient Knowledge Transfer to
  Limited Data Domains
MineGAN++: Mining Generative Models for Efficient Knowledge Transfer to Limited Data Domains
Yaxing Wang
Abel Gonzalez-Garcia
Chenshen Wu
Luis Herranz
Fahad Shahbaz Khan
Shangling Jui
Joost van de Weijer
68
6
0
28 Apr 2021
If your data distribution shifts, use self-learning
If your data distribution shifts, use self-learning
E. Rusak
Steffen Schneider
George Pachitariu
L. Eck
Peter V. Gehler
Oliver Bringmann
Wieland Brendel
Matthias Bethge
VLMOODTTA
177
33
0
27 Apr 2021
XLM-T: Multilingual Language Models in Twitter for Sentiment Analysis
  and Beyond
XLM-T: Multilingual Language Models in Twitter for Sentiment Analysis and Beyond
Francesco Barbieri
Luis Espinosa Anke
Jose Camacho-Collados
270
228
0
25 Apr 2021
Math Operation Embeddings for Open-ended Solution Analysis and Feedback
Math Operation Embeddings for Open-ended Solution Analysis and Feedback
Mengxue Zhang
Zichao Wang
Richard Baraniuk
Andrew Lan
65
10
0
25 Apr 2021
Revisiting Document Representations for Large-Scale Zero-Shot Learning
Revisiting Document Representations for Large-Scale Zero-Shot Learning
Jihyung Kil
Wei-Lun Chao
VLM
100
10
0
21 Apr 2021
Learn Continually, Generalize Rapidly: Lifelong Knowledge Accumulation
  for Few-shot Learning
Learn Continually, Generalize Rapidly: Lifelong Knowledge Accumulation for Few-shot Learning
Xisen Jin
Bill Yuchen Lin
Mohammad Rostami
Xiang Ren
BDLCLL
86
42
0
18 Apr 2021
Cross-Attention is All You Need: Adapting Pretrained Transformers for
  Machine Translation
Cross-Attention is All You Need: Adapting Pretrained Transformers for Machine Translation
Mozhdeh Gheini
Xiang Ren
Jonathan May
LRM
95
116
0
18 Apr 2021
The Power of Scale for Parameter-Efficient Prompt Tuning
The Power of Scale for Parameter-Efficient Prompt Tuning
Brian Lester
Rami Al-Rfou
Noah Constant
VPVLM
881
4,129
0
18 Apr 2021
Transductive Learning for Abstractive News Summarization
Transductive Learning for Abstractive News Summarization
Arthur Bravzinskas
Mengwen Liu
Ramesh Nallapati
Sujith Ravi
Markus Dreyer
AI4TS
36
1
0
17 Apr 2021
Neural Path Hunter: Reducing Hallucination in Dialogue Systems via Path
  Grounding
Neural Path Hunter: Reducing Hallucination in Dialogue Systems via Path Grounding
Nouha Dziri
Andrea Madotto
Osmar Zaiane
A. Bose
HILM
94
137
0
17 Apr 2021
Condenser: a Pre-training Architecture for Dense Retrieval
Condenser: a Pre-training Architecture for Dense Retrieval
Luyu Gao
Jamie Callan
AI4CE
71
269
0
16 Apr 2021
What to Pre-Train on? Efficient Intermediate Task Selection
What to Pre-Train on? Efficient Intermediate Task Selection
Clifton A. Poth
Jonas Pfeiffer
Andreas Rucklé
Iryna Gurevych
113
100
0
16 Apr 2021
Counter-Interference Adapter for Multilingual Machine Translation
Counter-Interference Adapter for Multilingual Machine Translation
Yaoming Zhu
Jiangtao Feng
Chengqi Zhao
Mingxuan Wang
Lei Li
66
63
0
16 Apr 2021
MetaXL: Meta Representation Transformation for Low-resource
  Cross-lingual Learning
MetaXL: Meta Representation Transformation for Low-resource Cross-lingual Learning
Mengzhou Xia
Guoqing Zheng
Subhabrata Mukherjee
Milad Shokouhi
Graham Neubig
Ahmed Hassan Awadallah
82
32
0
16 Apr 2021
Device-Cloud Collaborative Learning for Recommendation
Device-Cloud Collaborative Learning for Recommendation
Jiangchao Yao
Feng Wang
Kunyang Jia
Bo Han
Jingren Zhou
Hongxia Yang
104
63
0
14 Apr 2021
Relational World Knowledge Representation in Contextual Language Models:
  A Review
Relational World Knowledge Representation in Contextual Language Models: A Review
Tara Safavi
Danai Koutra
KELM
100
51
0
12 Apr 2021
Previous
123...5455565758
Next