ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2106.10199
  4. Cited By
BitFit: Simple Parameter-efficient Fine-tuning for Transformer-based
  Masked Language-models

BitFit: Simple Parameter-efficient Fine-tuning for Transformer-based Masked Language-models

18 June 2021
Elad Ben-Zaken
Shauli Ravfogel
Yoav Goldberg
ArXivPDFHTML

Papers citing "BitFit: Simple Parameter-efficient Fine-tuning for Transformer-based Masked Language-models"

50 / 821 papers shown
Title
Ladder Fine-tuning approach for SAM integrating complementary network
Ladder Fine-tuning approach for SAM integrating complementary network
Shurong Chai
R. Jain
Shiyu Teng
Jiaqing Liu
Yinhao Li
T. Tateyama
Yen-wei Chen
MedIm
30
31
0
22 Jun 2023
Parameter-efficient is not sufficient: Exploring Parameter, Memory, and
  Time Efficient Adapter Tuning for Dense Predictions
Parameter-efficient is not sufficient: Exploring Parameter, Memory, and Time Efficient Adapter Tuning for Dense Predictions
Dongshuo Yin
Xueting Han
Bin Li
Hao Feng
Jinghua Bai
VPVLM
36
18
0
16 Jun 2023
One-for-All: Generalized LoRA for Parameter-Efficient Fine-tuning
One-for-All: Generalized LoRA for Parameter-Efficient Fine-tuning
Arnav Chavan
Zhuang Liu
D. K. Gupta
Eric P. Xing
Zhiqiang Shen
36
87
0
13 Jun 2023
Morphosyntactic probing of multilingual BERT models
Morphosyntactic probing of multilingual BERT models
Judit Ács
Endre Hamerlik
Roy Schwartz
Noah A. Smith
András Kornai
35
9
0
09 Jun 2023
Git-Theta: A Git Extension for Collaborative Development of Machine
  Learning Models
Git-Theta: A Git Extension for Collaborative Development of Machine Learning Models
Nikhil Kandpal
Brian Lester
Mohammed Muqeeth
Anisha Mascarenhas
Monty Evans
Vishal Baskaran
Tenghao Huang
Haokun Liu
Colin Raffel
VLM
24
10
0
07 Jun 2023
ViDA: Homeostatic Visual Domain Adapter for Continual Test Time
  Adaptation
ViDA: Homeostatic Visual Domain Adapter for Continual Test Time Adaptation
Jiaming Liu
Senqiao Yang
Peidong Jia
Renrui Zhang
Ming Lu
Yandong Guo
Wei Xue
Shanghang Zhang
TTA
OOD
VLM
30
36
0
07 Jun 2023
Inference-Time Intervention: Eliciting Truthful Answers from a Language
  Model
Inference-Time Intervention: Eliciting Truthful Answers from a Language Model
Kenneth Li
Oam Patel
Fernanda Viégas
Hanspeter Pfister
Martin Wattenberg
KELM
HILM
43
485
0
06 Jun 2023
Sensitivity-Aware Finetuning for Accuracy Recovery on Deep Learning
  Hardware
Sensitivity-Aware Finetuning for Accuracy Recovery on Deep Learning Hardware
Lakshmi Nair
D. Bunandar
22
0
0
05 Jun 2023
Cross-Lingual Transfer with Target Language-Ready Task Adapters
Cross-Lingual Transfer with Target Language-Ready Task Adapters
Marinela Parović
Alan Ansell
Ivan Vulić
Anna Korhonen
40
9
0
05 Jun 2023
Prompt to be Consistent is Better than Self-Consistent? Few-Shot and
  Zero-Shot Fact Verification with Pre-trained Language Models
Prompt to be Consistent is Better than Self-Consistent? Few-Shot and Zero-Shot Fact Verification with Pre-trained Language Models
Fengzhu Zeng
Wei Gao
23
5
0
05 Jun 2023
bgGLUE: A Bulgarian General Language Understanding Evaluation Benchmark
bgGLUE: A Bulgarian General Language Understanding Evaluation Benchmark
Momchil Hardalov
Pepa Atanasova
Todor Mihaylov
G. Angelova
K. Simov
P. Osenova
Ves Stoyanov
Ivan Koychev
Preslav Nakov
Dragomir R. Radev
ELM
FedML
29
4
0
04 Jun 2023
Exploring the Impact of Model Scaling on Parameter-Efficient Tuning
Exploring the Impact of Model Scaling on Parameter-Efficient Tuning
Yusheng Su
Chi-Min Chan
Jiali Cheng
Yujia Qin
Yankai Lin
...
Ning Ding
Xingzhi Sun
Guotong Xie
Zhiyuan Liu
Maosong Sun
24
6
0
04 Jun 2023
Benchmarking Robustness of Adaptation Methods on Pre-trained
  Vision-Language Models
Benchmarking Robustness of Adaptation Methods on Pre-trained Vision-Language Models
Shuo Chen
Jindong Gu
Zhen Han
Yunpu Ma
Philip Torr
Volker Tresp
VPVLM
VLM
45
17
0
03 Jun 2023
UniDiff: Advancing Vision-Language Models with Generative and
  Discriminative Learning
UniDiff: Advancing Vision-Language Models with Generative and Discriminative Learning
Xiao Dong
Runhu Huang
Xiaoyong Wei
Zequn Jie
Jianxing Yu
Jian Yin
Xiaodan Liang
VLM
DiffM
42
1
0
01 Jun 2023
Jointly Reparametrized Multi-Layer Adaptation for Efficient and Private
  Tuning
Jointly Reparametrized Multi-Layer Adaptation for Efficient and Private Tuning
Umang Gupta
Aram Galstyan
Greg Ver Steeg
11
2
0
30 May 2023
PreQuant: A Task-agnostic Quantization Approach for Pre-trained Language
  Models
PreQuant: A Task-agnostic Quantization Approach for Pre-trained Language Models
Zhuocheng Gong
Jiahao Liu
Qifan Wang
Yang Yang
Jingang Wang
Wei Wu
Yunsen Xian
Dongyan Zhao
Rui Yan
MQ
35
5
0
30 May 2023
Universality and Limitations of Prompt Tuning
Universality and Limitations of Prompt Tuning
Yihan Wang
Jatin Chauhan
Wei Wang
Cho-Jui Hsieh
57
17
0
30 May 2023
Domain Specialization as the Key to Make Large Language Models
  Disruptive: A Comprehensive Survey
Domain Specialization as the Key to Make Large Language Models Disruptive: A Comprehensive Survey
Chen Ling
Xujiang Zhao
Jiaying Lu
Chengyuan Deng
Can Zheng
...
Chris White
Quanquan Gu
Jian Pei
Carl Yang
Liang Zhao
ALM
33
126
0
30 May 2023
Efficient Storage of Fine-Tuned Models via Low-Rank Approximation of
  Weight Residuals
Efficient Storage of Fine-Tuned Models via Low-Rank Approximation of Weight Residuals
Simo Ryu
S. Seo
Jaejun Yoo
37
5
0
28 May 2023
Plug-and-Play Knowledge Injection for Pre-trained Language Models
Plug-and-Play Knowledge Injection for Pre-trained Language Models
Zhengyan Zhang
Zhiyuan Zeng
Yankai Lin
Huadong Wang
Deming Ye
...
Xu Han
Zhiyuan Liu
Peng Li
Maosong Sun
Jie Zhou
KELM
46
11
0
28 May 2023
Stochastic Bridges as Effective Regularizers for Parameter-Efficient
  Tuning
Stochastic Bridges as Effective Regularizers for Parameter-Efficient Tuning
Weize Chen
Xu Han
Yankai Lin
Zhiyuan Liu
Maosong Sun
Jie Zhou
21
1
0
28 May 2023
Plug-and-Play Document Modules for Pre-trained Models
Plug-and-Play Document Modules for Pre-trained Models
Chaojun Xiao
Zhengyan Zhang
Xu Han
Chi-Min Chan
Yankai Lin
Zhiyuan Liu
Xiangyang Li
Zhonghua Li
Bo Zhao
Maosong Sun
KELM
34
5
0
28 May 2023
Fine-tuning Happens in Tiny Subspaces: Exploring Intrinsic Task-specific
  Subspaces of Pre-trained Language Models
Fine-tuning Happens in Tiny Subspaces: Exploring Intrinsic Task-specific Subspaces of Pre-trained Language Models
Zhong Zhang
Bang Liu
Junming Shao
30
7
0
27 May 2023
Weaker Than You Think: A Critical Look at Weakly Supervised Learning
Weaker Than You Think: A Critical Look at Weakly Supervised Learning
D. Zhu
Xiaoyu Shen
Marius Mosbach
Andreas Stephan
Dietrich Klakow
NoLa
39
9
0
27 May 2023
Do We Really Need a Large Number of Visual Prompts?
Do We Really Need a Large Number of Visual Prompts?
Youngeun Kim
Yuhang Li
Abhishek Moitra
Ruokai Yin
Priyadarshini Panda
VLM
VPVLM
46
5
0
26 May 2023
Few-shot Fine-tuning vs. In-context Learning: A Fair Comparison and
  Evaluation
Few-shot Fine-tuning vs. In-context Learning: A Fair Comparison and Evaluation
Marius Mosbach
Tiago Pimentel
Shauli Ravfogel
Dietrich Klakow
Yanai Elazar
52
124
0
26 May 2023
Parameter-Efficient Fine-Tuning without Introducing New Latency
Parameter-Efficient Fine-Tuning without Introducing New Latency
Baohao Liao
Yan Meng
Christof Monz
24
49
0
26 May 2023
Neural Architecture Search for Parameter-Efficient Fine-tuning of Large
  Pre-trained Language Models
Neural Architecture Search for Parameter-Efficient Fine-tuning of Large Pre-trained Language Models
Neal Lawton
Anoop Kumar
Govind Thattai
Aram Galstyan
Greg Ver Steeg
25
16
0
26 May 2023
READ: Recurrent Adaptation of Large Transformers
READ: Recurrent Adaptation of Large Transformers
Sida I. Wang
John Nguyen
Ke Li
Carole-Jean Wu
28
11
0
24 May 2023
MultiFusion: Fusing Pre-Trained Models for Multi-Lingual, Multi-Modal
  Image Generation
MultiFusion: Fusing Pre-Trained Models for Multi-Lingual, Multi-Modal Image Generation
Marco Bellagente
Manuel Brack
H. Teufel
Felix Friedrich
Bjorn Deiseroth
...
Koen Oostermeijer
Andres Felipe Cruz Salinas
P. Schramowski
Kristian Kersting
Samuel Weinbach
45
16
0
24 May 2023
Winner-Take-All Column Row Sampling for Memory Efficient Adaptation of
  Language Model
Winner-Take-All Column Row Sampling for Memory Efficient Adaptation of Language Model
Zirui Liu
Guanchu Wang
Shaochen Zhong
Zhaozhuo Xu
Daochen Zha
...
Zhimeng Jiang
Kaixiong Zhou
V. Chaudhary
Shuai Xu
Xia Hu
47
12
0
24 May 2023
Towards Adaptive Prefix Tuning for Parameter-Efficient Language Model
  Fine-tuning
Towards Adaptive Prefix Tuning for Parameter-Efficient Language Model Fine-tuning
Zhen-Ru Zhang
Chuanqi Tan
Haiyang Xu
Chengyu Wang
Jun Huang
Songfang Huang
33
29
0
24 May 2023
Bactrian-X: Multilingual Replicable Instruction-Following Models with
  Low-Rank Adaptation
Bactrian-X: Multilingual Replicable Instruction-Following Models with Low-Rank Adaptation
Haonan Li
Fajri Koto
Minghao Wu
Alham Fikri Aji
Timothy Baldwin
ALM
16
74
0
24 May 2023
Few-shot Unified Question Answering: Tuning Models or Prompts?
Few-shot Unified Question Answering: Tuning Models or Prompts?
Srijan Bansal
Semih Yavuz
Bo Pang
Meghana Moorthy Bhat
Yingbo Zhou
36
2
0
23 May 2023
QLoRA: Efficient Finetuning of Quantized LLMs
QLoRA: Efficient Finetuning of Quantized LLMs
Tim Dettmers
Artidoro Pagnoni
Ari Holtzman
Luke Zettlemoyer
ALM
73
2,361
0
23 May 2023
TaskWeb: Selecting Better Source Tasks for Multi-task NLP
TaskWeb: Selecting Better Source Tasks for Multi-task NLP
Joongwon Kim
Akari Asai
Gabriel Ilharco
Hannaneh Hajishirzi
29
11
0
22 May 2023
SPARSEFIT: Few-shot Prompting with Sparse Fine-tuning for Jointly
  Generating Predictions and Natural Language Explanations
SPARSEFIT: Few-shot Prompting with Sparse Fine-tuning for Jointly Generating Predictions and Natural Language Explanations
Jesus Solano
Oana-Maria Camburu
Pasquale Minervini
20
1
0
22 May 2023
TADA: Efficient Task-Agnostic Domain Adaptation for Transformers
TADA: Efficient Task-Agnostic Domain Adaptation for Transformers
Chia-Chien Hung
Lukas Lange
Jannik Strötgen
30
9
0
22 May 2023
Prefix Propagation: Parameter-Efficient Tuning for Long Sequences
Prefix Propagation: Parameter-Efficient Tuning for Long Sequences
Jonathan Li
Will Aitken
R. Bhambhoria
Xiao-Dan Zhu
17
14
0
20 May 2023
ToolkenGPT: Augmenting Frozen Language Models with Massive Tools via
  Tool Embeddings
ToolkenGPT: Augmenting Frozen Language Models with Massive Tools via Tool Embeddings
Shibo Hao
Tianyang Liu
Zhen Wang
Zhiting Hu
RALM
LLMAG
66
173
0
19 May 2023
A Parameter-Efficient Learning Approach to Arabic Dialect Identification
  with Pre-Trained General-Purpose Speech Model
A Parameter-Efficient Learning Approach to Arabic Dialect Identification with Pre-Trained General-Purpose Speech Model
S. Radhakrishnan
Chao-Han Huck Yang
S. Khan
N. Kiani
D. Gómez-Cabrero
Jesper N. Tegnér
28
12
0
18 May 2023
Ahead-of-Time P-Tuning
Ahead-of-Time P-Tuning
Daniil Gavrilov
Nikita Balagansky
40
1
0
18 May 2023
G-Adapter: Towards Structure-Aware Parameter-Efficient Transfer Learning
  for Graph Transformer Networks
G-Adapter: Towards Structure-Aware Parameter-Efficient Transfer Learning for Graph Transformer Networks
Anchun Gui
Jinqiang Ye
Han Xiao
27
19
0
17 May 2023
When Gradient Descent Meets Derivative-Free Optimization: A Match Made
  in Black-Box Scenario
When Gradient Descent Meets Derivative-Free Optimization: A Match Made in Black-Box Scenario
Chengcheng Han
Liqing Cui
Renyu Zhu
Jie Wang
Nuo Chen
Qiushi Sun
Xiang Li
Ming Gao
36
7
0
17 May 2023
Knowledge Card: Filling LLMs' Knowledge Gaps with Plug-in Specialized
  Language Models
Knowledge Card: Filling LLMs' Knowledge Gaps with Plug-in Specialized Language Models
Shangbin Feng
Weijia Shi
Yuyang Bai
Vidhisha Balachandran
Tianxing He
Yulia Tsvetkov
KELM
50
30
0
17 May 2023
Memorization for Good: Encryption with Autoregressive Language Models
Memorization for Good: Encryption with Autoregressive Language Models
Samuel Stevens
Yung-Chun Su
20
0
0
15 May 2023
Parameter-Efficient Fine-Tuning with Layer Pruning on Free-Text
  Sequence-to-Sequence Modeling
Parameter-Efficient Fine-Tuning with Layer Pruning on Free-Text Sequence-to-Sequence Modeling
Y. Zhu
Xuebing Yang
Yuanyuan Wu
Wensheng Zhang
MedIm
25
2
0
15 May 2023
Parameter-Efficient Fine-Tuning for Medical Image Analysis: The Missed
  Opportunity
Parameter-Efficient Fine-Tuning for Medical Image Analysis: The Missed Opportunity
Raman Dutt
Linus Ericsson
Pedro Sanchez
Sotirios A. Tsaftaris
Timothy M. Hospedales
MedIm
35
50
0
14 May 2023
Make Prompt-based Black-Box Tuning Colorful: Boosting Model
  Generalization from Three Orthogonal Perspectives
Make Prompt-based Black-Box Tuning Colorful: Boosting Model Generalization from Three Orthogonal Perspectives
Qiushi Sun
Chengcheng Han
Nuo Chen
Renyu Zhu
Jing Gong
Xiang Li
Ming Gao
VLM
27
8
0
14 May 2023
ArtGPT-4: Towards Artistic-understanding Large Vision-Language Models
  with Enhanced Adapter
ArtGPT-4: Towards Artistic-understanding Large Vision-Language Models with Enhanced Adapter
Zheng Yuan
HU Xue
Kun Wang
Yongming Liu
Kun Wang
VLM
MLLM
29
5
0
12 May 2023
Previous
123...111213...151617
Next