ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1903.05987
  4. Cited By
To Tune or Not to Tune? Adapting Pretrained Representations to Diverse
  Tasks

To Tune or Not to Tune? Adapting Pretrained Representations to Diverse Tasks

14 March 2019
Matthew E. Peters
Sebastian Ruder
Noah A. Smith
ArXivPDFHTML

Papers citing "To Tune or Not to Tune? Adapting Pretrained Representations to Diverse Tasks"

50 / 229 papers shown
Title
Neural Architecture Search for Parameter-Efficient Fine-tuning of Large
  Pre-trained Language Models
Neural Architecture Search for Parameter-Efficient Fine-tuning of Large Pre-trained Language Models
Neal Lawton
Anoop Kumar
Govind Thattai
Aram Galstyan
Greg Ver Steeg
25
16
0
26 May 2023
Learning Easily Updated General Purpose Text Representations with
  Adaptable Task-Specific Prefixes
Learning Easily Updated General Purpose Text Representations with Adaptable Task-Specific Prefixes
Kuan-Hao Huang
L Tan
Rui Hou
Sinong Wang
Amjad Almahairi
Ruty Rinott
AI4CE
33
0
0
22 May 2023
Advancing Neural Encoding of Portuguese with Transformer Albertina PT-*
Advancing Neural Encoding of Portuguese with Transformer Albertina PT-*
João Rodrigues
Luís Gomes
Joao Silva
António Branco
Rodrigo Santos
Henrique Lopes Cardoso
T. Osório
22
43
0
11 May 2023
The EarlyBIRD Catches the Bug: On Exploiting Early Layers of Encoder
  Models for More Efficient Code Classification
The EarlyBIRD Catches the Bug: On Exploiting Early Layers of Encoder Models for More Efficient Code Classification
Anastasiia Grishina
Max Hort
Leon Moonen
22
6
0
08 May 2023
Using Language Models on Low-end Hardware
Using Language Models on Low-end Hardware
Silin Gao
Beatriz Borges
Saya Kanno
Antoine Bosselut
21
0
0
03 May 2023
Reliable Gradient-free and Likelihood-free Prompt Tuning
Reliable Gradient-free and Likelihood-free Prompt Tuning
Maohao Shen
S. Ghosh
P. Sattigeri
Subhro Das
Yuheng Bu
G. Wornell
VLM
50
10
0
30 Apr 2023
Analyzing the Generalizability of Deep Contextualized Language
  Representations For Text Classification
Analyzing the Generalizability of Deep Contextualized Language Representations For Text Classification
Berfu Buyukoz
28
2
0
22 Mar 2023
How to prepare your task head for finetuning
How to prepare your task head for finetuning
Yi Ren
Shangmin Guo
Wonho Bae
Danica J. Sutherland
24
14
0
11 Feb 2023
Zero-Shot Learning for Requirements Classification: An Exploratory Study
Zero-Shot Learning for Requirements Classification: An Exploratory Study
Waad Alhoshan
Alessio Ferrari
Liping Zhao
VLM
9
39
0
09 Feb 2023
Rover: An online Spark SQL tuning service via generalized transfer
  learning
Rover: An online Spark SQL tuning service via generalized transfer learning
Yu Shen
Xinyuyang Ren
Yupeng Lu
Huaijun Jiang
Huanyong Xu
Di Peng
Yang Li
Wentao Zhang
Tengjiao Wang
34
10
0
08 Feb 2023
An Empirical Study on the Transferability of Transformer Modules in
  Parameter-Efficient Fine-Tuning
An Empirical Study on the Transferability of Transformer Modules in Parameter-Efficient Fine-Tuning
Mohammad AkbarTajari
S. Rajaee
Mohammad Taher Pilehvar
11
2
0
01 Feb 2023
A Stability Analysis of Fine-Tuning a Pre-Trained Model
A Stability Analysis of Fine-Tuning a Pre-Trained Model
Z. Fu
Anthony Man-Cho So
Nigel Collier
23
3
0
24 Jan 2023
Pretraining Without Attention
Pretraining Without Attention
Junxiong Wang
J. Yan
Albert Gu
Alexander M. Rush
27
48
0
20 Dec 2022
Pre-Trained Image Encoder for Generalizable Visual Reinforcement
  Learning
Pre-Trained Image Encoder for Generalizable Visual Reinforcement Learning
Zhecheng Yuan
Zhengrong Xue
Bo Yuan
Xueqian Wang
Yi Wu
Yang Gao
Huazhe Xu
SSL
OffRL
46
70
0
17 Dec 2022
Investigation of Japanese PnG BERT language model in text-to-speech
  synthesis for pitch accent language
Investigation of Japanese PnG BERT language model in text-to-speech synthesis for pitch accent language
Yusuke Yasuda
T. Toda
33
8
0
16 Dec 2022
Searching for Effective Multilingual Fine-Tuning Methods: A Case Study
  in Summarization
Searching for Effective Multilingual Fine-Tuning Methods: A Case Study in Summarization
Yiwei Qin
Graham Neubig
Pengfei Liu
23
3
0
12 Dec 2022
Zero-Shot Learning for Joint Intent and Slot Labeling
Zero-Shot Learning for Joint Intent and Slot Labeling
Rashmi Gangadharaiah
Balakrishnan Narayanaswamy
VLM
18
2
0
29 Nov 2022
Relationship of the language distance to English ability of a country
Relationship of the language distance to English ability of a country
Cao Xinxin
Lei Xiaolan
Ahmed Murtadha
22
1
0
15 Nov 2022
PATS: Sensitivity-aware Noisy Learning for Pretrained Language Models
PATS: Sensitivity-aware Noisy Learning for Pretrained Language Models
Yupeng Zhang
Hongzhi Zhang
Sirui Wang
Wei Wu
Zhoujun Li
AAML
37
1
0
22 Oct 2022
Clip-Tuning: Towards Derivative-free Prompt Learning with a Mixture of
  Rewards
Clip-Tuning: Towards Derivative-free Prompt Learning with a Mixture of Rewards
Yekun Chai
Shuohuan Wang
Yu Sun
Hao Tian
Hua Wu
Haifeng Wang
VLM
24
17
0
21 Oct 2022
Evidence > Intuition: Transferability Estimation for Encoder Selection
Evidence > Intuition: Transferability Estimation for Encoder Selection
Elisa Bassignana
Max Müller-Eberstein
Mike Zhang
Barbara Plank
38
8
0
20 Oct 2022
Hidden State Variability of Pretrained Language Models Can Guide
  Computation Reduction for Transfer Learning
Hidden State Variability of Pretrained Language Models Can Guide Computation Reduction for Transfer Learning
Shuo Xie
Jiahao Qiu
Ankita Pasad
Li Du
Qing Qu
Hongyuan Mei
35
16
0
18 Oct 2022
Post-hoc analysis of Arabic transformer models
Post-hoc analysis of Arabic transformer models
Ahmed Abdelali
Nadir Durrani
Fahim Dalvi
Hassan Sajjad
15
1
0
18 Oct 2022
Intersection of Parallels as an Early Stopping Criterion
Intersection of Parallels as an Early Stopping Criterion
Ali Vardasbi
Maarten de Rijke
Mostafa Dehghani
MoMe
41
5
0
19 Aug 2022
On Transfer of Adversarial Robustness from Pretraining to Downstream
  Tasks
On Transfer of Adversarial Robustness from Pretraining to Downstream Tasks
Laura Fee Nern
Harsh Raj
Maurice Georgi
Yash Sharma
AAML
31
2
0
07 Aug 2022
DeepEmotex: Classifying Emotion in Text Messages using Deep Transfer
  Learning
DeepEmotex: Classifying Emotion in Text Messages using Deep Transfer Learning
Maryam Hasan
Elke A. Rundensteiner
E. Agu
VLM
14
7
0
12 Jun 2022
Challenges and Opportunities in Offline Reinforcement Learning from
  Visual Observations
Challenges and Opportunities in Offline Reinforcement Learning from Visual Observations
Cong Lu
Philip J. Ball
Tim G. J. Rudner
Jack Parker-Holder
Michael A. Osborne
Yee Whye Teh
OffRL
32
52
0
09 Jun 2022
Annotation Error Detection: Analyzing the Past and Present for a More
  Coherent Future
Annotation Error Detection: Analyzing the Past and Present for a More Coherent Future
Jan-Christoph Klie
Bonnie Webber
Iryna Gurevych
42
43
0
05 Jun 2022
What Drives the Use of Metaphorical Language? Negative Insights from
  Abstractness, Affect, Discourse Coherence and Contextualized Word
  Representations
What Drives the Use of Metaphorical Language? Negative Insights from Abstractness, Affect, Discourse Coherence and Contextualized Word Representations
P. Piccirilli
Sabine Schulte im Walde
22
4
0
23 May 2022
Bridging the Domain Gap for Stance Detection for the Zulu language
Bridging the Domain Gap for Stance Detection for the Zulu language
Gcinizwe Dlamini
I. E. I. Bekkouch
A. Khan
Leon Derczynski
23
3
0
06 May 2022
Explain and Conquer: Personalised Text-based Reviews to Achieve
  Transparency
Explain and Conquer: Personalised Text-based Reviews to Achieve Transparency
Iñigo López-Riobóo Botana
V. Bolón-Canedo
B. Guijarro-Berdiñas
Amparo Alonso-Betanzos
26
2
0
03 May 2022
BERTops: Studying BERT Representations under a Topological Lens
BERTops: Studying BERT Representations under a Topological Lens
Jatin Chauhan
Manohar Kaul
24
3
0
02 May 2022
It Takes Two Flints to Make a Fire: Multitask Learning of Neural
  Relation and Explanation Classifiers
It Takes Two Flints to Make a Fire: Multitask Learning of Neural Relation and Explanation Classifiers
Zheng Tang
Mihai Surdeanu
27
6
0
25 Apr 2022
You Are What You Write: Preserving Privacy in the Era of Large Language
  Models
You Are What You Write: Preserving Privacy in the Era of Large Language Models
Richard Plant
V. Giuffrida
Dimitra Gkatzia
PILM
35
19
0
20 Apr 2022
PERFECT: Prompt-free and Efficient Few-shot Learning with Language
  Models
PERFECT: Prompt-free and Efficient Few-shot Learning with Language Models
Rabeeh Karimi Mahabadi
Luke Zettlemoyer
James Henderson
Marzieh Saeidi
Lambert Mathias
Ves Stoyanov
Majid Yazdani
VLM
34
69
0
03 Apr 2022
BERT-ASC: Implicit Aspect Representation Learning through
  Auxiliary-Sentence Construction for Sentiment Analysis
BERT-ASC: Implicit Aspect Representation Learning through Auxiliary-Sentence Construction for Sentiment Analysis
Ahmed Murtadha
Shengfeng Pan
Jianlin Su
Xinxin Cao
Wenze Zhang
Bo Wen
Yunfeng Liu
34
1
0
22 Mar 2022
The Unsurprising Effectiveness of Pre-Trained Vision Models for Control
The Unsurprising Effectiveness of Pre-Trained Vision Models for Control
Simone Parisi
Aravind Rajeswaran
Senthil Purushwalkam
Abhinav Gupta
LM&Ro
34
187
0
07 Mar 2022
Fine-Tuning can Distort Pretrained Features and Underperform
  Out-of-Distribution
Fine-Tuning can Distort Pretrained Features and Underperform Out-of-Distribution
Ananya Kumar
Aditi Raghunathan
Robbie Jones
Tengyu Ma
Percy Liang
OODD
50
642
0
21 Feb 2022
FAMIE: A Fast Active Learning Framework for Multilingual Information
  Extraction
FAMIE: A Fast Active Learning Framework for Multilingual Information Extraction
Minh Le Nguyen
Nghia Trung Ngo
Bonan Min
Thien Huu Nguyen
22
10
0
16 Feb 2022
UserBERT: Modeling Long- and Short-Term User Preferences via
  Self-Supervision
UserBERT: Modeling Long- and Short-Term User Preferences via Self-Supervision
Tianyu Li
Ali Cevahir
Derek Cho
Hao Gong
Duy Nguyen
B. Stenger
SSL
18
1
0
14 Feb 2022
Assessment of contextualised representations in detecting outcome
  phrases in clinical trials
Assessment of contextualised representations in detecting outcome phrases in clinical trials
Micheal Abaho
Danushka Bollegala
P. Williamson
S. Dodd
21
8
0
13 Feb 2022
A Semi-Supervised Deep Clustering Pipeline for Mining Intentions From
  Texts
A Semi-Supervised Deep Clustering Pipeline for Mining Intentions From Texts
Xinyu Chen
Ian Beaver
VLM
19
0
0
01 Feb 2022
Black-box Prompt Learning for Pre-trained Language Models
Black-box Prompt Learning for Pre-trained Language Models
Shizhe Diao
Zhichao Huang
Ruijia Xu
Xuechun Li
Yong Lin
Xiao Zhou
Tong Zhang
VLM
AAML
36
68
0
21 Jan 2022
Interpreting Arabic Transformer Models
Ahmed Abdelali
Nadir Durrani
Fahim Dalvi
Hassan Sajjad
41
2
0
19 Jan 2022
Black-Box Tuning for Language-Model-as-a-Service
Black-Box Tuning for Language-Model-as-a-Service
Tianxiang Sun
Yunfan Shao
Hong Qian
Xuanjing Huang
Xipeng Qiu
VLM
50
256
0
10 Jan 2022
Domain Adaptation with Pre-trained Transformers for Query Focused
  Abstractive Text Summarization
Domain Adaptation with Pre-trained Transformers for Query Focused Abstractive Text Summarization
Md Tahmid Rahman Laskar
Enamul Hoque
J. Huang
36
45
0
22 Dec 2021
An Empirical Investigation of the Role of Pre-training in Lifelong
  Learning
An Empirical Investigation of the Role of Pre-training in Lifelong Learning
Sanket Vaibhav Mehta
Darshan Patil
Sarath Chandar
Emma Strubell
CLL
45
136
0
16 Dec 2021
Measuring Context-Word Biases in Lexical Semantic Datasets
Measuring Context-Word Biases in Lexical Semantic Datasets
Qianchu Liu
Diana McCarthy
Anna Korhonen
31
2
0
13 Dec 2021
UniLog: Deploy One Model and Specialize it for All Log Analysis Tasks
UniLog: Deploy One Model and Specialize it for All Log Analysis Tasks
Yichen Zhu
Weibin Meng
Ying Liu
Shenglin Zhang
Tao Han
Shimin Tao
Dan Pei
MoE
41
14
0
06 Dec 2021
Towards More Robust Natural Language Understanding
Towards More Robust Natural Language Understanding
Xinliang Frederick Zhang
22
2
0
01 Dec 2021
Previous
12345
Next