Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1811.01088
Cited By
Sentence Encoders on STILTs: Supplementary Training on Intermediate Labeled-data Tasks
2 November 2018
Jason Phang
Thibault Févry
Samuel R. Bowman
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Sentence Encoders on STILTs: Supplementary Training on Intermediate Labeled-data Tasks"
50 / 119 papers shown
Title
Geographic Adaptation of Pretrained Language Models
Valentin Hofmann
Goran Glavavs
Nikola Ljubevsić
J. Pierrehumbert
Hinrich Schütze
VLM
21
16
0
16 Mar 2022
Hyperdecoders: Instance-specific decoders for multi-task NLP
Hamish Ivison
Matthew E. Peters
AI4CE
34
20
0
15 Mar 2022
Memory Efficient Continual Learning with Transformers
Beyza Ermis
Giovanni Zappella
Martin Wistuba
Aditya Rawal
Cédric Archambeau
CLL
34
43
0
09 Mar 2022
WANLI: Worker and AI Collaboration for Natural Language Inference Dataset Creation
Alisa Liu
Swabha Swayamdipta
Noah A. Smith
Yejin Choi
82
212
0
16 Jan 2022
Interpretable Low-Resource Legal Decision Making
R. Bhambhoria
Hui Liu
Samuel Dahan
Xiao-Dan Zhu
ELM
AILaw
32
9
0
01 Jan 2022
QuALITY: Question Answering with Long Input Texts, Yes!
Richard Yuanzhe Pang
Alicia Parrish
Nitish Joshi
Nikita Nangia
Jason Phang
...
Vishakh Padmakumar
Johnny Ma
Jana Thompson
He He
Sam Bowman
RALM
30
141
0
16 Dec 2021
Merging Models with Fisher-Weighted Averaging
Michael Matena
Colin Raffel
FedML
MoMe
50
352
0
18 Nov 2021
On Transferability of Prompt Tuning for Natural Language Processing
Yusheng Su
Xiaozhi Wang
Yujia Qin
Chi-Min Chan
Yankai Lin
...
Peng Li
Juanzi Li
Lei Hou
Maosong Sun
Jie Zhou
AAML
VLM
26
98
0
12 Nov 2021
Learning to Generalize Compositionally by Transferring Across Semantic Parsing Tasks
Wang Zhu
Peter Shaw
Tal Linzen
Fei Sha
35
7
0
09 Nov 2021
CLUES: Few-Shot Learning Evaluation in Natural Language Understanding
Subhabrata Mukherjee
Xiaodong Liu
Guoqing Zheng
Saghar Hosseini
Hao Cheng
Greg Yang
Christopher Meek
Ahmed Hassan Awadallah
Jianfeng Gao
ELM
33
11
0
04 Nov 2021
Diverse Distributions of Self-Supervised Tasks for Meta-Learning in NLP
Trapit Bansal
K. Gunasekaran
Tong Wang
Tsendsuren Munkhdalai
Andrew McCallum
SSL
OOD
51
19
0
02 Nov 2021
Leveraging the Inductive Bias of Large Language Models for Abstract Textual Reasoning
Christopher Rytting
David Wingate
AI4CE
LRM
18
26
0
05 Oct 2021
DeepA2: A Modular Framework for Deep Argument Analysis with Pretrained Neural Text2Text Language Models
Gregor Betz
Kyle Richardson
22
8
0
04 Oct 2021
STraTA: Self-Training with Task Augmentation for Better Few-shot Learning
Tu Vu
Minh-Thang Luong
Quoc V. Le
Grady Simon
Mohit Iyyer
131
61
0
13 Sep 2021
Raise a Child in Large Language Model: Towards Effective and Generalizable Fine-tuning
Runxin Xu
Fuli Luo
Zhiyuan Zhang
Chuanqi Tan
Baobao Chang
Songfang Huang
Fei Huang
LRM
151
178
0
13 Sep 2021
Rethinking Why Intermediate-Task Fine-Tuning Works
Ting-Yun Chang
Chi-Jen Lu
LRM
19
29
0
26 Aug 2021
Robust Transfer Learning with Pretrained Language Models through Adapters
Wenjuan Han
Bo Pang
Ying Nian Wu
16
54
0
05 Aug 2021
Domain-matched Pre-training Tasks for Dense Retrieval
Barlas Oğuz
Kushal Lakhotia
Anchit Gupta
Patrick Lewis
Vladimir Karpukhin
...
Xilun Chen
Sebastian Riedel
Wen-tau Yih
Sonal Gupta
Yashar Mehdad
RALM
30
66
0
28 Jul 2021
The Effectiveness of Intermediate-Task Training for Code-Switched Natural Language Understanding
Archiki Prasad
Mohammad Ali Rehan
Shreyasi Pathak
P. Jyothi
27
9
0
21 Jul 2021
CausalBERT: Injecting Causal Knowledge Into Pre-trained Models with Minimal Supervision
Zhongyang Li
Xiao Ding
Kuo Liao
Bing Qin
Ting Liu
CML
29
17
0
21 Jul 2021
Noise Stability Regularization for Improving BERT Fine-tuning
Hang Hua
Xingjian Li
Dejing Dou
Chengzhong Xu
Jiebo Luo
19
43
0
10 Jul 2021
The MultiBERTs: BERT Reproductions for Robustness Analysis
Thibault Sellam
Steve Yadlowsky
Jason W. Wei
Naomi Saphra
Alexander DÁmour
...
Iulia Turc
Jacob Eisenstein
Dipanjan Das
Ian Tenney
Ellie Pavlick
24
93
0
30 Jun 2021
Specializing Multilingual Language Models: An Empirical Study
Ethan C. Chau
Noah A. Smith
27
27
0
16 Jun 2021
Question Answering Infused Pre-training of General-Purpose Contextualized Representations
Robin Jia
M. Lewis
Luke Zettlemoyer
23
28
0
15 Jun 2021
Entailment as Few-Shot Learner
Sinong Wang
Han Fang
Madian Khabsa
Hanzi Mao
Hao Ma
35
183
0
29 Apr 2021
Multilingual and Cross-Lingual Intent Detection from Spoken Data
D. Gerz
Pei-hao Su
Razvan Kusztos
Avishek Mondal
M. Lis
Eshan Singhal
N. Mrksic
Tsung-Hsien Wen
Ivan Vulić
19
35
0
17 Apr 2021
AMMU : A Survey of Transformer-based Biomedical Pretrained Language Models
Katikapalli Subramanyam Kalyan
A. Rajasekharan
S. Sangeetha
LM&MA
MedIm
26
164
0
16 Apr 2021
What to Pre-Train on? Efficient Intermediate Task Selection
Clifton A. Poth
Jonas Pfeiffer
Andreas Rucklé
Iryna Gurevych
21
94
0
16 Apr 2021
How to Train BERT with an Academic Budget
Peter Izsak
Moshe Berchansky
Omer Levy
23
113
0
15 Apr 2021
FEWS: Large-Scale, Low-Shot Word Sense Disambiguation with the Dictionary
Terra Blevins
Mandar Joshi
Luke Zettlemoyer
36
19
0
16 Feb 2021
Making Pre-trained Language Models Better Few-shot Learners
Tianyu Gao
Adam Fisch
Danqi Chen
243
1,927
0
31 Dec 2020
A Closer Look at Few-Shot Crosslingual Transfer: The Choice of Shots Matters
Mengjie Zhao
Yi Zhu
Ehsan Shareghi
Ivan Vulić
Roi Reichart
Anna Korhonen
Hinrich Schütze
32
64
0
31 Dec 2020
A Simple and Efficient Multi-Task Learning Approach for Conditioned Dialogue Generation
Yan Zeng
J. Nie
21
5
0
21 Oct 2020
Pretrained Transformers for Text Ranking: BERT and Beyond
Jimmy J. Lin
Rodrigo Nogueira
Andrew Yates
VLM
244
612
0
13 Oct 2020
Effective Unsupervised Domain Adaptation with Adversarially Trained Language Models
Thuy-Trang Vu
Dinh Q. Phung
Gholamreza Haffari
16
24
0
05 Oct 2020
How to Motivate Your Dragon: Teaching Goal-Driven Agents to Speak and Act in Fantasy Worlds
Prithviraj Ammanabrolu
Jack Urbanek
Margaret Li
Arthur Szlam
Tim Rocktaschel
Jason Weston
LM&Ro
19
44
0
01 Oct 2020
Dataset Cartography: Mapping and Diagnosing Datasets with Training Dynamics
Swabha Swayamdipta
Roy Schwartz
Nicholas Lourie
Yizhong Wang
Hannaneh Hajishirzi
Noah A. Smith
Yejin Choi
44
429
0
22 Sep 2020
Repulsive Attention: Rethinking Multi-head Attention as Bayesian Inference
Bang An
Jie Lyu
Zhenyi Wang
Chunyuan Li
Changwei Hu
Fei Tan
Ruiyi Zhang
Yifan Hu
Changyou Chen
AAML
22
28
0
20 Sep 2020
Conditionally Adaptive Multi-Task Learning: Improving Transfer Learning in NLP Using Fewer Parameters & Less Data
Jonathan Pilault
Amine Elhattami
C. Pal
CLL
MoE
30
89
0
19 Sep 2020
Self-Supervised Meta-Learning for Few-Shot Natural Language Classification Tasks
Trapit Bansal
Rishikesh Jha
Tsendsuren Munkhdalai
Andrew McCallum
SSL
VLM
27
87
0
17 Sep 2020
Critical Thinking for Language Models
Gregor Betz
Christian Voigt
Kyle Richardson
SyDa
ReLM
LRM
AI4CE
23
35
0
15 Sep 2020
Transferability of Natural Language Inference to Biomedical Question Answering
Minbyul Jeong
Mujeen Sung
Gangwoo Kim
Donghyeon Kim
Wonjin Yoon
J. Yoo
Jaewoo Kang
19
38
0
01 Jul 2020
Revisiting Few-sample BERT Fine-tuning
Tianyi Zhang
Felix Wu
Arzoo Katiyar
Kilian Q. Weinberger
Yoav Artzi
41
441
0
10 Jun 2020
On the Stability of Fine-tuning BERT: Misconceptions, Explanations, and Strong Baselines
Marius Mosbach
Maksym Andriushchenko
Dietrich Klakow
31
352
0
08 Jun 2020
DeCLUTR: Deep Contrastive Learning for Unsupervised Textual Representations
John Giorgi
Osvald Nitski
Bo Wang
Gary D. Bader
SSL
39
490
0
05 Jun 2020
ADAHESSIAN: An Adaptive Second Order Optimizer for Machine Learning
Z. Yao
A. Gholami
Sheng Shen
Mustafa Mustafa
Kurt Keutzer
Michael W. Mahoney
ODL
39
275
0
01 Jun 2020
Language Models are Few-Shot Learners
Tom B. Brown
Benjamin Mann
Nick Ryder
Melanie Subbiah
Jared Kaplan
...
Christopher Berner
Sam McCandlish
Alec Radford
Ilya Sutskever
Dario Amodei
BDL
77
40,200
0
28 May 2020
AdapterFusion: Non-Destructive Task Composition for Transfer Learning
Jonas Pfeiffer
Aishwarya Kamath
Andreas Rucklé
Kyunghyun Cho
Iryna Gurevych
CLL
MoMe
47
817
0
01 May 2020
Pre-training Is (Almost) All You Need: An Application to Commonsense Reasoning
Alexandre Tamborrino
Nicola Pellicanò
B. Pannier
Pascal Voitot
Louise Naudin
LRM
22
62
0
29 Apr 2020
Recall and Learn: Fine-tuning Deep Pretrained Language Models with Less Forgetting
Sanyuan Chen
Yutai Hou
Yiming Cui
Wanxiang Che
Ting Liu
Xiangzhan Yu
KELM
CLL
21
213
0
27 Apr 2020
Previous
1
2
3
Next