ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1905.00537
  4. Cited By
SuperGLUE: A Stickier Benchmark for General-Purpose Language
  Understanding Systems
v1v2v3 (latest)

SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems

2 May 2019
Alex Jinpeng Wang
Yada Pruksachatkun
Nikita Nangia
Amanpreet Singh
Julian Michael
Felix Hill
Omer Levy
Samuel R. Bowman
    ELM
ArXiv (abs)PDFHTML

Papers citing "SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems"

50 / 1,500 papers shown
Title
Sparse Structure Search for Parameter-Efficient Tuning
Sparse Structure Search for Parameter-Efficient Tuning
Shengding Hu
Zhen Zhang
Ning Ding
Yadao Wang
Yasheng Wang
Zhiyuan Liu
Maosong Sun
76
17
0
15 Jun 2022
Language Models are General-Purpose Interfaces
Language Models are General-Purpose Interfaces
Y. Hao
Haoyu Song
Li Dong
Shaohan Huang
Zewen Chi
Wenhui Wang
Shuming Ma
Furu Wei
MLLM
76
102
0
13 Jun 2022
Improving Pre-trained Language Model Fine-tuning with Noise Stability
  Regularization
Improving Pre-trained Language Model Fine-tuning with Noise Stability Regularization
Hang Hua
Xingjian Li
Dejing Dou
Chengzhong Xu
Jiebo Luo
94
15
0
12 Jun 2022
Positional Label for Self-Supervised Vision Transformer
Positional Label for Self-Supervised Vision Transformer
Zhemin Zhang
Xun Gong
ViTMDE
59
6
0
10 Jun 2022
PEER: A Comprehensive and Multi-Task Benchmark for Protein Sequence
  Understanding
PEER: A Comprehensive and Multi-Task Benchmark for Protein Sequence Understanding
Minghao Xu
Zuobai Zhang
Jiarui Lu
Zhaocheng Zhu
Yang Zhang
Chang Ma
Runcheng Liu
Jian Tang
62
108
0
05 Jun 2022
Instance-wise Prompt Tuning for Pretrained Language Models
Instance-wise Prompt Tuning for Pretrained Language Models
Yuezihan Jiang
Hao Yang
Junyang Lin
Hanyu Zhao
An Yang
Chang Zhou
Hongxia Yang
Zhi-Xin Yang
Tengjiao Wang
VLM
59
7
0
04 Jun 2022
MMTM: Multi-Tasking Multi-Decoder Transformer for Math Word Problems
MMTM: Multi-Tasking Multi-Decoder Transformer for Math Word Problems
Keyur Faldu
Amit P. Sheth
Prashant Kikani
Darshan Patel
AIMat
52
1
0
02 Jun 2022
Prompt Injection: Parameterization of Fixed Inputs
Prompt Injection: Parameterization of Fixed Inputs
Eunbi Choi
Yongrae Jo
Joel Jang
Minjoon Seo
119
30
0
31 May 2022
VLUE: A Multi-Task Benchmark for Evaluating Vision-Language Models
VLUE: A Multi-Task Benchmark for Evaluating Vision-Language Models
Wangchunshu Zhou
Yan Zeng
Shizhe Diao
Xinsong Zhang
CoGeVLM
97
13
0
30 May 2022
CEBaB: Estimating the Causal Effects of Real-World Concepts on NLP Model
  Behavior
CEBaB: Estimating the Causal Effects of Real-World Concepts on NLP Model Behavior
Eldar David Abraham
Karel DÓosterlinck
Amir Feder
Y. Gat
Atticus Geiger
Christopher Potts
Roi Reichart
Zhengxuan Wu
CML
122
47
0
27 May 2022
Evaluating Multimodal Interactive Agents
Evaluating Multimodal Interactive Agents
Josh Abramson
Arun Ahuja
Federico Carnevale
Petko Georgiev
Alex Goldin
...
Adam Santoro
Tamara von Glehn
Greg Wayne
Nathaniel Wong
Chen Yan
56
3
0
26 May 2022
Rethinking Fano's Inequality in Ensemble Learning
Rethinking Fano's Inequality in Ensemble Learning
Terufumi Morishita
Gaku Morio
Shota Horiguchi
Hiroaki Ozaki
N. Nukaga
FedML
28
3
0
25 May 2022
Sparse Mixers: Combining MoE and Mixing to build a more efficient BERT
Sparse Mixers: Combining MoE and Mixing to build a more efficient BERT
James Lee-Thorp
Joshua Ainslie
MoE
94
12
0
24 May 2022
TALM: Tool Augmented Language Models
TALM: Tool Augmented Language Models
Aaron T Parisi
Yao-Min Zhao
Noah Fiedel
KELMRALMLLMAG
108
148
0
24 May 2022
ATTEMPT: Parameter-Efficient Multi-task Tuning via Attentional Mixtures
  of Soft Prompts
ATTEMPT: Parameter-Efficient Multi-task Tuning via Attentional Mixtures of Soft Prompts
Akari Asai
Mohammadreza Salehi
Matthew E. Peters
Hannaneh Hajishirzi
192
102
0
24 May 2022
Diverse Lottery Tickets Boost Ensemble from a Single Pretrained Model
Diverse Lottery Tickets Boost Ensemble from a Single Pretrained Model
Sosuke Kobayashi
Shun Kiyono
Jun Suzuki
Kentaro Inui
MoMe
77
9
0
24 May 2022
FlexiBERT: Are Current Transformer Architectures too Homogeneous and
  Rigid?
FlexiBERT: Are Current Transformer Architectures too Homogeneous and Rigid?
Shikhar Tuli
Bhishma Dedhia
Shreshth Tuli
N. Jha
94
14
0
23 May 2022
Simple Recurrence Improves Masked Language Models
Simple Recurrence Improves Masked Language Models
Tao Lei
Ran Tian
Jasmijn Bastings
Ankur P. Parikh
151
4
0
23 May 2022
Diversity Over Size: On the Effect of Sample and Topic Sizes for
  Argument Mining Datasets
Diversity Over Size: On the Effect of Sample and Topic Sizes for Argument Mining Datasets
Benjamin Schiller
Johannes Daxenberger
Iryna Gurevych
85
1
0
23 May 2022
PyRelationAL: a python library for active learning research and
  development
PyRelationAL: a python library for active learning research and development
P. Scherer
Thomas Gaudelet
Alison Pouplin
Alice Del Vecchio
S. SurajM
Oliver Bolton
Jyothish Soman
J. Taylor-King
Lindsay Edwards
KELM
49
0
0
23 May 2022
Vector-Quantized Input-Contextualized Soft Prompts for Natural Language
  Understanding
Vector-Quantized Input-Contextualized Soft Prompts for Natural Language Understanding
Rishabh Bhardwaj
Amrita Saha
Guosheng Lin
Soujanya Poria
VLMVPVLM
51
7
0
23 May 2022
Improving Short Text Classification With Augmented Data Using GPT-3
Improving Short Text Classification With Augmented Data Using GPT-3
Salvador Balkus
Donghui Yan
61
37
0
23 May 2022
Instruction Induction: From Few Examples to Natural Language Task
  Descriptions
Instruction Induction: From Few Examples to Natural Language Task Descriptions
Or Honovich
Uri Shaham
Samuel R. Bowman
Omer Levy
ELMLRM
280
146
0
22 May 2022
Life after BERT: What do Other Muppets Understand about Language?
Life after BERT: What do Other Muppets Understand about Language?
Vladislav Lialin
Kevin Zhao
Namrata Shivagunde
Anna Rumshisky
110
6
0
21 May 2022
Revisiting Pre-trained Language Models and their Evaluation for Arabic
  Natural Language Understanding
Revisiting Pre-trained Language Models and their Evaluation for Arabic Natural Language Understanding
Abbas Ghaddar
Yimeng Wu
Sunyam Bagga
Ahmad Rashid
Khalil Bibi
...
Zhefeng Wang
Baoxing Huai
Xin Jiang
Qun Liu
Philippe Langlais
63
6
0
21 May 2022
Are Prompt-based Models Clueless?
Are Prompt-based Models Clueless?
Pride Kavumba
Ryo Takahashi
Yusuke Oda
VLM
208
13
0
19 May 2022
Learning Rate Curriculum
Learning Rate Curriculum
Florinel-Alin Croitoru
Nicolae-Cătălin Ristea
Radu Tudor Ionescu
N. Sebe
74
9
0
18 May 2022
Predicting Human Psychometric Properties Using Computational Language
  Models
Predicting Human Psychometric Properties Using Computational Language Models
Antonio Laverghetta
Animesh Nighojkar
Jamshidbek Mirzakhalov
John Licato
62
9
0
12 May 2022
Making Pretrained Language Models Good Long-tailed Learners
Making Pretrained Language Models Good Long-tailed Learners
Chen Zhang
Lei Ren
Jingang Wang
Wei Wu
Dawei Song
RALMVLM
81
9
0
11 May 2022
Towards Unified Prompt Tuning for Few-shot Text Classification
Towards Unified Prompt Tuning for Few-shot Text Classification
Jiadong Wang
Chengyu Wang
Fuli Luo
Chuanqi Tan
Minghui Qiu
Fei Yang
Qiuhui Shi
Songfang Huang
Ming Gao
VLM
70
28
0
11 May 2022
UL2: Unifying Language Learning Paradigms
UL2: Unifying Language Learning Paradigms
Yi Tay
Mostafa Dehghani
Vinh Q. Tran
Xavier Garcia
Jason W. Wei
...
Tal Schuster
H. Zheng
Denny Zhou
N. Houlsby
Donald Metzler
AI4CE
139
313
0
10 May 2022
Task-specific Compression for Multi-task Language Models using
  Attribution-based Pruning
Task-specific Compression for Multi-task Language Models using Attribution-based Pruning
Nakyeong Yang
Yunah Jang
Hwanhee Lee
Seohyeong Jung
Kyomin Jung
28
8
0
09 May 2022
Beyond Distributional Hypothesis: Let Language Models Learn Meaning-Text
  Correspondence
Beyond Distributional Hypothesis: Let Language Models Learn Meaning-Text Correspondence
Myeongjun Jang
Frank Mtumbuka
Thomas Lukasiewicz
82
10
0
08 May 2022
HumanAL: Calibrating Human Matching Beyond a Single Task
HumanAL: Calibrating Human Matching Beyond a Single Task
Roee Shraga
HAI
52
6
0
06 May 2022
Knowledge Distillation of Russian Language Models with Reduction of
  Vocabulary
Knowledge Distillation of Russian Language Models with Reduction of Vocabulary
A. Kolesnikova
Yuri Kuratov
Vasily Konovalov
Andrey Kravchenko
VLM
42
10
0
04 May 2022
Are All the Datasets in Benchmark Necessary? A Pilot Study of Dataset
  Evaluation for Text Classification
Are All the Datasets in Benchmark Necessary? A Pilot Study of Dataset Evaluation for Text Classification
Yanghua Xiao
Jinlan Fu
See-Kiong Ng
Pengfei Liu
48
2
0
04 May 2022
Optimizing Mixture of Experts using Dynamic Recompilations
Optimizing Mixture of Experts using Dynamic Recompilations
Ferdinand Kossmann
Zhihao Jia
A. Aiken
96
5
0
04 May 2022
Mixed-effects transformers for hierarchical adaptation
Mixed-effects transformers for hierarchical adaptation
Julia White
Noah D. Goodman
Robert D. Hawkins
46
2
0
03 May 2022
Improving In-Context Few-Shot Learning via Self-Supervised Training
Improving In-Context Few-Shot Learning via Self-Supervised Training
Mingda Chen
Jingfei Du
Ramakanth Pasunuru
Todor Mihaylov
Srini Iyer
Ves Stoyanov
Zornitsa Kozareva
SSLAI4MH
105
67
0
03 May 2022
ElitePLM: An Empirical Study on General Language Ability Evaluation of
  Pretrained Language Models
ElitePLM: An Empirical Study on General Language Ability Evaluation of Pretrained Language Models
Junyi Li
Tianyi Tang
Zheng Gong
Lixin Yang
Zhuohao Yu
Zhongfu Chen
Jingyuan Wang
Wayne Xin Zhao
Ji-Rong Wen
LM&MAELM
49
8
0
03 May 2022
OPT: Open Pre-trained Transformer Language Models
OPT: Open Pre-trained Transformer Language Models
Susan Zhang
Stephen Roller
Naman Goyal
Mikel Artetxe
Moya Chen
...
Daniel Simig
Punit Singh Koura
Anjali Sridhar
Tianlu Wang
Luke Zettlemoyer
VLMOSLMAI4CE
392
3,707
0
02 May 2022
BERTops: Studying BERT Representations under a Topological Lens
BERTops: Studying BERT Representations under a Topological Lens
Jatin Chauhan
Manohar Kaul
77
4
0
02 May 2022
Probing Cross-Lingual Lexical Knowledge from Multilingual Sentence
  Encoders
Probing Cross-Lingual Lexical Knowledge from Multilingual Sentence Encoders
Ivan Vulić
Goran Glavaš
Fangyu Liu
Nigel Collier
Edoardo Ponti
Anna Korhonen
96
9
0
30 Apr 2022
Prompt Consistency for Zero-Shot Task Generalization
Prompt Consistency for Zero-Shot Task Generalization
Chunting Zhou
Junxian He
Xuezhe Ma
Taylor Berg-Kirkpatrick
Graham Neubig
VLM
108
79
0
29 Apr 2022
Czech Dataset for Cross-lingual Subjectivity Classification
Czech Dataset for Cross-lingual Subjectivity Classification
Pavel Přibáň
J. Steinberger
74
5
0
29 Apr 2022
GRIT: General Robust Image Task Benchmark
GRIT: General Robust Image Task Benchmark
Tanmay Gupta
Ryan Marten
Aniruddha Kembhavi
Derek Hoiem
VLMOODObjD
75
33
0
28 Apr 2022
Super-Prompting: Utilizing Model-Independent Contextual Data to Reduce
  Data Annotation Required in Visual Commonsense Tasks
Super-Prompting: Utilizing Model-Independent Contextual Data to Reduce Data Annotation Required in Visual Commonsense Tasks
Navid Rezaei
Marek Reformat
VLM
48
2
0
25 Apr 2022
Exploring the Role of Task Transferability in Large-Scale Multi-Task
  Learning
Exploring the Role of Task Transferability in Large-Scale Multi-Task Learning
Vishakh Padmakumar
Leonard Lausen
Miguel Ballesteros
Sheng Zha
He He
George Karypis
104
20
0
23 Apr 2022
Grad-SAM: Explaining Transformers via Gradient Self-Attention Maps
Grad-SAM: Explaining Transformers via Gradient Self-Attention Maps
Oren Barkan
Edan Hauon
Avi Caciularu
Ori Katz
Itzik Malkiel
Omri Armstrong
Noam Koenigstein
88
40
0
23 Apr 2022
Hierarchical Label-wise Attention Transformer Model for Explainable ICD
  Coding
Hierarchical Label-wise Attention Transformer Model for Explainable ICD Coding
Leibo Liu
Óscar Pérez
Anthony N. Nguyen
Vicki Bennett
Louisa R Jorm
70
25
0
22 Apr 2022
Previous
123...192021...282930
Next