ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1905.00537
  4. Cited By
SuperGLUE: A Stickier Benchmark for General-Purpose Language
  Understanding Systems
v1v2v3 (latest)

SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems

2 May 2019
Alex Jinpeng Wang
Yada Pruksachatkun
Nikita Nangia
Amanpreet Singh
Julian Michael
Felix Hill
Omer Levy
Samuel R. Bowman
    ELM
ArXiv (abs)PDFHTML

Papers citing "SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems"

50 / 1,500 papers shown
Title
Masked prediction tasks: a parameter identifiability view
Masked prediction tasks: a parameter identifiability view
Bingbin Liu
Daniel J. Hsu
Pradeep Ravikumar
Andrej Risteski
SSLOOD
67
4
0
18 Feb 2022
ST-MoE: Designing Stable and Transferable Sparse Expert Models
ST-MoE: Designing Stable and Transferable Sparse Expert Models
Barret Zoph
Irwan Bello
Sameer Kumar
Nan Du
Yanping Huang
J. Dean
Noam M. Shazeer
W. Fedus
MoE
218
205
0
17 Feb 2022
SGPT: GPT Sentence Embeddings for Semantic Search
SGPT: GPT Sentence Embeddings for Semantic Search
Niklas Muennighoff
RALM
166
190
0
17 Feb 2022
A Survey of Knowledge-Intensive NLP with Pre-Trained Language Models
A Survey of Knowledge-Intensive NLP with Pre-Trained Language Models
Da Yin
Li Dong
Hao Cheng
Xiaodong Liu
Kai-Wei Chang
Furu Wei
Jianfeng Gao
KELM
73
34
0
17 Feb 2022
An Equivalence Between Data Poisoning and Byzantine Gradient Attacks
An Equivalence Between Data Poisoning and Byzantine Gradient Attacks
Sadegh Farhadkhani
R. Guerraoui
L. Hoang
Oscar Villemaud
FedML
69
25
0
17 Feb 2022
Revisiting Parameter-Efficient Tuning: Are We Really There Yet?
Revisiting Parameter-Efficient Tuning: Are We Really There Yet?
Guanzheng Chen
Fangyu Liu
Zaiqiao Meng
Shangsong Liang
64
95
0
16 Feb 2022
Russian SuperGLUE 1.1: Revising the Lessons not Learned by Russian NLP
  models
Russian SuperGLUE 1.1: Revising the Lessons not Learned by Russian NLP models
Alena Fenogenova
Maria Tikhonova
Vladislav Mikhailov
Tatiana Shavrina
Anton A. Emelyanov
Denis Shevelev
Alexander Kukushkin
Valentin Malykh
Ekaterina Artemova
AAMLVLMELM
51
2
0
15 Feb 2022
MuLD: The Multitask Long Document Benchmark
MuLD: The Multitask Long Document Benchmark
G. Hudson
Noura Al Moubayed
105
11
0
15 Feb 2022
A Survey on Model Compression and Acceleration for Pretrained Language
  Models
A Survey on Model Compression and Acceleration for Pretrained Language Models
Canwen Xu
Julian McAuley
106
61
0
15 Feb 2022
Scaling Laws Under the Microscope: Predicting Transformer Performance
  from Small Scale Experiments
Scaling Laws Under the Microscope: Predicting Transformer Performance from Small Scale Experiments
Maor Ivgi
Y. Carmon
Jonathan Berant
80
17
0
13 Feb 2022
Slovene SuperGLUE Benchmark: Translation and Evaluation
Slovene SuperGLUE Benchmark: Translation and Evaluation
Aleš Žagar
Marko Robnik-Šikonja
58
10
0
10 Feb 2022
Generating Training Data with Language Models: Towards Zero-Shot
  Language Understanding
Generating Training Data with Language Models: Towards Zero-Shot Language Understanding
Yu Meng
Jiaxin Huang
Yu Zhang
Jiawei Han
SyDa
79
235
0
09 Feb 2022
TimeLMs: Diachronic Language Models from Twitter
TimeLMs: Diachronic Language Models from Twitter
Daniel Loureiro
Francesco Barbieri
Leonardo Neves
Luis Espinosa Anke
Jose Camacho-Collados
119
266
0
08 Feb 2022
What are the best systems? New perspectives on NLP Benchmarking
What are the best systems? New perspectives on NLP Benchmarking
Pierre Colombo
Nathan Noiry
Ekhine Irurozki
Stephan Clémençon
201
42
0
08 Feb 2022
Conversational Agents: Theory and Applications
Conversational Agents: Theory and Applications
M. Wahde
M. Virgolin
LLMAG
68
26
0
07 Feb 2022
GatorTron: A Large Clinical Language Model to Unlock Patient Information
  from Unstructured Electronic Health Records
GatorTron: A Large Clinical Language Model to Unlock Patient Information from Unstructured Electronic Health Records
Xi Yang
Aokun Chen
Nima M. Pournejatian
Hoo-Chang Shin
Kaleb E. Smith
...
Duane A. Mitchell
W. Hogan
E. Shenkman
Jiang Bian
Yonghui Wu
AI4MHLM&MA
102
551
0
02 Feb 2022
Co-training Improves Prompt-based Learning for Large Language Models
Co-training Improves Prompt-based Learning for Large Language Models
Hunter Lang
Monica Agrawal
Yoon Kim
David Sontag
VLMLRM
213
41
0
02 Feb 2022
Correcting diacritics and typos with a ByT5 transformer model
Correcting diacritics and typos with a ByT5 transformer model
Lukas Stankevicius
M. Lukoševičius
J. Kapočiūtė-Dzikienė
Monika Briediene
Tomas Krilavičius
62
21
0
31 Jan 2022
Protum: A New Method For Prompt Tuning Based on "[MASK]"
Protum: A New Method For Prompt Tuning Based on "[MASK]"
Pan He
Yuxi Chen
Yan Wang
Yanru Zhang
AAML
46
3
0
28 Jan 2022
IGLUE: A Benchmark for Transfer Learning across Modalities, Tasks, and
  Languages
IGLUE: A Benchmark for Transfer Learning across Modalities, Tasks, and Languages
Emanuele Bugliarello
Fangyu Liu
Jonas Pfeiffer
Siva Reddy
Desmond Elliott
Edoardo Ponti
Ivan Vulić
MLLMVLMELM
117
64
0
27 Jan 2022
An Application of Pseudo-Log-Likelihoods to Natural Language Scoring
An Application of Pseudo-Log-Likelihoods to Natural Language Scoring
Darren Abramson
Ali Emami
66
3
0
23 Jan 2022
Uncovering More Shallow Heuristics: Probing the Natural Language
  Inference Capacities of Transformer-Based Pre-Trained Language Models Using
  Syllogistic Patterns
Uncovering More Shallow Heuristics: Probing the Natural Language Inference Capacities of Transformer-Based Pre-Trained Language Models Using Syllogistic Patterns
Reto Gubelmann
Siegfried Handschuh
ReLMLRM
80
6
0
19 Jan 2022
Instance-aware Prompt Learning for Language Understanding and Generation
Instance-aware Prompt Learning for Language Understanding and Generation
Feihu Jin
Jinliang Lu
Jiajun Zhang
Chengqing Zong
57
33
0
18 Jan 2022
RuMedBench: A Russian Medical Language Understanding Benchmark
RuMedBench: A Russian Medical Language Understanding Benchmark
Pavel Blinov
A. Reshetnikova
A. Nesterov
Galina Zubkova
V. Kokh
ELM
72
15
0
17 Jan 2022
Efficient Data-Plane Memory Scheduling for In-Network Aggregation
Efficient Data-Plane Memory Scheduling for In-Network Aggregation
Hao Wang
Yifan Zhang
Chon-In Lao
Yanfang Le
Wenfei Wu
Kai Chen
38
16
0
17 Jan 2022
Ensemble Transformer for Efficient and Accurate Ranking Tasks: an
  Application to Question Answering Systems
Ensemble Transformer for Efficient and Accurate Ranking Tasks: an Application to Question Answering Systems
Yoshitomo Matsubara
Luca Soldaini
Eric Lind
Alessandro Moschitti
89
7
0
15 Jan 2022
DeepSpeed-MoE: Advancing Mixture-of-Experts Inference and Training to
  Power Next-Generation AI Scale
DeepSpeed-MoE: Advancing Mixture-of-Experts Inference and Training to Power Next-Generation AI Scale
Samyam Rajbhandari
Conglong Li
Z. Yao
Minjia Zhang
Reza Yazdani Aminabadi
A. A. Awan
Jeff Rasley
Yuxiong He
135
309
0
14 Jan 2022
The Dark Side of the Language: Pre-trained Transformers in the DarkNet
The Dark Side of the Language: Pre-trained Transformers in the DarkNet
Leonardo Ranaldi
Aria Nourbakhsh
Arianna Patrizi
Elena Sofia Ruzzetti
Dario Onorati
Francesca Fallucchi
Fabio Massimo Zanzotto
VLM
63
21
0
14 Jan 2022
Pretrained Language Models for Text Generation: A Survey
Pretrained Language Models for Text Generation: A Survey
Junyi Li
Tianyi Tang
Wayne Xin Zhao
J. Nie
Ji-Rong Wen
AI4CE
173
151
0
14 Jan 2022
How Does Data Corruption Affect Natural Language Understanding Models? A
  Study on GLUE datasets
How Does Data Corruption Affect Natural Language Understanding Models? A Study on GLUE datasets
Aarne Talman
Marianna Apidianaki
S. Chatzikyriakidis
Jörg Tiedemann
ELM
57
0
0
12 Jan 2022
SCROLLS: Standardized CompaRison Over Long Language Sequences
SCROLLS: Standardized CompaRison Over Long Language Sequences
Uri Shaham
Elad Segal
Maor Ivgi
Avia Efrat
Ori Yoran
...
Ankit Gupta
Wenhan Xiong
Mor Geva
Jonathan Berant
Omer Levy
RALM
125
139
0
10 Jan 2022
Does QA-based intermediate training help fine-tuning language models for
  text classification?
Does QA-based intermediate training help fine-tuning language models for text classification?
Shiwei Zhang
Xiuzhen Zhang
49
2
0
30 Dec 2021
CUGE: A Chinese Language Understanding and Generation Evaluation
  Benchmark
CUGE: A Chinese Language Understanding and Generation Evaluation Benchmark
Yuan Yao
Qingxiu Dong
Jian Guan
Boxi Cao
Zhengyan Zhang
...
Zhiyuan Liu
Xianpei Han
Erhong Yang
Zhifang Sui
Maosong Sun
ALMELM
76
17
0
27 Dec 2021
CRASS: A Novel Data Set and Benchmark to Test Counterfactual Reasoning
  of Large Language Models
CRASS: A Novel Data Set and Benchmark to Test Counterfactual Reasoning of Large Language Models
Jorg Frohberg
Frank Binder
SLR
126
30
0
22 Dec 2021
Few-shot Learning with Multilingual Language Models
Few-shot Learning with Multilingual Language Models
Xi Lin
Todor Mihaylov
Mikel Artetxe
Tianlu Wang
Shuohui Chen
...
Luke Zettlemoyer
Zornitsa Kozareva
Mona T. Diab
Ves Stoyanov
Xian Li
BDLELMLRM
153
308
0
20 Dec 2021
Pushing the Limits of Rule Reasoning in Transformers through Natural
  Language Satisfiability
Pushing the Limits of Rule Reasoning in Transformers through Natural Language Satisfiability
Kyle Richardson
Ashish Sabharwal
ReLMLRM
85
25
0
16 Dec 2021
Know Thy Strengths: Comprehensive Dialogue State Tracking Diagnostics
Know Thy Strengths: Comprehensive Dialogue State Tracking Diagnostics
Hyundong Justin Cho
Chinnadhurai Sankar
Christopher Lin
Kaushik Ram Sadagopan
Shahin Shayandeh
Asli Celikyilmaz
Jonathan May
Ahmad Beirami
115
10
0
15 Dec 2021
Fine-Tuning Large Neural Language Models for Biomedical Natural Language
  Processing
Fine-Tuning Large Neural Language Models for Biomedical Natural Language Processing
Robert Tinn
Hao Cheng
Yu Gu
Naoto Usuyama
Xiaodong Liu
Tristan Naumann
Jianfeng Gao
Hoifung Poon
LM&MA
60
117
0
15 Dec 2021
LMTurk: Few-Shot Learners as Crowdsourcing Workers in a
  Language-Model-as-a-Service Framework
LMTurk: Few-Shot Learners as Crowdsourcing Workers in a Language-Model-as-a-Service Framework
Mengjie Zhao
Fei Mi
Yasheng Wang
Minglei Li
Xin Jiang
Qun Liu
Hinrich Schütze
RALM
115
11
0
14 Dec 2021
GLaM: Efficient Scaling of Language Models with Mixture-of-Experts
GLaM: Efficient Scaling of Language Models with Mixture-of-Experts
Nan Du
Yanping Huang
Andrew M. Dai
Simon Tong
Dmitry Lepikhin
...
Kun Zhang
Quoc V. Le
Yonghui Wu
Zhiwen Chen
Claire Cui
ALMMoE
269
832
0
13 Dec 2021
Measuring Context-Word Biases in Lexical Semantic Datasets
Measuring Context-Word Biases in Lexical Semantic Datasets
Qianchu Liu
Diana McCarthy
Anna Korhonen
71
2
0
13 Dec 2021
Technical Language Supervision for Intelligent Fault Diagnosis in
  Process Industry
Technical Language Supervision for Intelligent Fault Diagnosis in Process Industry
Karl Lowenmark
C. Taal
S. Schnabel
Marcus Liwicki
Fredrik Sandin
45
7
0
11 Dec 2021
Analysis and Prediction of NLP Models Via Task Embeddings
Analysis and Prediction of NLP Models Via Task Embeddings
Damien Sileo
Marie-Francine Moens
55
4
0
10 Dec 2021
FLAVA: A Foundational Language And Vision Alignment Model
FLAVA: A Foundational Language And Vision Alignment Model
Amanpreet Singh
Ronghang Hu
Vedanuj Goswami
Guillaume Couairon
Wojciech Galuba
Marcus Rohrbach
Douwe Kiela
CLIPVLM
133
719
0
08 Dec 2021
Ethical and social risks of harm from Language Models
Ethical and social risks of harm from Language Models
Laura Weidinger
John F. J. Mellor
Maribeth Rauh
Conor Griffin
J. Uesato
...
Lisa Anne Hendricks
William S. Isaac
Sean Legassick
G. Irving
Iason Gabriel
PILM
191
1,045
0
08 Dec 2021
JABER and SABER: Junior and Senior Arabic BERt
JABER and SABER: Junior and Senior Arabic BERt
Abbas Ghaddar
Yimeng Wu
Ahmad Rashid
Khalil Bibi
Mehdi Rezagholizadeh
...
Zhefeng Wang
Baoxing Huai
Xin Jiang
Qun Liu
Philippe Langlais
55
5
0
08 Dec 2021
Quantifying Adaptability in Pre-trained Language Models with 500 Tasks
Quantifying Adaptability in Pre-trained Language Models with 500 Tasks
Belinda Z. Li
Jane A. Yu
Madian Khabsa
Luke Zettlemoyer
A. Halevy
Jacob Andreas
ELM
89
17
0
06 Dec 2021
How not to Lie with a Benchmark: Rearranging NLP Leaderboards
How not to Lie with a Benchmark: Rearranging NLP Leaderboards
Tatiana Shavrina
Valentin Malykh
ALMELM
505
12
0
02 Dec 2021
Towards More Robust Natural Language Understanding
Towards More Robust Natural Language Understanding
Xinliang Frederick Zhang
65
2
0
01 Dec 2021
Dyna-bAbI: unlocking bAbI's potential with dynamic synthetic
  benchmarking
Dyna-bAbI: unlocking bAbI's potential with dynamic synthetic benchmarking
Ronen Tamari
Kyle Richardson
Aviad Sar-Shalom
Noam Kahlon
Nelson F. Liu
Reut Tsarfaty
Dafna Shahaf
115
5
0
30 Nov 2021
Previous
123...212223...282930
Next