Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2105.04876
Cited By
Benchmarking down-scaled (not so large) pre-trained language models
11 May 2021
Matthias Aßenmacher
P. Schulze
C. Heumann
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Benchmarking down-scaled (not so large) pre-trained language models"
4 / 4 papers shown
Title
Drawing Causal Inferences About Performance Effects in NLP
Sandra Wankmüller
CML
16
1
0
14 Sep 2022
Scaling Laws for Neural Language Models
Jared Kaplan
Sam McCandlish
T. Henighan
Tom B. Brown
B. Chess
R. Child
Scott Gray
Alec Radford
Jeff Wu
Dario Amodei
264
4,489
0
23 Jan 2020
Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
M. Shoeybi
M. Patwary
Raul Puri
P. LeGresley
Jared Casper
Bryan Catanzaro
MoE
245
1,826
0
17 Sep 2019
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding
Alex Jinpeng Wang
Amanpreet Singh
Julian Michael
Felix Hill
Omer Levy
Samuel R. Bowman
ELM
299
6,984
0
20 Apr 2018
1