ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2203.17189
  4. Cited By
Scaling Up Models and Data with $\texttt{t5x}$ and $\texttt{seqio}$

Scaling Up Models and Data with t5x\texttt{t5x}t5x and seqio\texttt{seqio}seqio

31 March 2022
Adam Roberts
Hyung Won Chung
Anselm Levskaya
Gaurav Mishra
James Bradbury
D. Andor
Sharan Narang
Brian Lester
Colin Gaffney
Afroz Mohiuddin
Curtis Hawthorne
Aitor Lewkowycz
Alexandru Salcianu
Marc van Zee
Jacob Austin
Sebastian Goodman
Livio Baldini Soares
Haitang Hu
Sasha Tsvyashchenko
Aakanksha Chowdhery
Jasmijn Bastings
Jannis Bulian
Xavier Garcia
Jianmo Ni
A. Chen
Kathleen Kenealy
J. Clark
Stephan Lee
Daniel H Garrette
James Lee-Thorp
Colin Raffel
Noam M. Shazeer
Marvin Ritter
Maarten Bosma
Alexandre Passos
Jeremy B. Maitin-Shepard
Noah Fiedel
Mark Omernick
Brennan Saeta
Ryan Sepassi
A. Spiridonov
Joshua Newlan
Andrea Gesmundo
    ALM
ArXivPDFHTML

Papers citing "Scaling Up Models and Data with $\texttt{t5x}$ and $\texttt{seqio}$"

7 / 57 papers shown
Title
TALM: Tool Augmented Language Models
TALM: Tool Augmented Language Models
Aaron T Parisi
Yao-Min Zhao
Noah Fiedel
KELM
RALM
LLMAG
32
144
0
24 May 2022
RankGen: Improving Text Generation with Large Ranking Models
RankGen: Improving Text Generation with Large Ranking Models
Kalpesh Krishna
Yapei Chang
John Wieting
Mohit Iyyer
AIMat
24
68
0
19 May 2022
UL2: Unifying Language Learning Paradigms
UL2: Unifying Language Learning Paradigms
Yi Tay
Mostafa Dehghani
Vinh Q. Tran
Xavier Garcia
Jason W. Wei
...
Tal Schuster
H. Zheng
Denny Zhou
N. Houlsby
Donald Metzler
AI4CE
59
297
0
10 May 2022
GPT-NeoX-20B: An Open-Source Autoregressive Language Model
GPT-NeoX-20B: An Open-Source Autoregressive Language Model
Sid Black
Stella Biderman
Eric Hallahan
Quentin G. Anthony
Leo Gao
...
Shivanshu Purohit
Laria Reynolds
J. Tow
Benqi Wang
Samuel Weinbach
99
802
0
14 Apr 2022
What Language Model Architecture and Pretraining Objective Work Best for
  Zero-Shot Generalization?
What Language Model Architecture and Pretraining Objective Work Best for Zero-Shot Generalization?
Thomas Wang
Adam Roberts
Daniel Hesslow
Teven Le Scao
Hyung Won Chung
Iz Beltagy
Julien Launay
Colin Raffel
33
167
0
12 Apr 2022
Deduplicating Training Data Makes Language Models Better
Deduplicating Training Data Makes Language Models Better
Katherine Lee
Daphne Ippolito
A. Nystrom
Chiyuan Zhang
Douglas Eck
Chris Callison-Burch
Nicholas Carlini
SyDa
242
593
0
14 Jul 2021
Megatron-LM: Training Multi-Billion Parameter Language Models Using
  Model Parallelism
Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
M. Shoeybi
M. Patwary
Raul Puri
P. LeGresley
Jared Casper
Bryan Catanzaro
MoE
245
1,826
0
17 Sep 2019
Previous
12