ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2412.15285
  4. Cited By
Maximize Your Data's Potential: Enhancing LLM Accuracy with Two-Phase
  Pretraining

Maximize Your Data's Potential: Enhancing LLM Accuracy with Two-Phase Pretraining

18 December 2024
Steven Feng
Shrimai Prabhumoye
John Kamalu
Jane Polak Scowcroft
M. Patwary
Mohammad Shoeybi
Bryan Catanzaro
ArXiv (abs)PDFHTML

Papers citing "Maximize Your Data's Potential: Enhancing LLM Accuracy with Two-Phase Pretraining"

4 / 4 papers shown
Title
Curriculum-Guided Layer Scaling for Language Model Pretraining
Curriculum-Guided Layer Scaling for Language Model Pretraining
Karanpartap Singh
Neil Band
Ehsan Adeli
ALMLRM
41
0
0
13 Jun 2025
Chameleon: A Flexible Data-mixing Framework for Language Model Pretraining and Finetuning
Chameleon: A Flexible Data-mixing Framework for Language Model Pretraining and Finetuning
Wanyun Xie
F. Tonin
Volkan Cevher
36
0
0
30 May 2025
ChemPile: A 250GB Diverse and Curated Dataset for Chemical Foundation Models
ChemPile: A 250GB Diverse and Curated Dataset for Chemical Foundation Models
Adrian Mirza
Nawaf Alampara
Martiño Ríos-García
Mohamed Abdelalim
Jack Butler
...
Mark Worrall
Adamo Young
Philippe Schwaller
Michael Pieler
Kevin Maik Jablonka
149
0
0
18 May 2025
MIND: Math Informed syNthetic Dialogues for Pretraining LLMs
MIND: Math Informed syNthetic Dialogues for Pretraining LLMs
Syeda Nahida Akter
Shrimai Prabhumoye
John Kamalu
S. Satheesh
Eric Nyberg
M. Patwary
Mohammad Shoeybi
Bryan Catanzaro
LRMSyDaReLM
167
2
0
15 Oct 2024
1