Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2404.10933
Cited By
LLMem: Estimating GPU Memory Usage for Fine-Tuning Pre-Trained LLMs
16 April 2024
Taeho Kim
Yanming Wang
Vatshank Chaturvedi
Lokesh Gupta
Seyeon Kim
Yongin Kwon
Sangtae Ha
Re-assign community
ArXiv
PDF
HTML
Papers citing
"LLMem: Estimating GPU Memory Usage for Fine-Tuning Pre-Trained LLMs"
5 / 5 papers shown
Title
Accurate GPU Memory Prediction for Deep Learning Jobs through Dynamic Analysis
Jiabo Shi
Yehia Elkhatib
3DH
VLM
30
0
0
04 Apr 2025
KD-LoRA: A Hybrid Approach to Efficient Fine-Tuning with LoRA and Knowledge Distillation
Rambod Azimi
Rishav Rishav
M. Teichmann
Samira Ebrahimi Kahou
ALM
28
0
0
28 Oct 2024
ZeRO-Offload: Democratizing Billion-Scale Model Training
Jie Ren
Samyam Rajbhandari
Reza Yazdani Aminabadi
Olatunji Ruwase
Shuangyang Yang
Minjia Zhang
Dong Li
Yuxiong He
MoE
177
414
0
18 Jan 2021
Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
M. Shoeybi
M. Patwary
Raul Puri
P. LeGresley
Jared Casper
Bryan Catanzaro
MoE
245
1,821
0
17 Sep 2019
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding
Alex Jinpeng Wang
Amanpreet Singh
Julian Michael
Felix Hill
Omer Levy
Samuel R. Bowman
ELM
297
6,959
0
20 Apr 2018
1