ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2408.00741
  4. Cited By
DynamoLLM: Designing LLM Inference Clusters for Performance and Energy
  Efficiency

DynamoLLM: Designing LLM Inference Clusters for Performance and Energy Efficiency

1 August 2024
Jovan Stojkovic
Chaojie Zhang
Íñigo Goiri
Josep Torrellas
Esha Choukse
ArXivPDFHTML

Papers citing "DynamoLLM: Designing LLM Inference Clusters for Performance and Energy Efficiency"

10 / 10 papers shown
Title
The ML.ENERGY Benchmark: Toward Automated Inference Energy Measurement and Optimization
The ML.ENERGY Benchmark: Toward Automated Inference Energy Measurement and Optimization
Jae-Won Chung
Jiachen Liu
Jeff J. Ma
Ruofan Wu
Oh Jun Kweon
Yuxuan Xia
Zhiyu Wu
Mosharaf Chowdhury
31
0
0
09 May 2025
CarbonCall: Sustainability-Aware Function Calling for Large Language Models on Edge Devices
CarbonCall: Sustainability-Aware Function Calling for Large Language Models on Edge Devices
Varatheepan Paramanayakam
Andreas Karatzas
Iraklis Anagnostopoulos
Dimitrios Stamoulis
34
0
0
29 Apr 2025
HyGen: Efficient LLM Serving via Elastic Online-Offline Request Co-location
HyGen: Efficient LLM Serving via Elastic Online-Offline Request Co-location
Ting Sun
Penghan Wang
Fan Lai
148
1
0
15 Jan 2025
iServe: An Intent-based Serving System for LLMs
iServe: An Intent-based Serving System for LLMs
Dimitrios Liakopoulos
Tianrui Hu
Prasoon Sinha
N. Yadwadkar
VLM
179
0
0
08 Jan 2025
FreeRide: Harvesting Bubbles in Pipeline Parallelism
FreeRide: Harvesting Bubbles in Pipeline Parallelism
Jiashu Zhang
Zihan Pan
Molly
Xu
Khuzaima S. Daudjee
90
0
0
11 Sep 2024
Efficient LLM Scheduling by Learning to Rank
Efficient LLM Scheduling by Learning to Rank
Yichao Fu
Siqi Zhu
Runlong Su
Aurick Qiao
Ion Stoica
Hao Zhang
55
19
0
28 Aug 2024
Efficient Interactive LLM Serving with Proxy Model-based Sequence Length
  Prediction
Efficient Interactive LLM Serving with Proxy Model-based Sequence Length Prediction
Haoran Qiu
Weichao Mao
Archit Patke
Shengkun Cui
Saurabh Jha
Chen Wang
Hubertus Franke
Zbigniew T. Kalbarczyk
Tamer Basar
Ravishankar K. Iyer
30
24
0
12 Apr 2024
LLM in a flash: Efficient Large Language Model Inference with Limited
  Memory
LLM in a flash: Efficient Large Language Model Inference with Limited Memory
Keivan Alizadeh-Vahid
Iman Mirzadeh
Dmitry Belenko
Karen Khatamifard
Minsik Cho
C. C. D. Mundo
Mohammad Rastegari
Mehrdad Farajtabar
77
112
0
12 Dec 2023
A Study of Generative Large Language Model for Medical Research and
  Healthcare
A Study of Generative Large Language Model for Medical Research and Healthcare
C.A.I. Peng
Xi Yang
Aokun Chen
Kaleb E. Smith
Nima M. Pournejatian
...
W. Hogan
E. Shenkman
Yi Guo
Jiang Bian
Yonghui Wu
LM&MA
ELM
AI4MH
155
244
0
22 May 2023
Making AI Less "Thirsty": Uncovering and Addressing the Secret Water Footprint of AI Models
Making AI Less "Thirsty": Uncovering and Addressing the Secret Water Footprint of AI Models
Pengfei Li
Jianyi Yang
M. A. Islam
Shaolei Ren
81
122
0
06 Apr 2023
1