ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2501.12162
  4. Cited By
AdaServe: Accelerating Multi-SLO LLM Serving with SLO-Customized Speculative Decoding
v1v2 (latest)

AdaServe: Accelerating Multi-SLO LLM Serving with SLO-Customized Speculative Decoding

21 January 2025
Zikun Li
Zhuofu Chen
Remi Delacourt
Gabriele Oliaro
Zeyu Wang
Qinghan Chen
Shuhuai Lin
April Yang
Zhihao Zhang
Zhuoming Chen
Sean Lai
Xinhao Cheng
Xupeng Miao
Zhihao Jia
ArXiv (abs)PDFHTML

Papers citing "AdaServe: Accelerating Multi-SLO LLM Serving with SLO-Customized Speculative Decoding"

2 / 2 papers shown
Title
SLED: A Speculative LLM Decoding Framework for Efficient Edge Serving
SLED: A Speculative LLM Decoding Framework for Efficient Edge Serving
Xiangchen Li
Dimitrios Spatharakis
Saeid Ghafouri
Jiakun Fan
Dimitrios Nikolopoulos
Deepu John
Bo Ji
Dimitrios S. Nikolopoulos
52
0
0
11 Jun 2025
MagicDec: Breaking the Latency-Throughput Tradeoff for Long Context Generation with Speculative Decoding
MagicDec: Breaking the Latency-Throughput Tradeoff for Long Context Generation with Speculative Decoding
Jian Chen
Vashisth Tiwari
Ranajoy Sadhukhan
Zhuoming Chen
Jinyuan Shi
Ian En-Hsu Yen
Ian En-Hsu Yen
Avner May
Tianqi Chen
Beidi Chen
LRM
154
32
0
20 Aug 2024
1