ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2504.06261
  4. Cited By
Hogwild! Inference: Parallel LLM Generation via Concurrent Attention
v1v2v3 (latest)

Hogwild! Inference: Parallel LLM Generation via Concurrent Attention

8 April 2025
Gleb Rodionov
Roman Garipov
Alina Shutova
George Yakushev
Erik Schultheis
Vage Egiazarian
Anton Sinitsin
Denis Kuznedelev
Dan Alistarh
    LRM
ArXiv (abs)PDFHTML

Papers citing "Hogwild! Inference: Parallel LLM Generation via Concurrent Attention"

10 / 10 papers shown
Title
Multiverse: Your Language Models Secretly Decide How to Parallelize and Merge Generation
Multiverse: Your Language Models Secretly Decide How to Parallelize and Merge Generation
Xinyu Yang
Yuwei An
Hongyi Liu
Tianqi Chen
Beidi Chen
SyDaLRM
122
0
0
11 Jun 2025
Get Experience from Practice: LLM Agents with Record & Replay
Get Experience from Practice: LLM Agents with Record & Replay
Erhu Feng
Wenbo Zhou
Zibin Liu
Le Chen
Yunpeng Dong
...
Yisheng Zhao
Dong Du
Zhichao Hua
Yubin Xia
Haibo Chen
117
0
0
23 May 2025
Group Think: Multiple Concurrent Reasoning Agents Collaborating at Token Level Granularity
Group Think: Multiple Concurrent Reasoning Agents Collaborating at Token Level Granularity
Chan-Jan Hsu
Davide Buffelli
Jamie McGowan
Feng-Ting Liao
Yi-Chang Chen
Sattar Vakili
Da-shan Shiu
LLMAGLRM
106
1
0
16 May 2025
Parallel Scaling Law for Language Models
Parallel Scaling Law for Language Models
Mouxiang Chen
Binyuan Hui
Zeyu Cui
Jiaxi Yang
Dayiheng Liu
Jianling Sun
Junyang Lin
Zhongxin Liu
MoELRM
91
2
0
15 May 2025
THOUGHTTERMINATOR: Benchmarking, Calibrating, and Mitigating Overthinking in Reasoning Models
THOUGHTTERMINATOR: Benchmarking, Calibrating, and Mitigating Overthinking in Reasoning Models
Xiao Pu
Michael Stephen Saxon
Wenyue Hua
William Y. Wang
LRM
103
7
0
17 Apr 2025
Accelerate Parallelizable Reasoning via Parallel Decoding within One Sequence
Accelerate Parallelizable Reasoning via Parallel Decoding within One Sequence
Yijiong Yu
LRMAIMat
171
1
0
26 Mar 2025
Learning to Keep a Promise: Scaling Language Model Decoding Parallelism with Learned Asynchronous Decoding
Learning to Keep a Promise: Scaling Language Model Decoding Parallelism with Learned Asynchronous Decoding
Tian Jin
Ellie Y. Cheng
Zack Ankner
Nikunj Saunshi
Blake M. Elias
Amir Yazdanbakhsh
Jonathan Ragan-Kelley
Suvinay Subramanian
Michael Carbin
135
5
0
24 Feb 2025
Dynamic Parallel Tree Search for Efficient LLM Reasoning
Dynamic Parallel Tree Search for Efficient LLM Reasoning
Yifu Ding
Wentao Jiang
Shunyu Liu
Yongcheng Jing
Jinpei Guo
...
Zengmao Wang
Ziqiang Liu
Di Lin
Xianglong Liu
Dacheng Tao
LRM
119
11
0
22 Feb 2025
LIMO: Less is More for Reasoning
LIMO: Less is More for Reasoning
Yixin Ye
Zhen Huang
Yang Xiao
Ethan Chern
Shijie Xia
Pengfei Liu
LRM
168
166
0
05 Feb 2025
EAGLE: Speculative Sampling Requires Rethinking Feature Uncertainty
EAGLE: Speculative Sampling Requires Rethinking Feature Uncertainty
Yuhui Li
Fangyun Wei
Chao Zhang
Hongyang R. Zhang
150
165
0
26 Jan 2024
1