Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2503.16672
Cited By
Accelerating Transformer Inference and Training with 2:4 Activation Sparsity
20 March 2025
Daniel Haziza
Timothy Chou
Dhruv Choudhary
Luca Wehrstedt
Francisco Massa
Jiecao Yu
Geonhwa Jeong
Supriya Rao
Patrick Labatut
Jesse Cai
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Accelerating Transformer Inference and Training with 2:4 Activation Sparsity"
3 / 3 papers shown
Title
SLoPe: Double-Pruned Sparse Plus Lazy Low-Rank Adapter Pretraining of LLMs
Mohammad Mozaffari
Amir Yazdanbakhsh
Zhao Zhang
M. Dehnavi
140
7
0
28 Jan 2025
Nemotron-4 340B Technical Report
Nvidia
:
Bo Adler
Niket Agarwal
Ashwath Aithal
...
Jimmy Zhang
Jing Zhang
Vivienne Zhang
Yian Zhang
Chen Zhu
117
69
0
17 Jun 2024
Primer: Searching for Efficient Transformers for Language Modeling
David R. So
Wojciech Mañke
Hanxiao Liu
Zihang Dai
Noam M. Shazeer
Quoc V. Le
VLM
264
156
0
17 Sep 2021
1