DAQ: Density-Aware Post-Training Weight-Only Quantization For LLMs
v1v2 (latest)

DAQ: Density-Aware Post-Training Weight-Only Quantization For LLMs

    MQ

Papers citing "DAQ: Density-Aware Post-Training Weight-Only Quantization For LLMs"

18 / 18 papers shown
Title
The case for 4-bit precision: k-bit Inference Scaling Laws
The case for 4-bit precision: k-bit Inference Scaling Laws
Tim Dettmers
Luke Zettlemoyer
95
234
0
19 Dec 2022

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.