Quantizable Transformers: Removing Outliers by Helping Attention Heads
  Do Nothing
v1v2 (latest)

Quantizable Transformers: Removing Outliers by Helping Attention Heads Do Nothing

    MQ

Papers citing "Quantizable Transformers: Removing Outliers by Helping Attention Heads Do Nothing"

19 / 19 papers shown
Title
Rethinking Channel Dimensions to Isolate Outliers for Low-bit Weight Quantization of Large Language Models
Rethinking Channel Dimensions to Isolate Outliers for Low-bit Weight Quantization of Large Language Models
Jung Hwan Heo
Jeonghoon Kim
Beomseok Kwon
Byeongwook Kim
Se Jung Kwon
Dongsoo Lee
129
10
0
27 Sep 2023

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.