
v1v2 (latest)
Quantizable Transformers: Removing Outliers by Helping Attention Heads Do Nothing
Papers citing "Quantizable Transformers: Removing Outliers by Helping Attention Heads Do Nothing"
19 / 19 papers shown
Title |
---|
![]() Rethinking Channel Dimensions to Isolate Outliers for Low-bit Weight Quantization of Large Language Models Jung Hwan Heo Jeonghoon Kim Beomseok Kwon Byeongwook Kim Se Jung Kwon Dongsoo Lee |