L4Q: Parameter Efficient Quantization-Aware Fine-Tuning on Large
  Language Models

L4Q: Parameter Efficient Quantization-Aware Fine-Tuning on Large Language Models

Papers citing "L4Q: Parameter Efficient Quantization-Aware Fine-Tuning on Large Language Models"

14 / 14 papers shown
Title
Rethinking Channel Dimensions to Isolate Outliers for Low-bit Weight Quantization of Large Language Models
Rethinking Channel Dimensions to Isolate Outliers for Low-bit Weight Quantization of Large Language Models
Jung Hwan Heo
Jeonghoon Kim
Beomseok Kwon
Byeongwook Kim
Se Jung Kwon
Dongsoo Lee
77
10
0
27 Sep 2023

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.