ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1802.08635
  4. Cited By
Loss-aware Weight Quantization of Deep Networks

Loss-aware Weight Quantization of Deep Networks

23 February 2018
Lu Hou
James T. Kwok
    MQ
ArXivPDFHTML

Papers citing "Loss-aware Weight Quantization of Deep Networks"

50 / 69 papers shown
Title
Radio: Rate-Distortion Optimization for Large Language Model Compression
Radio: Rate-Distortion Optimization for Large Language Model Compression
Sean I. Young
MQ
26
0
0
05 May 2025
Foundations of Large Language Model Compression -- Part 1: Weight
  Quantization
Foundations of Large Language Model Compression -- Part 1: Weight Quantization
Sean I. Young
MQ
50
1
0
03 Sep 2024
The Binary Quantized Neural Network for Dense Prediction via Specially
  Designed Upsampling and Attention
The Binary Quantized Neural Network for Dense Prediction via Specially Designed Upsampling and Attention
Xingyu Ding
Li Shan
Guiqin Zhao
Meiqi Wu
Wenzhang Zhou
Wei Li
MQ
39
0
0
28 May 2024
One-Step Forward and Backtrack: Overcoming Zig-Zagging in Loss-Aware
  Quantization Training
One-Step Forward and Backtrack: Overcoming Zig-Zagging in Loss-Aware Quantization Training
Lianbo Ma
Yuee Zhou
Jianlun Ma
Guo-Ding Yu
Qing Li
MQ
25
1
0
30 Jan 2024
Overcoming Distribution Mismatch in Quantizing Image Super-Resolution
  Networks
Overcoming Distribution Mismatch in Quantizing Image Super-Resolution Networks
Chee Hong
Kyoung Mu Lee
SupR
MQ
27
1
0
25 Jul 2023
Learning to Learn with Indispensable Connections
Learning to Learn with Indispensable Connections
Sambhavi Tiwari
Manas Gogoi
Shekhar Verma
Krishna Pratap Singh
CLL
23
0
0
06 Apr 2023
Ternary Quantization: A Survey
Ternary Quantization: A Survey
Danyang Liu
Xue Liu
MQ
26
3
0
02 Mar 2023
Hyperspherical Loss-Aware Ternary Quantization
Hyperspherical Loss-Aware Ternary Quantization
Dan Liu
Xue Liu
MQ
27
0
0
24 Dec 2022
Adaptive Low-Precision Training for Embeddings in Click-Through Rate
  Prediction
Adaptive Low-Precision Training for Embeddings in Click-Through Rate Prediction
Shiwei Li
Huifeng Guo
Luyao Hou
Wei Zhang
Xing Tang
Ruiming Tang
Rui Zhang
Rui Li
MQ
129
9
0
12 Dec 2022
Class-based Quantization for Neural Networks
Class-based Quantization for Neural Networks
Wenhao Sun
Grace Li Zhang
Huaxi Gu
Bing Li
Ulf Schlichtmann
MQ
24
7
0
27 Nov 2022
AskewSGD : An Annealed interval-constrained Optimisation method to train
  Quantized Neural Networks
AskewSGD : An Annealed interval-constrained Optimisation method to train Quantized Neural Networks
Louis Leconte
S. Schechtman
Eric Moulines
29
4
0
07 Nov 2022
A Closer Look at Hardware-Friendly Weight Quantization
A Closer Look at Hardware-Friendly Weight Quantization
Sungmin Bae
Piotr Zielinski
S. Chatterjee
MQ
32
0
0
07 Oct 2022
CADyQ: Content-Aware Dynamic Quantization for Image Super-Resolution
CADyQ: Content-Aware Dynamic Quantization for Image Super-Resolution
Chee Hong
Sungyong Baik
Heewon Kim
Seungjun Nah
Kyoung Mu Lee
SupR
MQ
31
32
0
21 Jul 2022
QReg: On Regularization Effects of Quantization
QReg: On Regularization Effects of Quantization
Mohammadhossein Askarihemmat
Reyhane Askari Hemmat
Alexander Hoffman
Ivan Lazarevich
Ehsan Saboori
Olivier Mastropietro
Sudhakar Sah
Yvon Savaria
J. David
MQ
37
5
0
24 Jun 2022
Exploring Extreme Parameter Compression for Pre-trained Language Models
Exploring Extreme Parameter Compression for Pre-trained Language Models
Yuxin Ren
Benyou Wang
Lifeng Shang
Xin Jiang
Qun Liu
33
18
0
20 May 2022
FxP-QNet: A Post-Training Quantizer for the Design of Mixed
  Low-Precision DNNs with Dynamic Fixed-Point Representation
FxP-QNet: A Post-Training Quantizer for the Design of Mixed Low-Precision DNNs with Dynamic Fixed-Point Representation
Ahmad Shawahna
S. M. Sait
A. El-Maleh
Irfan Ahmad
MQ
20
6
0
22 Mar 2022
DQ-BART: Efficient Sequence-to-Sequence Model via Joint Distillation and
  Quantization
DQ-BART: Efficient Sequence-to-Sequence Model via Joint Distillation and Quantization
Zheng Li
Zijian Wang
Ming Tan
Ramesh Nallapati
Parminder Bhatia
Andrew O. Arnold
Bing Xiang
Dan Roth
MQ
26
40
0
21 Mar 2022
Compression of Generative Pre-trained Language Models via Quantization
Compression of Generative Pre-trained Language Models via Quantization
Chaofan Tao
Lu Hou
Wei Zhang
Lifeng Shang
Xin Jiang
Qun Liu
Ping Luo
Ngai Wong
MQ
38
103
0
21 Mar 2022
Sharpness-aware Quantization for Deep Neural Networks
Sharpness-aware Quantization for Deep Neural Networks
Jing Liu
Jianfei Cai
Bohan Zhuang
MQ
27
24
0
24 Nov 2021
Towards Efficient Post-training Quantization of Pre-trained Language
  Models
Towards Efficient Post-training Quantization of Pre-trained Language Models
Haoli Bai
Lu Hou
Lifeng Shang
Xin Jiang
Irwin King
M. Lyu
MQ
79
47
0
30 Sep 2021
Training Spiking Neural Networks Using Lessons From Deep Learning
Training Spiking Neural Networks Using Lessons From Deep Learning
Jason Eshraghian
Max Ward
Emre Neftci
Xinxin Wang
Gregor Lenz
Girish Dwivedi
Bennamoun
Doo Seok Jeong
Wei D. Lu
40
432
0
27 Sep 2021
Distance-aware Quantization
Distance-aware Quantization
Dohyung Kim
Junghyup Lee
Bumsub Ham
MQ
15
28
0
16 Aug 2021
Adaptive Precision Training (AdaPT): A dynamic fixed point quantized
  training approach for DNNs
Adaptive Precision Training (AdaPT): A dynamic fixed point quantized training approach for DNNs
Lorenz Kummer
Kevin Sidak
Tabea Reichmann
Wilfried Gansterer
MQ
27
5
0
28 Jul 2021
Trustworthy AI: A Computational Perspective
Trustworthy AI: A Computational Perspective
Haochen Liu
Yiqi Wang
Wenqi Fan
Xiaorui Liu
Yaxin Li
Shaili Jain
Yunhao Liu
Anil K. Jain
Jiliang Tang
FaML
104
196
0
12 Jul 2021
NUQSGD: Provably Communication-efficient Data-parallel SGD via Nonuniform Quantization
Ali Ramezani-Kebrya
Fartash Faghri
Ilya Markov
V. Aksenov
Dan Alistarh
Daniel M. Roy
MQ
65
30
0
28 Apr 2021
Random and Adversarial Bit Error Robustness: Energy-Efficient and Secure
  DNN Accelerators
Random and Adversarial Bit Error Robustness: Energy-Efficient and Secure DNN Accelerators
David Stutz
Nandhini Chandramoorthy
Matthias Hein
Bernt Schiele
AAML
MQ
24
18
0
16 Apr 2021
BinaryBERT: Pushing the Limit of BERT Quantization
BinaryBERT: Pushing the Limit of BERT Quantization
Haoli Bai
Wei Zhang
Lu Hou
Lifeng Shang
Jing Jin
Xin Jiang
Qun Liu
Michael Lyu
Irwin King
MQ
142
221
0
31 Dec 2020
Direct Quantization for Training Highly Accurate Low Bit-width Deep
  Neural Networks
Direct Quantization for Training Highly Accurate Low Bit-width Deep Neural Networks
Ziquan Liu
Wuguannan Yao
Qiao Li
Antoni B. Chan
MQ
24
9
0
26 Dec 2020
DAQ: Channel-Wise Distribution-Aware Quantization for Deep Image
  Super-Resolution Networks
DAQ: Channel-Wise Distribution-Aware Quantization for Deep Image Super-Resolution Networks
Chee Hong
Heewon Kim
Sungyong Baik
Junghun Oh
Kyoung Mu Lee
OOD
SupR
MQ
24
41
0
21 Dec 2020
Learning Quantized Neural Nets by Coarse Gradient Method for Non-linear
  Classification
Learning Quantized Neural Nets by Coarse Gradient Method for Non-linear Classification
Ziang Long
Penghang Yin
Jack Xin
MQ
32
3
0
23 Nov 2020
Stochastic Precision Ensemble: Self-Knowledge Distillation for Quantized
  Deep Neural Networks
Stochastic Precision Ensemble: Self-Knowledge Distillation for Quantized Deep Neural Networks
Yoonho Boo
Sungho Shin
Jungwook Choi
Wonyong Sung
MQ
30
29
0
30 Sep 2020
BAMSProd: A Step towards Generalizing the Adaptive Optimization Methods
  to Deep Binary Model
BAMSProd: A Step towards Generalizing the Adaptive Optimization Methods to Deep Binary Model
Junjie Liu
Dongchao Wen
Deyu Wang
Wei Tao
Tse-Wei Chen
Kinya Osa
Masami Kato
MQ
29
1
0
29 Sep 2020
TernaryBERT: Distillation-aware Ultra-low Bit BERT
TernaryBERT: Distillation-aware Ultra-low Bit BERT
Wei Zhang
Lu Hou
Yichun Yin
Lifeng Shang
Xiao Chen
Xin Jiang
Qun Liu
MQ
33
208
0
27 Sep 2020
One Weight Bitwidth to Rule Them All
One Weight Bitwidth to Rule Them All
Ting-Wu Chin
P. Chuang
Vikas Chandra
Diana Marculescu
MQ
28
25
0
22 Aug 2020
Compression of Deep Learning Models for Text: A Survey
Compression of Deep Learning Models for Text: A Survey
Manish Gupta
Puneet Agrawal
VLM
MedIm
AI4CE
17
115
0
12 Aug 2020
FATNN: Fast and Accurate Ternary Neural Networks
FATNN: Fast and Accurate Ternary Neural Networks
Peng Chen
Bohan Zhuang
Chunhua Shen
MQ
6
15
0
12 Aug 2020
DBQ: A Differentiable Branch Quantizer for Lightweight Deep Neural
  Networks
DBQ: A Differentiable Branch Quantizer for Lightweight Deep Neural Networks
Hassan Dbouk
Hetul Sanghvi
M. Mehendale
Naresh R Shanbhag
MQ
19
9
0
19 Jul 2020
Quantized Neural Networks: Characterization and Holistic Optimization
Quantized Neural Networks: Characterization and Holistic Optimization
Yoonho Boo
Sungho Shin
Wonyong Sung
MQ
45
8
0
31 May 2020
WaveQ: Gradient-Based Deep Quantization of Neural Networks through
  Sinusoidal Adaptive Regularization
WaveQ: Gradient-Based Deep Quantization of Neural Networks through Sinusoidal Adaptive Regularization
Ahmed T. Elthakeb
Prannoy Pilligundla
Fatemehsadat Mireshghallah
T. Elgindi
Charles-Alban Deledalle
H. Esmaeilzadeh
MQ
17
10
0
29 Feb 2020
Optimal Gradient Quantization Condition for Communication-Efficient
  Distributed Training
Optimal Gradient Quantization Condition for Communication-Efficient Distributed Training
An Xu
Zhouyuan Huo
Heng-Chiao Huang
MQ
11
6
0
25 Feb 2020
Switchable Precision Neural Networks
Switchable Precision Neural Networks
Luis Guerra
Bohan Zhuang
Ian Reid
Tom Drummond
MQ
30
20
0
07 Feb 2020
SQWA: Stochastic Quantized Weight Averaging for Improving the
  Generalization Capability of Low-Precision Deep Neural Networks
SQWA: Stochastic Quantized Weight Averaging for Improving the Generalization Capability of Low-Precision Deep Neural Networks
Sungho Shin
Yoonho Boo
Wonyong Sung
MQ
22
3
0
02 Feb 2020
Least squares binary quantization of neural networks
Least squares binary quantization of neural networks
Hadi Pouransari
Zhucheng Tu
Oncel Tuzel
MQ
17
32
0
09 Jan 2020
Sparse Weight Activation Training
Sparse Weight Activation Training
Md Aamir Raihan
Tor M. Aamodt
34
73
0
07 Jan 2020
Towards Unified INT8 Training for Convolutional Neural Network
Towards Unified INT8 Training for Convolutional Neural Network
Feng Zhu
Ruihao Gong
F. Yu
Xianglong Liu
Yanfei Wang
Zhelong Li
Xiuqi Yang
Junjie Yan
MQ
35
150
0
29 Dec 2019
Adaptive Loss-aware Quantization for Multi-bit Networks
Adaptive Loss-aware Quantization for Multi-bit Networks
Zhongnan Qu
Zimu Zhou
Yun Cheng
Lothar Thiele
MQ
36
53
0
18 Dec 2019
Quantization Networks
Quantization Networks
Jiwei Yang
Xu Shen
Jun Xing
Xinmei Tian
Houqiang Li
Bing Deng
Jianqiang Huang
Xiansheng Hua
MQ
30
339
0
21 Nov 2019
Loss Aware Post-training Quantization
Loss Aware Post-training Quantization
Yury Nahshan
Brian Chmiel
Chaim Baskin
Evgenii Zheltonozhskii
Ron Banner
A. Bronstein
A. Mendelson
MQ
31
163
0
17 Nov 2019
Scientific Image Restoration Anywhere
Scientific Image Restoration Anywhere
V. Abeykoon
Zhengchun Liu
R. Kettimuthu
Geoffrey C. Fox
Ian Foster
19
19
0
12 Nov 2019
HAWQ-V2: Hessian Aware trace-Weighted Quantization of Neural Networks
HAWQ-V2: Hessian Aware trace-Weighted Quantization of Neural Networks
Zhen Dong
Z. Yao
Yaohui Cai
Daiyaan Arfeen
A. Gholami
Michael W. Mahoney
Kurt Keutzer
MQ
34
274
0
10 Nov 2019
12
Next