ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2203.12091
  4. Cited By
FxP-QNet: A Post-Training Quantizer for the Design of Mixed
  Low-Precision DNNs with Dynamic Fixed-Point Representation

FxP-QNet: A Post-Training Quantizer for the Design of Mixed Low-Precision DNNs with Dynamic Fixed-Point Representation

22 March 2022
Ahmad Shawahna
S. M. Sait
A. El-Maleh
Irfan Ahmad
    MQ
ArXiv (abs)PDFHTML

Papers citing "FxP-QNet: A Post-Training Quantizer for the Design of Mixed Low-Precision DNNs with Dynamic Fixed-Point Representation"

50 / 62 papers shown
Title
How to Reach Real-Time AI on Consumer Devices? Solutions for
  Programmable and Custom Architectures
How to Reach Real-Time AI on Consumer Devices? Solutions for Programmable and Custom Architectures
Stylianos I. Venieris
Ioannis Panopoulos
Ilias Leontiadis
I. Venieris
71
6
0
21 Jun 2021
Hardware-Centric AutoML for Mixed-Precision Quantization
Hardware-Centric AutoML for Mixed-Precision Quantization
Kuan-Chieh Wang
Zhijian Liu
Chengyue Wu
Ji Lin
Song Han
MQ
60
14
0
11 Aug 2020
PROFIT: A Novel Training Method for sub-4-bit MobileNet Models
PROFIT: A Novel Training Method for sub-4-bit MobileNet Models
Eunhyeok Park
S. Yoo
MQ
52
85
0
11 Aug 2020
Multi-Precision Policy Enforced Training (MuPPET): A precision-switching
  strategy for quantised fixed-point training of CNNs
Multi-Precision Policy Enforced Training (MuPPET): A precision-switching strategy for quantised fixed-point training of CNNs
A. Rajagopal
D. A. Vink
Stylianos I. Venieris
C. Bouganis
MQ
66
15
0
16 Jun 2020
Binary Neural Networks: A Survey
Binary Neural Networks: A Survey
Haotong Qin
Ruihao Gong
Xianglong Liu
Xiao Bai
Jingkuan Song
N. Sebe
MQ
126
468
0
31 Mar 2020
Memory-Driven Mixed Low Precision Quantization For Enabling Deep Network
  Inference On Microcontrollers
Memory-Driven Mixed Low Precision Quantization For Enabling Deep Network Inference On Microcontrollers
Manuele Rusci
Alessandro Capotondi
Luca Benini
MQ
81
75
0
30 May 2019
HAWQ: Hessian AWare Quantization of Neural Networks with Mixed-Precision
HAWQ: Hessian AWare Quantization of Neural Networks with Mixed-Precision
Zhen Dong
Z. Yao
A. Gholami
Michael W. Mahoney
Kurt Keutzer
MQ
85
526
0
29 Apr 2019
Learned Step Size Quantization
Learned Step Size Quantization
S. K. Esser
J. McKinstry
Deepika Bablani
R. Appuswamy
D. Modha
MQ
75
806
0
21 Feb 2019
Low-bit Quantization of Neural Networks for Efficient Inference
Low-bit Quantization of Neural Networks for Efficient Inference
Yoni Choukroun
Eli Kravchik
Fan Yang
P. Kisilev
MQ
72
363
0
18 Feb 2019
Are All Layers Created Equal?
Are All Layers Created Equal?
Chiyuan Zhang
Samy Bengio
Y. Singer
64
140
0
06 Feb 2019
Same, Same But Different - Recovering Neural Network Quantization Error
  Through Weight Factorization
Same, Same But Different - Recovering Neural Network Quantization Error Through Weight Factorization
Eldad Meller
Alexander Finkelstein
Uri Almog
Mark Grobman
MQ
56
86
0
05 Feb 2019
FPGA-based Accelerators of Deep Learning Networks for Learning and
  Classification: A Review
FPGA-based Accelerators of Deep Learning Networks for Learning and Classification: A Review
Ahmad Shawahna
S. M. Sait
A. El-Maleh
67
378
0
01 Jan 2019
Training Competitive Binary Neural Networks from Scratch
Training Competitive Binary Neural Networks from Scratch
Joseph Bethge
Marvin Bornstein
Adrian Loy
Haojin Yang
Christoph Meinel
MQ
59
33
0
05 Dec 2018
Mixed Precision Quantization of ConvNets via Differentiable Neural
  Architecture Search
Mixed Precision Quantization of ConvNets via Differentiable Neural Architecture Search
Bichen Wu
Yanghan Wang
Peizhao Zhang
Yuandong Tian
Peter Vajda
Kurt Keutzer
MQ
68
273
0
30 Nov 2018
HAQ: Hardware-Aware Automated Quantization with Mixed Precision
HAQ: Hardware-Aware Automated Quantization with Mixed Precision
Kuan-Chieh Wang
Zhijian Liu
Chengyue Wu
Ji Lin
Song Han
MQ
127
882
0
21 Nov 2018
A Unified Framework of DNN Weight Pruning and Weight
  Clustering/Quantization Using ADMM
A Unified Framework of DNN Weight Pruning and Weight Clustering/Quantization Using ADMM
David Cortes
Tianyun Zhang
Kaiqi Zhang
Jiayu Li
Jiaming Xie
Yun Liang
Sijia Liu
Xinyu Lin
Yanzhi Wang
MQ
38
45
0
05 Nov 2018
Quantization for Rapid Deployment of Deep Neural Networks
Quantization for Rapid Deployment of Deep Neural Networks
J. Lee
Sangwon Ha
Saerom Choi
Won-Jo Lee
Seungwon Lee
MQ
44
49
0
12 Oct 2018
Post-training 4-bit quantization of convolution networks for
  rapid-deployment
Post-training 4-bit quantization of convolution networks for rapid-deployment
Ron Banner
Yury Nahshan
Elad Hoffer
Daniel Soudry
MQ
50
94
0
02 Oct 2018
Learning to Quantize Deep Networks by Optimizing Quantization Intervals
  with Task Loss
Learning to Quantize Deep Networks by Optimizing Quantization Intervals with Task Loss
S. Jung
Changyong Son
Seohyung Lee
JinWoo Son
Youngjun Kwak
Jae-Joon Han
Sung Ju Hwang
Changkyu Choi
MQ
48
375
0
17 Aug 2018
LQ-Nets: Learned Quantization for Highly Accurate and Compact Deep
  Neural Networks
LQ-Nets: Learned Quantization for Highly Accurate and Compact Deep Neural Networks
Dongqing Zhang
Jiaolong Yang
Dongqiangzi Ye
G. Hua
MQ
62
703
0
26 Jul 2018
Bridging the Accuracy Gap for 2-bit Quantized Neural Networks (QNN)
Bridging the Accuracy Gap for 2-bit Quantized Neural Networks (QNN)
Jungwook Choi
P. Chuang
Zhuo Wang
Swagath Venkataramani
Vijayalakshmi Srinivasan
K. Gopalakrishnan
MQ
42
76
0
17 Jul 2018
SYQ: Learning Symmetric Quantization For Efficient Deep Neural Networks
SYQ: Learning Symmetric Quantization For Efficient Deep Neural Networks
Julian Faraone
Nicholas J. Fraser
Michaela Blott
Philip H. W. Leong
MQ
77
133
0
01 Jul 2018
Quantizing deep convolutional networks for efficient inference: A
  whitepaper
Quantizing deep convolutional networks for efficient inference: A whitepaper
Raghuraman Krishnamoorthi
MQ
141
1,016
0
21 Jun 2018
CascadeCNN: Pushing the performance limits of quantisation
CascadeCNN: Pushing the performance limits of quantisation
Alexandros Kouris
Stylianos I. Venieris
C. Bouganis
MQ
47
24
0
22 May 2018
PACT: Parameterized Clipping Activation for Quantized Neural Networks
PACT: Parameterized Clipping Activation for Quantized Neural Networks
Jungwook Choi
Zhuo Wang
Swagath Venkataramani
P. Chuang
Vijayalakshmi Srinivasan
K. Gopalakrishnan
MQ
65
953
0
16 May 2018
Loss-aware Weight Quantization of Deep Networks
Loss-aware Weight Quantization of Deep Networks
Lu Hou
James T. Kwok
MQ
82
127
0
23 Feb 2018
Quantization and Training of Neural Networks for Efficient
  Integer-Arithmetic-Only Inference
Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference
Benoit Jacob
S. Kligys
Bo Chen
Menglong Zhu
Matthew Tang
Andrew G. Howard
Hartwig Adam
Dmitry Kalenichenko
MQ
156
3,130
0
15 Dec 2017
Bit Fusion: Bit-Level Dynamically Composable Architecture for
  Accelerating Deep Neural Networks
Bit Fusion: Bit-Level Dynamically Composable Architecture for Accelerating Deep Neural Networks
Hardik Sharma
Jongse Park
Naveen Suda
Liangzhen Lai
Benson Chau
Joo-Young Kim
Vikas Chandra
H. Esmaeilzadeh
MQ
61
491
0
05 Dec 2017
Apprentice: Using Knowledge Distillation Techniques To Improve
  Low-Precision Network Accuracy
Apprentice: Using Knowledge Distillation Techniques To Improve Low-Precision Network Accuracy
Asit K. Mishra
Debbie Marr
FedML
65
331
0
15 Nov 2017
Towards Effective Low-bitwidth Convolutional Neural Networks
Towards Effective Low-bitwidth Convolutional Neural Networks
Bohan Zhuang
Chunhua Shen
Mingkui Tan
Lingqiao Liu
Ian Reid
MQ
70
232
0
01 Nov 2017
Mixed Precision Training
Mixed Precision Training
Paulius Micikevicius
Sharan Narang
Jonah Alben
G. Diamos
Erich Elsen
...
Boris Ginsburg
Michael Houston
Oleksii Kuchaiev
Ganesh Venkatesh
Hao Wu
157
1,799
0
10 Oct 2017
WRPN: Wide Reduced-Precision Networks
WRPN: Wide Reduced-Precision Networks
Asit K. Mishra
Eriko Nurvitadhi
Jeffrey J. Cook
Debbie Marr
MQ
78
267
0
04 Sep 2017
Balanced Quantization: An Effective and Efficient Approach to Quantized
  Neural Networks
Balanced Quantization: An Effective and Efficient Approach to Quantized Neural Networks
Shuchang Zhou
Yuzhi Wang
He Wen
Qinyao He
Yuheng Zou
MQ
85
110
0
22 Jun 2017
Ternary Neural Networks with Fine-Grained Quantization
Ternary Neural Networks with Fine-Grained Quantization
Naveen Mellempudi
Abhisek Kundu
Dheevatsa Mudigere
Dipankar Das
Bharat Kaul
Pradeep Dubey
MQ
94
111
0
02 May 2017
Efficient Processing of Deep Neural Networks: A Tutorial and Survey
Efficient Processing of Deep Neural Networks: A Tutorial and Survey
Vivienne Sze
Yu-hsin Chen
Tien-Ju Yang
J. Emer
AAML3DV
120
3,022
0
27 Mar 2017
Mask R-CNN
Mask R-CNN
Kaiming He
Georgia Gkioxari
Piotr Dollár
Ross B. Girshick
ObjD
352
27,195
0
20 Mar 2017
Low-Precision Batch-Normalized Activations
Low-Precision Batch-Normalized Activations
Benjamin Graham
MQ
49
9
0
27 Feb 2017
Incremental Network Quantization: Towards Lossless CNNs with
  Low-Precision Weights
Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights
Aojun Zhou
Anbang Yao
Yiwen Guo
Lin Xu
Yurong Chen
MQ
395
1,051
0
10 Feb 2017
Deep Learning with Low Precision by Half-wave Gaussian Quantization
Deep Learning with Low Precision by Half-wave Gaussian Quantization
Zhaowei Cai
Xiaodong He
Jian Sun
Nuno Vasconcelos
MQ
136
505
0
03 Feb 2017
Trained Ternary Quantization
Trained Ternary Quantization
Chenzhuo Zhu
Song Han
Huizi Mao
W. Dally
MQ
137
1,035
0
04 Dec 2016
SqueezeDet: Unified, Small, Low Power Fully Convolutional Neural
  Networks for Real-Time Object Detection for Autonomous Driving
SqueezeDet: Unified, Small, Low Power Fully Convolutional Neural Networks for Real-Time Object Detection for Autonomous Driving
Bichen Wu
Alvin Wan
F. Iandola
Peter H. Jin
Kurt Keutzer
90
513
0
04 Dec 2016
Loss-aware Binarization of Deep Networks
Loss-aware Binarization of Deep Networks
Lu Hou
Quanming Yao
James T. Kwok
MQ
70
220
0
05 Nov 2016
Quantized Neural Networks: Training Neural Networks with Low Precision
  Weights and Activations
Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations
Itay Hubara
Matthieu Courbariaux
Daniel Soudry
Ran El-Yaniv
Yoshua Bengio
MQ
149
1,866
0
22 Sep 2016
Densely Connected Convolutional Networks
Densely Connected Convolutional Networks
Gao Huang
Zhuang Liu
Laurens van der Maaten
Kilian Q. Weinberger
PINN3DV
775
36,813
0
25 Aug 2016
DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low
  Bitwidth Gradients
DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients
Shuchang Zhou
Yuxin Wu
Zekun Ni
Xinyu Zhou
He Wen
Yuheng Zou
MQ
122
2,088
0
20 Jun 2016
Fully Convolutional Networks for Semantic Segmentation
Fully Convolutional Networks for Semantic Segmentation
Evan Shelhamer
Jonathan Long
Trevor Darrell
VOSSSeg
741
37,862
0
20 May 2016
Ternary Weight Networks
Ternary Weight Networks
Fengfu Li
Bin Liu
Xiaoxing Wang
Bo Zhang
Junchi Yan
MQ
76
525
0
16 May 2016
XNOR-Net: ImageNet Classification Using Binary Convolutional Neural
  Networks
XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks
Mohammad Rastegari
Vicente Ordonez
Joseph Redmon
Ali Farhadi
MQ
170
4,357
0
16 Mar 2016
Convolutional Neural Networks using Logarithmic Data Representation
Convolutional Neural Networks using Logarithmic Data Representation
Daisuke Miyashita
Edward H. Lee
B. Murmann
MQ
81
428
0
03 Mar 2016
EIE: Efficient Inference Engine on Compressed Deep Neural Network
EIE: Efficient Inference Engine on Compressed Deep Neural Network
Song Han
Xingyu Liu
Huizi Mao
Jing Pu
A. Pedram
M. Horowitz
W. Dally
127
2,457
0
04 Feb 2016
12
Next