ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1810.05723
  4. Cited By
Post-training 4-bit quantization of convolution networks for
  rapid-deployment
v1v2v3 (latest)

Post-training 4-bit quantization of convolution networks for rapid-deployment

2 October 2018
Ron Banner
Yury Nahshan
Elad Hoffer
Daniel Soudry
    MQ
ArXiv (abs)PDFHTMLGithub (243★)

Papers citing "Post-training 4-bit quantization of convolution networks for rapid-deployment"

12 / 12 papers shown
Title
Improving Post Training Neural Quantization: Layer-wise Calibration and
  Integer Programming
Improving Post Training Neural Quantization: Layer-wise Calibration and Integer Programming
Itay Hubara
Yury Nahshan
Y. Hanani
Ron Banner
Daniel Soudry
MQ
98
127
0
14 Jun 2020
Discovering Low-Precision Networks Close to Full-Precision Networks for
  Efficient Embedded Inference
Discovering Low-Precision Networks Close to Full-Precision Networks for Efficient Embedded Inference
J. McKinstry
S. K. Esser
R. Appuswamy
Deepika Bablani
John V. Arthur
Izzet B. Yildiz
D. Modha
MQ
54
94
0
11 Sep 2018
Quantizing deep convolutional networks for efficient inference: A
  whitepaper
Quantizing deep convolutional networks for efficient inference: A whitepaper
Raghuraman Krishnamoorthi
MQ
141
1,019
0
21 Jun 2018
Scalable Methods for 8-bit Training of Neural Networks
Scalable Methods for 8-bit Training of Neural Networks
Ron Banner
Itay Hubara
Elad Hoffer
Daniel Soudry
MQ
84
339
0
25 May 2018
PACT: Parameterized Clipping Activation for Quantized Neural Networks
PACT: Parameterized Clipping Activation for Quantized Neural Networks
Jungwook Choi
Zhuo Wang
Swagath Venkataramani
P. Chuang
Vijayalakshmi Srinivasan
K. Gopalakrishnan
MQ
65
954
0
16 May 2018
Highly Efficient 8-bit Low Precision Inference of Convolutional Neural
  Networks with IntelCaffe
Highly Efficient 8-bit Low Precision Inference of Convolutional Neural Networks with IntelCaffe
Jiong Gong
Haihao Shen
Guoming Zhang
Xiaoli Liu
Shane Li
Ge Jin
Niharika Maheshwari
Evarist Fomenko
Eden Segal
BDLMQ
38
33
0
04 May 2018
Value-aware Quantization for Training and Inference of Neural Networks
Value-aware Quantization for Training and Inference of Neural Networks
Eunhyeok Park
S. Yoo
Peter Vajda
MQ
47
161
0
20 Apr 2018
Training and Inference with Integers in Deep Neural Networks
Training and Inference with Integers in Deep Neural Networks
Shuang Wu
Guoqi Li
F. Chen
Luping Shi
MQ
65
391
0
13 Feb 2018
Quantization and Training of Neural Networks for Efficient
  Integer-Arithmetic-Only Inference
Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference
Benoit Jacob
S. Kligys
Bo Chen
Menglong Zhu
Matthew Tang
Andrew G. Howard
Hartwig Adam
Dmitry Kalenichenko
MQ
156
3,138
0
15 Dec 2017
Towards Accurate Binary Convolutional Neural Network
Towards Accurate Binary Convolutional Neural Network
Xiaofan Lin
Cong Zhao
Wei Pan
MQ
84
647
0
30 Nov 2017
DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low
  Bitwidth Gradients
DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients
Shuchang Zhou
Yuxin Wu
Zekun Ni
Xinyu Zhou
He Wen
Yuheng Zou
MQ
122
2,090
0
20 Jun 2016
XNOR-Net: ImageNet Classification Using Binary Convolutional Neural
  Networks
XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks
Mohammad Rastegari
Vicente Ordonez
Joseph Redmon
Ali Farhadi
MQ
172
4,368
0
16 Mar 2016
1