ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1606.06160
  4. Cited By
DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low
  Bitwidth Gradients

DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients

20 June 2016
Shuchang Zhou
Yuxin Wu
Zekun Ni
Xinyu Zhou
He Wen
Yuheng Zou
    MQ
ArXivPDFHTML

Papers citing "DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients"

44 / 444 papers shown
Title
Distribution-Aware Binarization of Neural Networks for Sketch
  Recognition
Distribution-Aware Binarization of Neural Networks for Sketch Recognition
Ameya Prabhu
Vishal Batchu
Sri Aurobindo Munagala
Rohit Gajawada
A. Namboodiri
MQ
27
5
0
09 Apr 2018
Training DNNs with Hybrid Block Floating Point
Training DNNs with Hybrid Block Floating Point
M. Drumond
Tao R. Lin
Martin Jaggi
Babak Falsafi
25
95
0
04 Apr 2018
Diagonalwise Refactorization: An Efficient Training Method for Depthwise
  Convolutions
Diagonalwise Refactorization: An Efficient Training Method for Depthwise Convolutions
Zheng Qin
Zhaoning Zhang
Dongsheng Li
Yiming Zhang
Yuxing Peng
25
28
0
27 Mar 2018
Merging and Evolution: Improving Convolutional Neural Networks for
  Mobile Applications
Merging and Evolution: Improving Convolutional Neural Networks for Mobile Applications
Zheng Qin
Zhaoning Zhang
Shiqing Zhang
Hao Yu
Yuxing Peng
11
7
0
24 Mar 2018
Toolflows for Mapping Convolutional Neural Networks on FPGAs: A Survey
  and Future Directions
Toolflows for Mapping Convolutional Neural Networks on FPGAs: A Survey and Future Directions
Stylianos I. Venieris
Alexandros Kouris
C. Bouganis
19
184
0
15 Mar 2018
Deep Neural Network Compression with Single and Multiple Level
  Quantization
Deep Neural Network Compression with Single and Multiple Level Quantization
Yuhui Xu
Yongzhuang Wang
Aojun Zhou
Weiyao Lin
H. Xiong
MQ
20
114
0
06 Mar 2018
Demystifying Parallel and Distributed Deep Learning: An In-Depth
  Concurrency Analysis
Demystifying Parallel and Distributed Deep Learning: An In-Depth Concurrency Analysis
Tal Ben-Nun
Torsten Hoefler
GNN
33
704
0
26 Feb 2018
PBGen: Partial Binarization of Deconvolution-Based Generators for Edge
  Intelligence
PBGen: Partial Binarization of Deconvolution-Based Generators for Edge Intelligence
Jinglan Liu
Jiaxin Zhang
Yukun Ding
Xiaowei Xu
Meng Jiang
Yiyu Shi
41
4
0
26 Feb 2018
Loss-aware Weight Quantization of Deep Networks
Loss-aware Weight Quantization of Deep Networks
Lu Hou
James T. Kwok
MQ
35
127
0
23 Feb 2018
Training and Inference with Integers in Deep Neural Networks
Training and Inference with Integers in Deep Neural Networks
Shuang Wu
Guoqi Li
F. Chen
Luping Shi
MQ
41
389
0
13 Feb 2018
BinaryRelax: A Relaxation Approach For Training Deep Neural Networks
  With Quantized Weights
BinaryRelax: A Relaxation Approach For Training Deep Neural Networks With Quantized Weights
Penghang Yin
Shuai Zhang
J. Lyu
Stanley Osher
Y. Qi
Jack Xin
MQ
27
78
0
19 Jan 2018
Fix your classifier: the marginal value of training the last weight
  layer
Fix your classifier: the marginal value of training the last weight layer
Elad Hoffer
Itay Hubara
Daniel Soudry
35
101
0
14 Jan 2018
Quantization and Training of Neural Networks for Efficient
  Integer-Arithmetic-Only Inference
Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference
Benoit Jacob
S. Kligys
Bo Chen
Menglong Zhu
Matthew Tang
Andrew G. Howard
Hartwig Adam
Dmitry Kalenichenko
MQ
92
3,062
0
15 Dec 2017
High performance ultra-low-precision convolutions on mobile devices
High performance ultra-low-precision convolutions on mobile devices
Andrew Tulloch
Yangqing Jia
HAI
MQ
18
27
0
06 Dec 2017
Deep Gradient Compression: Reducing the Communication Bandwidth for
  Distributed Training
Deep Gradient Compression: Reducing the Communication Bandwidth for Distributed Training
Chengyue Wu
Song Han
Huizi Mao
Yu Wang
W. Dally
74
1,388
0
05 Dec 2017
Bit Fusion: Bit-Level Dynamically Composable Architecture for
  Accelerating Deep Neural Networks
Bit Fusion: Bit-Level Dynamically Composable Architecture for Accelerating Deep Neural Networks
Hardik Sharma
Jongse Park
Naveen Suda
Liangzhen Lai
Benson Chau
Joo-Young Kim
Vikas Chandra
H. Esmaeilzadeh
MQ
32
488
0
05 Dec 2017
Deep Expander Networks: Efficient Deep Networks from Graph Theory
Deep Expander Networks: Efficient Deep Networks from Graph Theory
Ameya Prabhu
G. Varma
A. Namboodiri
GNN
30
71
0
23 Nov 2017
ADaPTION: Toolbox and Benchmark for Training Convolutional Neural
  Networks with Reduced Numerical Precision Weights and Activation
ADaPTION: Toolbox and Benchmark for Training Convolutional Neural Networks with Reduced Numerical Precision Weights and Activation
Moritz B. Milde
Daniel Neil
Alessandro Aimar
T. Delbruck
Giacomo Indiveri
MQ
42
9
0
13 Nov 2017
Attacking Binarized Neural Networks
Attacking Binarized Neural Networks
A. Galloway
Graham W. Taylor
M. Moussa
MQ
AAML
14
104
0
01 Nov 2017
Minimum Energy Quantized Neural Networks
Minimum Energy Quantized Neural Networks
Bert Moons
Koen Goetschalckx
Nick Van Berckelaer
Marian Verhelst
MQ
33
123
0
01 Nov 2017
Towards Effective Low-bitwidth Convolutional Neural Networks
Towards Effective Low-bitwidth Convolutional Neural Networks
Bohan Zhuang
Chunhua Shen
Mingkui Tan
Lingqiao Liu
Ian Reid
MQ
39
232
0
01 Nov 2017
Deep Learning as a Mixed Convex-Combinatorial Optimization Problem
Deep Learning as a Mixed Convex-Combinatorial Optimization Problem
A. Friesen
Pedro M. Domingos
26
20
0
31 Oct 2017
Gradient Sparsification for Communication-Efficient Distributed
  Optimization
Gradient Sparsification for Communication-Efficient Distributed Optimization
Jianqiao Wangni
Jialei Wang
Ji Liu
Tong Zhang
15
522
0
26 Oct 2017
Mixed Precision Training
Mixed Precision Training
Paulius Micikevicius
Sharan Narang
Jonah Alben
G. Diamos
Erich Elsen
...
Boris Ginsburg
Michael Houston
Oleksii Kuchaiev
Ganesh Venkatesh
Hao Wu
90
1,767
0
10 Oct 2017
Flexible Network Binarization with Layer-wise Priority
He Wang
Yi Tian Xu
Bingbing Ni
Hongteng Xu
MQ
36
10
0
13 Sep 2017
WRPN: Wide Reduced-Precision Networks
WRPN: Wide Reduced-Precision Networks
Asit K. Mishra
Eriko Nurvitadhi
Jeffrey J. Cook
Debbie Marr
MQ
39
266
0
04 Sep 2017
BitNet: Bit-Regularized Deep Neural Networks
BitNet: Bit-Regularized Deep Neural Networks
Aswin Raghavan
Mohamed R. Amer
S. Chai
Graham Taylor
MQ
38
10
0
16 Aug 2017
Streaming Architecture for Large-Scale Quantized Neural Networks on an
  FPGA-Based Dataflow Platform
Streaming Architecture for Large-Scale Quantized Neural Networks on an FPGA-Based Dataflow Platform
Chaim Baskin
Natan Liss
Evgenii Zheltonozhskii
A. Bronstein
A. Mendelson
GNN
MQ
45
35
0
31 Jul 2017
Model compression as constrained optimization, with application to
  neural nets. Part II: quantization
Model compression as constrained optimization, with application to neural nets. Part II: quantization
M. A. Carreira-Perpiñán
Yerlan Idelbayev
MQ
28
37
0
13 Jul 2017
SEP-Nets: Small and Effective Pattern Networks
SEP-Nets: Small and Effective Pattern Networks
Zhe Li
Xiaoyu Wang
Xutao Lv
Tianbao Yang
30
12
0
13 Jun 2017
NullHop: A Flexible Convolutional Neural Network Accelerator Based on
  Sparse Representations of Feature Maps
NullHop: A Flexible Convolutional Neural Network Accelerator Based on Sparse Representations of Feature Maps
Alessandro Aimar
Hesham Mostafa
Enrico Calabrese
A. Rios-Navarro
Ricardo Tapiador-Morales
...
Moritz B. Milde
Federico Corradi
A. Linares-Barranco
Shih-Chii Liu
T. Delbruck
93
243
0
05 Jun 2017
TernGrad: Ternary Gradients to Reduce Communication in Distributed Deep
  Learning
TernGrad: Ternary Gradients to Reduce Communication in Distributed Deep Learning
W. Wen
Cong Xu
Feng Yan
Chunpeng Wu
Yandan Wang
Yiran Chen
Hai Helen Li
52
984
0
22 May 2017
More is Less: A More Complicated Network with Less Inference Complexity
More is Less: A More Complicated Network with Less Inference Complexity
Xuanyi Dong
Junshi Huang
Yi Yang
Shuicheng Yan
26
288
0
25 Mar 2017
Deep Convolutional Neural Network Inference with Floating-point Weights
  and Fixed-point Activations
Deep Convolutional Neural Network Inference with Floating-point Weights and Fixed-point Activations
Liangzhen Lai
Naveen Suda
Vikas Chandra
MQ
33
85
0
08 Mar 2017
Binarized Convolutional Landmark Localizers for Human Pose Estimation
  and Face Alignment with Limited Resources
Binarized Convolutional Landmark Localizers for Human Pose Estimation and Face Alignment with Limited Resources
Adrian Bulat
Georgios Tzimiropoulos
CVBM
3DV
32
191
0
02 Mar 2017
Incremental Network Quantization: Towards Lossless CNNs with
  Low-Precision Weights
Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights
Aojun Zhou
Anbang Yao
Yiwen Guo
Lin Xu
Yurong Chen
MQ
337
1,049
0
10 Feb 2017
Deep Learning with Low Precision by Half-wave Gaussian Quantization
Deep Learning with Low Precision by Half-wave Gaussian Quantization
Zhaowei Cai
Xiaodong He
Jian Sun
Nuno Vasconcelos
MQ
50
503
0
03 Feb 2017
Mixed Low-precision Deep Learning Inference using Dynamic Fixed Point
Mixed Low-precision Deep Learning Inference using Dynamic Fixed Point
Naveen Mellempudi
Abhisek Kundu
Dipankar Das
Dheevatsa Mudigere
Bharat Kaul
MQ
35
30
0
31 Jan 2017
Scaling Binarized Neural Networks on Reconfigurable Logic
Scaling Binarized Neural Networks on Reconfigurable Logic
Nicholas J. Fraser
Yaman Umuroglu
Giulio Gambardella
Michaela Blott
Philip H. W. Leong
Magnus Jahre
K. Vissers
MQ
20
57
0
12 Jan 2017
Trained Ternary Quantization
Trained Ternary Quantization
Chenzhuo Zhu
Song Han
Huizi Mao
W. Dally
MQ
92
1,035
0
04 Dec 2016
FINN: A Framework for Fast, Scalable Binarized Neural Network Inference
FINN: A Framework for Fast, Scalable Binarized Neural Network Inference
Yaman Umuroglu
Nicholas J. Fraser
Giulio Gambardella
Michaela Blott
Philip H. W. Leong
Magnus Jahre
K. Vissers
MQ
53
984
0
01 Dec 2016
Training Bit Fully Convolutional Network for Fast Semantic Segmentation
Training Bit Fully Convolutional Network for Fast Semantic Segmentation
He Wen
Shuchang Zhou
Zhe Liang
Yuxiang Zhang
Dieqiao Feng
Xinyu Zhou
Cong Yao
MQ
SSeg
37
10
0
01 Dec 2016
Quantized Neural Networks: Training Neural Networks with Low Precision
  Weights and Activations
Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations
Itay Hubara
Matthieu Courbariaux
Daniel Soudry
Ran El-Yaniv
Yoshua Bengio
MQ
54
1,846
0
22 Sep 2016
Ternary Neural Networks for Resource-Efficient AI Applications
Ternary Neural Networks for Resource-Efficient AI Applications
Hande Alemdar
V. Leroy
Adrien Prost-Boucle
F. Pétrot
24
204
0
01 Sep 2016
Previous
123456789