ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1702.03044
  4. Cited By
Incremental Network Quantization: Towards Lossless CNNs with
  Low-Precision Weights

Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights

10 February 2017
Aojun Zhou
Anbang Yao
Yiwen Guo
Lin Xu
Yurong Chen
    MQ
ArXivPDFHTML

Papers citing "Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights"

50 / 464 papers shown
Title
Joint Pruning & Quantization for Extremely Sparse Neural Networks
Joint Pruning & Quantization for Extremely Sparse Neural Networks
Po-Hsiang Yu
Sih-Sian Wu
Jan P. Klopp
Liang-Gee Chen
Shao-Yi Chien
MQ
17
14
0
05 Oct 2020
BCNN: A Binary CNN with All Matrix Ops Quantized to 1 Bit Precision
BCNN: A Binary CNN with All Matrix Ops Quantized to 1 Bit Precision
A. Redfern
Lijun Zhu
Molly K. Newquist
MQ
25
12
0
01 Oct 2020
MS-RANAS: Multi-Scale Resource-Aware Neural Architecture Search
MS-RANAS: Multi-Scale Resource-Aware Neural Architecture Search
Cristian Cioflan
Radu Timofte
9
2
0
29 Sep 2020
BAMSProd: A Step towards Generalizing the Adaptive Optimization Methods
  to Deep Binary Model
BAMSProd: A Step towards Generalizing the Adaptive Optimization Methods to Deep Binary Model
Junjie Liu
Dongchao Wen
Deyu Wang
Wei Tao
Tse-Wei Chen
Kinya Osa
Masami Kato
MQ
19
1
0
29 Sep 2020
On Efficient Constructions of Checkpoints
On Efficient Constructions of Checkpoints
Yu Chen
Zhenming Liu
Bin Ren
Xin Jin
OffRL
29
12
0
28 Sep 2020
Holistic Filter Pruning for Efficient Deep Neural Networks
Holistic Filter Pruning for Efficient Deep Neural Networks
Lukas Enderich
Fabian Timm
Wolfram Burgard
24
7
0
17 Sep 2020
MSP: An FPGA-Specific Mixed-Scheme, Multi-Precision Deep Neural Network
  Quantization Framework
MSP: An FPGA-Specific Mixed-Scheme, Multi-Precision Deep Neural Network Quantization Framework
Sung-En Chang
Yanyu Li
Mengshu Sun
Weiwen Jiang
Runbin Shi
Xue Lin
Yanzhi Wang
MQ
19
7
0
16 Sep 2020
An FPGA Accelerated Method for Training Feed-forward Neural Networks
  Using Alternating Direction Method of Multipliers and LSMR
An FPGA Accelerated Method for Training Feed-forward Neural Networks Using Alternating Direction Method of Multipliers and LSMR
Seyedeh Niusha Alavi Foumani
Ce Guo
Wayne Luk
14
3
0
06 Sep 2020
Layer-specific Optimization for Mixed Data Flow with Mixed Precision in
  FPGA Design for CNN-based Object Detectors
Layer-specific Optimization for Mixed Data Flow with Mixed Precision in FPGA Design for CNN-based Object Detectors
Duy-Thanh Nguyen
Hyun Kim
Hyuk-Jae Lee
MQ
17
59
0
03 Sep 2020
Transform Quantization for CNN (Convolutional Neural Network)
  Compression
Transform Quantization for CNN (Convolutional Neural Network) Compression
Sean I. Young
Wang Zhe
David S. Taubman
B. Girod
MQ
29
69
0
02 Sep 2020
Dual Precision Deep Neural Network
Dual Precision Deep Neural Network
J. Park
J. Choi
J. Ko
14
1
0
02 Sep 2020
One Weight Bitwidth to Rule Them All
One Weight Bitwidth to Rule Them All
Ting-Wu Chin
P. Chuang
Vikas Chandra
Diana Marculescu
MQ
20
25
0
22 Aug 2020
Channel-wise Hessian Aware trace-Weighted Quantization of Neural
  Networks
Channel-wise Hessian Aware trace-Weighted Quantization of Neural Networks
Xu Qian
Victor Li
Darren Crews
MQ
16
9
0
19 Aug 2020
SPINN: Synergistic Progressive Inference of Neural Networks over Device
  and Cloud
SPINN: Synergistic Progressive Inference of Neural Networks over Device and Cloud
Stefanos Laskaridis
Stylianos I. Venieris
Mario Almeida
Ilias Leontiadis
Nicholas D. Lane
28
265
0
14 Aug 2020
Towards Modality Transferable Visual Information Representation with
  Optimal Model Compression
Towards Modality Transferable Visual Information Representation with Optimal Model Compression
Rongqun Lin
Linwei Zhu
Shiqi Wang
Sam Kwong
22
2
0
13 Aug 2020
Towards Lossless Binary Convolutional Neural Networks Using Piecewise
  Approximation
Towards Lossless Binary Convolutional Neural Networks Using Piecewise Approximation
Baozhou Zhu
Zaid Al-Ars
Wei Pan
MQ
14
8
0
08 Aug 2020
Search What You Want: Barrier Panelty NAS for Mixed Precision
  Quantization
Search What You Want: Barrier Panelty NAS for Mixed Precision Quantization
Haibao Yu
Qi Han
Jianbo Li
Jianping Shi
Guangliang Cheng
Bin Fan
MQ
19
61
0
20 Jul 2020
HMQ: Hardware Friendly Mixed Precision Quantization Block for CNNs
HMQ: Hardware Friendly Mixed Precision Quantization Block for CNNs
H. Habi
Roy H. Jennings
Arnon Netzer
MQ
21
65
0
20 Jul 2020
Compression strategies and space-conscious representations for deep
  neural networks
Compression strategies and space-conscious representations for deep neural networks
Giosuè Cataldo Marinò
G. Ghidoli
Marco Frasca
D. Malchiodi
4
10
0
15 Jul 2020
AQD: Towards Accurate Fully-Quantized Object Detection
AQD: Towards Accurate Fully-Quantized Object Detection
Peng Chen
Jing Liu
Bohan Zhuang
Mingkui Tan
Chunhua Shen
MQ
29
10
0
14 Jul 2020
Term Revealing: Furthering Quantization at Run Time on Quantized DNNs
Term Revealing: Furthering Quantization at Run Time on Quantized DNNs
H. T. Kung
Bradley McDanel
S. Zhang
MQ
13
9
0
13 Jul 2020
AUSN: Approximately Uniform Quantization by Adaptively Superimposing
  Non-uniform Distribution for Deep Neural Networks
AUSN: Approximately Uniform Quantization by Adaptively Superimposing Non-uniform Distribution for Deep Neural Networks
Fangxin Liu
Wenbo Zhao
Yanzhi Wang
Changzhi Dai
Li Jiang
MQ
17
3
0
08 Jul 2020
EasyQuant: Post-training Quantization via Scale Optimization
EasyQuant: Post-training Quantization via Scale Optimization
Di Wu
Qingming Tang
Yongle Zhao
Ming Zhang
Ying Fu
Debing Zhang
MQ
22
75
0
30 Jun 2020
Learning compositional functions via multiplicative weight updates
Learning compositional functions via multiplicative weight updates
Jeremy Bernstein
Jiawei Zhao
M. Meister
Xuan Li
Anima Anandkumar
Yisong Yue
10
26
0
25 Jun 2020
Distilling Object Detectors with Task Adaptive Regularization
Distilling Object Detectors with Task Adaptive Regularization
Ruoyu Sun
Fuhui Tang
Xiaopeng Zhang
H. Xiong
Qi Tian
ObjD
4
56
0
23 Jun 2020
Efficient Integer-Arithmetic-Only Convolutional Neural Networks
Efficient Integer-Arithmetic-Only Convolutional Neural Networks
Hengrui Zhao
Dong Liu
Houqiang Li
MQ
23
4
0
21 Jun 2020
Improving Post Training Neural Quantization: Layer-wise Calibration and
  Integer Programming
Improving Post Training Neural Quantization: Layer-wise Calibration and Integer Programming
Itay Hubara
Yury Nahshan
Y. Hanani
Ron Banner
Daniel Soudry
MQ
24
122
0
14 Jun 2020
CoDeNet: Efficient Deployment of Input-Adaptive Object Detection on
  Embedded FPGAs
CoDeNet: Efficient Deployment of Input-Adaptive Object Detection on Embedded FPGAs
Zhen Dong
Dequan Wang
Qijing Huang
Yizhao Gao
Yaohui Cai
Tian Li
Bichen Wu
Kurt Keutzer
J. Wawrzynek
ObjD
31
1
0
12 Jun 2020
Neural Network Activation Quantization with Bitwise Information
  Bottlenecks
Neural Network Activation Quantization with Bitwise Information Bottlenecks
Xichuan Zhou
Kui Liu
Cong Shi
Haijun Liu
Ji Liu
MQ
19
1
0
09 Jun 2020
Conditional Neural Architecture Search
Conditional Neural Architecture Search
Sheng-Chun Kao
Arun Ramamurthy
Reed Williams
T. Krishna
10
0
0
06 Jun 2020
Generative Design of Hardware-aware DNNs
Generative Design of Hardware-aware DNNs
Sheng-Chun Kao
Arun Ramamurthy
T. Krishna
MQ
11
2
0
06 Jun 2020
An Overview of Neural Network Compression
An Overview of Neural Network Compression
James OÑeill
AI4CE
45
98
0
05 Jun 2020
AnalogNet: Convolutional Neural Network Inference on Analog Focal Plane
  Sensor Processors
AnalogNet: Convolutional Neural Network Inference on Analog Focal Plane Sensor Processors
Matthew Z. Wong
Benoît Guillard
Riku Murai
Sajad Saeedi
Paul H. J. Kelly
15
15
0
02 Jun 2020
VecQ: Minimal Loss DNN Model Compression With Vectorized Weight
  Quantization
VecQ: Minimal Loss DNN Model Compression With Vectorized Weight Quantization
Cheng Gong
Yao Chen
Ye Lu
Tao Li
Cong Hao
Deming Chen
MQ
14
44
0
18 May 2020
Binarizing MobileNet via Evolution-based Searching
Binarizing MobileNet via Evolution-based Searching
Hai T. Phan
Zechun Liu
Dang T. Huynh
Marios Savvides
Kwang-Ting Cheng
Zhiqiang Shen
3DV
MQ
32
43
0
13 May 2020
An Experimental Study of Reduced-Voltage Operation in Modern FPGAs for
  Neural Network Acceleration
An Experimental Study of Reduced-Voltage Operation in Modern FPGAs for Neural Network Acceleration
Behzad Salami
Erhan Baturay Onural
Ismail Emir Yüksel
Fahrettin Koc
Oguz Ergin
A. Cristal
O. Unsal
H. Sarbazi-Azad
O. Mutlu
22
45
0
04 May 2020
NTIRE 2020 Challenge on Image and Video Deblurring
NTIRE 2020 Challenge on Image and Video Deblurring
Seungjun Nah
Sanghyun Son
Radu Timofte
Kyoung Mu Lee
64
32
0
04 May 2020
TRP: Trained Rank Pruning for Efficient Deep Neural Networks
TRP: Trained Rank Pruning for Efficient Deep Neural Networks
Yuhui Xu
Yuxi Li
Shuai Zhang
W. Wen
Botao Wang
Y. Qi
Yiran Chen
Weiyao Lin
H. Xiong
AAML
25
71
0
30 Apr 2020
Real-Time Apple Detection System Using Embedded Systems With Hardware
  Accelerators: An Edge AI Application
Real-Time Apple Detection System Using Embedded Systems With Hardware Accelerators: An Edge AI Application
Vittorio Mazzia
Francesco Salvetti
Aleem Khaliq
Marcello Chiaberge
22
152
0
28 Apr 2020
Fully Embedding Fast Convolutional Networks on Pixel Processor Arrays
Fully Embedding Fast Convolutional Networks on Pixel Processor Arrays
Laurie Bose
Jianing Chen
S. Carey
Piotr Dudek
W. Mayol-Cuevas
20
37
0
27 Apr 2020
A Unified DNN Weight Compression Framework Using Reweighted Optimization
  Methods
A Unified DNN Weight Compression Framework Using Reweighted Optimization Methods
Tianyun Zhang
Xiaolong Ma
Zheng Zhan
Shangli Zhou
Minghai Qin
Fei Sun
Yen-kuang Chen
Caiwen Ding
M. Fardad
Yanzhi Wang
25
5
0
12 Apr 2020
LadaBERT: Lightweight Adaptation of BERT through Hybrid Model
  Compression
LadaBERT: Lightweight Adaptation of BERT through Hybrid Model Compression
Yihuan Mao
Yujing Wang
Chufan Wu
Chen Zhang
Yang-Feng Wang
Yaming Yang
Quanlu Zhang
Yunhai Tong
Jing Bai
14
72
0
08 Apr 2020
A Learning Framework for n-bit Quantized Neural Networks toward FPGAs
A Learning Framework for n-bit Quantized Neural Networks toward FPGAs
Jun Chen
L. Liu
Yong Liu
Xianfang Zeng
MQ
18
26
0
06 Apr 2020
Binary Neural Networks: A Survey
Binary Neural Networks: A Survey
Haotong Qin
Ruihao Gong
Xianglong Liu
Xiao Bai
Jingkuan Song
N. Sebe
MQ
50
457
0
31 Mar 2020
Rethinking Depthwise Separable Convolutions: How Intra-Kernel
  Correlations Lead to Improved MobileNets
Rethinking Depthwise Separable Convolutions: How Intra-Kernel Correlations Lead to Improved MobileNets
D. Haase
Manuel Amthor
20
132
0
30 Mar 2020
DA-NAS: Data Adapted Pruning for Efficient Neural Architecture Search
DA-NAS: Data Adapted Pruning for Efficient Neural Architecture Search
Xiyang Dai
Dongdong Chen
Mengchen Liu
Yinpeng Chen
Lu Yuan
24
20
0
27 Mar 2020
CoCoPIE: Making Mobile AI Sweet As PIE --Compression-Compilation
  Co-Design Goes a Long Way
CoCoPIE: Making Mobile AI Sweet As PIE --Compression-Compilation Co-Design Goes a Long Way
Shaoshan Liu
Bin Ren
Xipeng Shen
Yanzhi Wang
4
18
0
14 Mar 2020
Kernel Quantization for Efficient Network Compression
Kernel Quantization for Efficient Network Compression
Zhongzhi Yu
Yemin Shi
Tiejun Huang
Yizhou Yu
MQ
21
3
0
11 Mar 2020
ReActNet: Towards Precise Binary Neural Network with Generalized
  Activation Functions
ReActNet: Towards Precise Binary Neural Network with Generalized Activation Functions
Zechun Liu
Zhiqiang Shen
Marios Savvides
Kwang-Ting Cheng
MQ
25
347
0
07 Mar 2020
Propagating Asymptotic-Estimated Gradients for Low Bitwidth Quantized
  Neural Networks
Propagating Asymptotic-Estimated Gradients for Low Bitwidth Quantized Neural Networks
Jun Chen
Yong Liu
Hao Zhang
Shengnan Hou
Jian Yang
MQ
17
7
0
04 Mar 2020
Previous
123456...8910
Next