ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1702.03044
  4. Cited By
Incremental Network Quantization: Towards Lossless CNNs with
  Low-Precision Weights

Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights

10 February 2017
Aojun Zhou
Anbang Yao
Yiwen Guo
Lin Xu
Yurong Chen
    MQ
ArXivPDFHTML

Papers citing "Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights"

50 / 464 papers shown
Title
ImageSig: A signature transform for ultra-lightweight image recognition
ImageSig: A signature transform for ultra-lightweight image recognition
Mohamed Ramzy Ibrahim
Terry Lyons
VLM
19
7
0
13 May 2022
A Survey on AI Sustainability: Emerging Trends on Learning Algorithms
  and Research Challenges
A Survey on AI Sustainability: Emerging Trends on Learning Algorithms and Research Challenges
Zhenghua Chen
Min-man Wu
Alvin Chan
Xiaoli Li
Yew-Soon Ong
14
6
0
08 May 2022
PVNAS: 3D Neural Architecture Search with Point-Voxel Convolution
PVNAS: 3D Neural Architecture Search with Point-Voxel Convolution
Zhijian Liu
Haotian Tang
Shengyu Zhao
Kevin Shao
Song Han
3DPC
15
40
0
25 Apr 2022
Joint Coreset Construction and Quantization for Distributed Machine
  Learning
Joint Coreset Construction and Quantization for Distributed Machine Learning
Hanlin Lu
Changchang Liu
Shiqiang Wang
T. He
V. Narayanan
Kevin S. Chan
Stephen Pasteris
6
2
0
13 Apr 2022
E^2TAD: An Energy-Efficient Tracking-based Action Detector
E^2TAD: An Energy-Efficient Tracking-based Action Detector
Xin Hu
Zhenyu Wu
Haoyuan Miao
Siqi Fan
Taiyu Long
...
Pengcheng Pi
Yi Wu
Zhou Ren
Zhangyang Wang
G. Hua
21
2
0
09 Apr 2022
ShiftNAS: Towards Automatic Generation of Advanced Mulitplication-Less
  Neural Networks
ShiftNAS: Towards Automatic Generation of Advanced Mulitplication-Less Neural Networks
Xiaoxuan Lou
Guowen Xu
Kangjie Chen
Guanlin Li
Jiwei Li
Tianwei Zhang
OOD
MQ
30
0
0
07 Apr 2022
LilNetX: Lightweight Networks with EXtreme Model Compression and
  Structured Sparsification
LilNetX: Lightweight Networks with EXtreme Model Compression and Structured Sparsification
Sharath Girish
Kamal Gupta
Saurabh Singh
Abhinav Shrivastava
28
11
0
06 Apr 2022
It's All In the Teacher: Zero-Shot Quantization Brought Closer to the
  Teacher
It's All In the Teacher: Zero-Shot Quantization Brought Closer to the Teacher
Kanghyun Choi
Hye Yoon Lee
Deokki Hong
Joonsang Yu
Noseong Park
Youngsok Kim
Jinho Lee
MQ
19
31
0
31 Mar 2022
REx: Data-Free Residual Quantization Error Expansion
REx: Data-Free Residual Quantization Error Expansion
Edouard Yvinec
Arnaud Dapgony
Matthieu Cord
Kévin Bailly
MQ
26
8
0
28 Mar 2022
FxP-QNet: A Post-Training Quantizer for the Design of Mixed
  Low-Precision DNNs with Dynamic Fixed-Point Representation
FxP-QNet: A Post-Training Quantizer for the Design of Mixed Low-Precision DNNs with Dynamic Fixed-Point Representation
Ahmad Shawahna
S. M. Sait
A. El-Maleh
Irfan Ahmad
MQ
18
6
0
22 Mar 2022
Online Continual Learning for Embedded Devices
Online Continual Learning for Embedded Devices
Tyler L. Hayes
Christopher Kanan
CLL
25
54
0
21 Mar 2022
Hardware Approximate Techniques for Deep Neural Network Accelerators: A
  Survey
Hardware Approximate Techniques for Deep Neural Network Accelerators: A Survey
Giorgos Armeniakos
Georgios Zervakis
Dimitrios Soudris
J. Henkel
201
93
0
16 Mar 2022
LDP: Learnable Dynamic Precision for Efficient Deep Neural Network
  Training and Inference
LDP: Learnable Dynamic Precision for Efficient Deep Neural Network Training and Inference
Zhongzhi Yu
Y. Fu
Shang Wu
Mengquan Li
Haoran You
Yingyan Lin
24
1
0
15 Mar 2022
DNN Training Acceleration via Exploring GPGPU Friendly Sparsity
DNN Training Acceleration via Exploring GPGPU Friendly Sparsity
Zhuoran Song
Yihong Xu
Han Li
Naifeng Jing
Xiaoyao Liang
Li Jiang
27
3
0
11 Mar 2022
Standard Deviation-Based Quantization for Deep Neural Networks
Standard Deviation-Based Quantization for Deep Neural Networks
Amir Ardakani
A. Ardakani
B. Meyer
J. Clark
W. Gross
MQ
41
1
0
24 Feb 2022
Distilled Neural Networks for Efficient Learning to Rank
Distilled Neural Networks for Efficient Learning to Rank
F. M. Nardini
Cosimo Rulli
Salvatore Trani
Rossano Venturini
FedML
24
16
0
22 Feb 2022
ICSML: Industrial Control Systems ML Framework for native inference
  using IEC 61131-3 code
ICSML: Industrial Control Systems ML Framework for native inference using IEC 61131-3 code
Constantine Doumanidis
Prashant Hari Narayan Rajput
Michail Maniatakos
12
2
0
21 Feb 2022
Pruning Networks with Cross-Layer Ranking & k-Reciprocal Nearest Filters
Pruning Networks with Cross-Layer Ranking & k-Reciprocal Nearest Filters
Mingbao Lin
Liujuan Cao
Yu-xin Zhang
Ling Shao
Chia-Wen Lin
Rongrong Ji
22
51
0
15 Feb 2022
Vau da muntanialas: Energy-efficient multi-die scalable acceleration of
  RNN inference
Vau da muntanialas: Energy-efficient multi-die scalable acceleration of RNN inference
G. Paulin
Francesco Conti
Lukas Cavigelli
Luca Benini
22
8
0
14 Feb 2022
F8Net: Fixed-Point 8-bit Only Multiplication for Network Quantization
F8Net: Fixed-Point 8-bit Only Multiplication for Network Quantization
Qing Jin
Jian Ren
Richard Zhuang
Sumant Hanumante
Zhengang Li
Zhiyu Chen
Yanzhi Wang
Kai-Min Yang
Sergey Tulyakov
MQ
24
48
0
10 Feb 2022
Robust Binary Models by Pruning Randomly-initialized Networks
Robust Binary Models by Pruning Randomly-initialized Networks
Chen Liu
Ziqi Zhao
Sabine Süsstrunk
Mathieu Salzmann
TPM
AAML
MQ
19
4
0
03 Feb 2022
Post-training Quantization for Neural Networks with Provable Guarantees
Post-training Quantization for Neural Networks with Provable Guarantees
Jinjie Zhang
Yixuan Zhou
Rayan Saab
MQ
23
31
0
26 Jan 2022
AutoMC: Automated Model Compression based on Domain Knowledge and
  Progressive search strategy
AutoMC: Automated Model Compression based on Domain Knowledge and Progressive search strategy
Chunnan Wang
Hongzhi Wang
Xiangyu Shi
19
0
0
24 Jan 2022
UDC: Unified DNAS for Compressible TinyML Models
UDC: Unified DNAS for Compressible TinyML Models
Igor Fedorov
Ramon Matas
Hokchhay Tann
Chu Zhou
Matthew Mattina
P. Whatmough
AI4CE
21
13
0
15 Jan 2022
Sub-mW Keyword Spotting on an MCU: Analog Binary Feature Extraction and
  Binary Neural Networks
Sub-mW Keyword Spotting on an MCU: Analog Binary Feature Extraction and Binary Neural Networks
G. Cerutti
Lukas Cavigelli
Renzo Andri
Michele Magno
Elisabetta Farella
Luca Benini
26
14
0
10 Jan 2022
Finding the Task-Optimal Low-Bit Sub-Distribution in Deep Neural
  Networks
Finding the Task-Optimal Low-Bit Sub-Distribution in Deep Neural Networks
Runpei Dong
Zhanhong Tan
Mengdi Wu
Linfeng Zhang
Kaisheng Ma
MQ
33
11
0
30 Dec 2021
BMPQ: Bit-Gradient Sensitivity Driven Mixed-Precision Quantization of
  DNNs from Scratch
BMPQ: Bit-Gradient Sensitivity Driven Mixed-Precision Quantization of DNNs from Scratch
Souvik Kundu
Shikai Wang
Qirui Sun
P. Beerel
Massoud Pedram
MQ
15
18
0
24 Dec 2021
SoK: Privacy-preserving Deep Learning with Homomorphic Encryption
SoK: Privacy-preserving Deep Learning with Homomorphic Encryption
Robert Podschwadt
Daniel Takabi
Peizhao Hu
FedML
16
6
0
23 Dec 2021
A Generalized Zero-Shot Quantization of Deep Convolutional Neural
  Networks via Learned Weights Statistics
A Generalized Zero-Shot Quantization of Deep Convolutional Neural Networks via Learned Weights Statistics
Prasen Kumar Sharma
Arun Abraham
V. N. Rajendiran
MQ
25
7
0
06 Dec 2021
Finding Deviated Behaviors of the Compressed DNN Models for Image
  Classifications
Finding Deviated Behaviors of the Compressed DNN Models for Image Classifications
Yongqiang Tian
Wuqi Zhang
Ming Wen
S. Cheung
Chengnian Sun
Shiqing Ma
Yu Jiang
18
6
0
06 Dec 2021
Mixed Precision of Quantization of Transformer Language Models for
  Speech Recognition
Mixed Precision of Quantization of Transformer Language Models for Speech Recognition
Junhao Xu
Shoukang Hu
Jianwei Yu
Xunying Liu
Helen M. Meng
MQ
35
15
0
29 Nov 2021
Toward Compact Parameter Representations for Architecture-Agnostic
  Neural Network Compression
Toward Compact Parameter Representations for Architecture-Agnostic Neural Network Compression
Yuezhou Sun
Wenlong Zhao
Lijun Zhang
Xiao Liu
Hui Guan
Matei A. Zaharia
23
0
0
19 Nov 2021
Iterative Training: Finding Binary Weight Deep Neural Networks with
  Layer Binarization
Iterative Training: Finding Binary Weight Deep Neural Networks with Layer Binarization
Cheng-Chou Lan
MQ
14
0
0
13 Nov 2021
SEOFP-NET: Compression and Acceleration of Deep Neural Networks for
  Speech Enhancement Using Sign-Exponent-Only Floating-Points
SEOFP-NET: Compression and Acceleration of Deep Neural Networks for Speech Enhancement Using Sign-Exponent-Only Floating-Points
Yu-Chen Lin
Cheng Yu
Y. Hsu
Szu-Wei Fu
Yu Tsao
Tei-Wei Kuo
14
6
0
08 Nov 2021
RMSMP: A Novel Deep Neural Network Quantization Framework with Row-wise
  Mixed Schemes and Multiple Precisions
RMSMP: A Novel Deep Neural Network Quantization Framework with Row-wise Mixed Schemes and Multiple Precisions
Sung-En Chang
Yanyu Li
Mengshu Sun
Weiwen Jiang
Sijia Liu
Yanzhi Wang
Xue Lin
MQ
8
10
0
30 Oct 2021
Demystifying and Generalizing BinaryConnect
Demystifying and Generalizing BinaryConnect
Abhishek Sharma
Yaoliang Yu
Eyyub Sari
Mahdi Zolnouri
V. Nia
MQ
17
8
0
25 Oct 2021
Applications and Techniques for Fast Machine Learning in Science
Applications and Techniques for Fast Machine Learning in Science
A. Deiana
Nhan Tran
Joshua C. Agar
Michaela Blott
G. D. Guglielmo
...
Ashish Sharma
S. Summers
Pietro Vischia
J. Vlimant
Olivia Weng
11
71
0
25 Oct 2021
Sub-bit Neural Networks: Learning to Compress and Accelerate Binary
  Neural Networks
Sub-bit Neural Networks: Learning to Compress and Accelerate Binary Neural Networks
Yikai Wang
Yi Yang
Fuchun Sun
Anbang Yao
MQ
23
15
0
18 Oct 2021
Towards Mixed-Precision Quantization of Neural Networks via Constrained
  Optimization
Towards Mixed-Precision Quantization of Neural Networks via Constrained Optimization
Weihan Chen
Peisong Wang
Jian Cheng
MQ
42
61
0
13 Oct 2021
CBP: Backpropagation with constraint on weight precision using a
  pseudo-Lagrange multiplier method
CBP: Backpropagation with constraint on weight precision using a pseudo-Lagrange multiplier method
Guhyun Kim
D. Jeong
MQ
36
2
0
06 Oct 2021
SECDA: Efficient Hardware/Software Co-Design of FPGA-based DNN
  Accelerators for Edge Inference
SECDA: Efficient Hardware/Software Co-Design of FPGA-based DNN Accelerators for Edge Inference
Jude Haris
Perry Gibson
José Cano
Nicolas Bohm Agostini
David Kaeli
36
19
0
01 Oct 2021
OMPQ: Orthogonal Mixed Precision Quantization
OMPQ: Orthogonal Mixed Precision Quantization
Yuexiao Ma
Taisong Jin
Xiawu Zheng
Yan Wang
Huixia Li
Yongjian Wu
Guannan Jiang
Wei Zhang
Rongrong Ji
MQ
17
33
0
16 Sep 2021
Elastic Significant Bit Quantization and Acceleration for Deep Neural
  Networks
Elastic Significant Bit Quantization and Acceleration for Deep Neural Networks
Cheng Gong
Ye Lu
Kunpeng Xie
Zongming Jin
Tao Li
Yanzhi Wang
MQ
25
7
0
08 Sep 2021
Efficient Visual Recognition with Deep Neural Networks: A Survey on
  Recent Advances and New Directions
Efficient Visual Recognition with Deep Neural Networks: A Survey on Recent Advances and New Directions
Yang Wu
Dingheng Wang
Xiaotong Lu
Fan Yang
Guoqi Li
W. Dong
Jianbo Shi
29
18
0
30 Aug 2021
Compact representations of convolutional neural networks via weight
  pruning and quantization
Compact representations of convolutional neural networks via weight pruning and quantization
Giosuè Cataldo Marinò
A. Petrini
D. Malchiodi
Marco Frasca
MQ
19
4
0
28 Aug 2021
Controlled GAN-Based Creature Synthesis via a Challenging Game Art
  Dataset -- Addressing the Noise-Latent Trade-Off
Controlled GAN-Based Creature Synthesis via a Challenging Game Art Dataset -- Addressing the Noise-Latent Trade-Off
Vaibhav Vavilala
David A. Forsyth
8
3
0
19 Aug 2021
Distance-aware Quantization
Distance-aware Quantization
Dohyung Kim
Junghyup Lee
Bumsub Ham
MQ
13
28
0
16 Aug 2021
Bias Loss for Mobile Neural Networks
Bias Loss for Mobile Neural Networks
L. Abrahamyan
Valentin Ziatchin
Yiming Chen
Nikos Deligiannis
13
14
0
23 Jul 2021
Model compression as constrained optimization, with application to
  neural nets. Part V: combining compressions
Model compression as constrained optimization, with application to neural nets. Part V: combining compressions
Miguel Á. Carreira-Perpiñán
Yerlan Idelbayev
25
6
0
09 Jul 2021
$S^3$: Sign-Sparse-Shift Reparametrization for Effective Training of
  Low-bit Shift Networks
S3S^3S3: Sign-Sparse-Shift Reparametrization for Effective Training of Low-bit Shift Networks
Xinlin Li
Bang Liu
Yaoliang Yu
Wulong Liu
Chunjing Xu
V. Nia
MQ
28
5
0
07 Jul 2021
Previous
123456...8910
Next