ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1702.03044
  4. Cited By
Incremental Network Quantization: Towards Lossless CNNs with
  Low-Precision Weights

Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights

10 February 2017
Aojun Zhou
Anbang Yao
Yiwen Guo
Lin Xu
Yurong Chen
    MQ
ArXivPDFHTML

Papers citing "Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights"

50 / 464 papers shown
Title
Radio: Rate-Distortion Optimization for Large Language Model Compression
Radio: Rate-Distortion Optimization for Large Language Model Compression
Sean I. Young
MQ
21
0
0
05 May 2025
Pushing the Limits of Low-Bit Optimizers: A Focus on EMA Dynamics
Pushing the Limits of Low-Bit Optimizers: A Focus on EMA Dynamics
Cong Xu
Wenbin Liang
Mo Yu
Anan Liu
K. Zhang
Lizhuang Ma
J. Wang
J. Wang
W. Zhang
MQ
51
0
0
01 May 2025
Precision Neural Network Quantization via Learnable Adaptive Modules
Precision Neural Network Quantization via Learnable Adaptive Modules
Wenqiang Zhou
Zhendong Yu
X. Liu
Jiaming Yang
Rong Xiao
Tao Wang
Chenwei Tang
Jiancheng Lv
MQ
46
0
0
24 Apr 2025
Tin-Tin: Towards Tiny Learning on Tiny Devices with Integer-based Neural Network Training
Tin-Tin: Towards Tiny Learning on Tiny Devices with Integer-based Neural Network Training
Yi Hu
Jinhang Zuo
Eddie Zhang
Bob Iannucci
Carlee Joe-Wong
24
0
0
13 Apr 2025
ViM-VQ: Efficient Post-Training Vector Quantization for Visual Mamba
Juncan Deng
Shuaiting Li
Zeyu Wang
Kedong Xu
Hong Gu
Kejie Huang
MQ
60
0
0
12 Mar 2025
Verification of Bit-Flip Attacks against Quantized Neural Networks
Verification of Bit-Flip Attacks against Quantized Neural Networks
Yedi Zhang
Lei Huang
Pengfei Gao
Fu Song
Jun Sun
Jin Song Dong
AAML
47
0
0
22 Feb 2025
GSQ-Tuning: Group-Shared Exponents Integer in Fully Quantized Training for LLMs On-Device Fine-tuning
GSQ-Tuning: Group-Shared Exponents Integer in Fully Quantized Training for LLMs On-Device Fine-tuning
Sifan Zhou
Shuo Wang
Zhihang Yuan
Mingjia Shi
Yuzhang Shang
Dawei Yang
ALM
MQ
85
0
0
18 Feb 2025
Forget the Data and Fine-Tuning! Just Fold the Network to Compress
Forget the Data and Fine-Tuning! Just Fold the Network to Compress
Dong Wang
Haris Šikić
Lothar Thiele
O. Saukh
48
0
0
17 Feb 2025
Fast Matrix Multiplications for Lookup Table-Quantized LLMs
Fast Matrix Multiplications for Lookup Table-Quantized LLMs
Han Guo
William Brandon
Radostin Cholakov
Jonathan Ragan-Kelley
Eric P. Xing
Yoon Kim
MQ
83
12
0
20 Jan 2025
Improving Quantization-aware Training of Low-Precision Network via Block
  Replacement on Full-Precision Counterpart
Improving Quantization-aware Training of Low-Precision Network via Block Replacement on Full-Precision Counterpart
Chengting Yu
Shu Yang
Fengzhao Zhang
Hanzhi Ma
Aili Wang
Er-ping Li
MQ
77
2
0
20 Dec 2024
Data Generation for Hardware-Friendly Post-Training Quantization
Data Generation for Hardware-Friendly Post-Training Quantization
Lior Dikstein
Ariel Lapid
Arnon Netzer
H. Habi
MQ
136
0
0
29 Oct 2024
QT-DoG: Quantization-aware Training for Domain Generalization
QT-DoG: Quantization-aware Training for Domain Generalization
Saqib Javed
Hieu Le
Mathieu Salzmann
OOD
MQ
28
1
0
08 Oct 2024
Accelerating PoT Quantization on Edge Devices
Accelerating PoT Quantization on Edge Devices
Rappy Saha
Jude Haris
José Cano
MQ
18
0
0
30 Sep 2024
InfantCryNet: A Data-driven Framework for Intelligent Analysis of Infant Cries
InfantCryNet: A Data-driven Framework for Intelligent Analysis of Infant Cries
Mengze Hong
Chen Jason Zhang
Lingxiao Yang
Yuanfeng Song
Di Jiang
39
2
0
29 Sep 2024
FSL-HDnn: A 5.7 TOPS/W End-to-end Few-shot Learning Classifier
  Accelerator with Feature Extraction and Hyperdimensional Computing
FSL-HDnn: A 5.7 TOPS/W End-to-end Few-shot Learning Classifier Accelerator with Feature Extraction and Hyperdimensional Computing
Haichao Yang
Chang Eun Song
Weihong Xu
Behnam Khaleghi
Uday Mallappa
Monil Shah
Keming Fan
Mingu Kang
Tajana Rosing
19
1
0
17 Sep 2024
Foundations of Large Language Model Compression -- Part 1: Weight
  Quantization
Foundations of Large Language Model Compression -- Part 1: Weight Quantization
Sean I. Young
MQ
40
1
0
03 Sep 2024
1-Bit FQT: Pushing the Limit of Fully Quantized Training to 1-bit
1-Bit FQT: Pushing the Limit of Fully Quantized Training to 1-bit
Chang Gao
J. Chen
Kang Zhao
Jiaqi Wang
Liping Jing
MQ
38
2
0
26 Aug 2024
CCSRP: Robust Pruning of Spiking Neural Networks through Cooperative
  Coevolution
CCSRP: Robust Pruning of Spiking Neural Networks through Cooperative Coevolution
J. Reif
Jiakang Li
Songning Lai
Alexander Fay
AAML
27
0
0
18 Jul 2024
Tiled Bit Networks: Sub-Bit Neural Network Compression Through Reuse of
  Learnable Binary Vectors
Tiled Bit Networks: Sub-Bit Neural Network Compression Through Reuse of Learnable Binary Vectors
Matt Gorbett
Hossein Shirazi
Indrakshi Ray
MQ
43
0
0
16 Jul 2024
Quality Scalable Quantization Methodology for Deep Learning on Edge
Quality Scalable Quantization Methodology for Deep Learning on Edge
S. Khaliq
Rehan Hafiz
MQ
35
1
0
15 Jul 2024
Towards Lightweight Speaker Verification via Adaptive Neural Network
  Quantization
Towards Lightweight Speaker Verification via Adaptive Neural Network Quantization
Bei Liu
Haoyu Wang
Yanmin Qian
MQ
28
0
0
08 Jun 2024
ReDistill: Residual Encoded Distillation for Peak Memory Reduction of CNNs
ReDistill: Residual Encoded Distillation for Peak Memory Reduction of CNNs
Fang Chen
Gourav Datta
Mujahid Al Rafi
Hyeran Jeon
Meng Tang
91
1
0
06 Jun 2024
From Algorithm to Hardware: A Survey on Efficient and Safe Deployment of
  Deep Neural Networks
From Algorithm to Hardware: A Survey on Efficient and Safe Deployment of Deep Neural Networks
Xue Geng
Zhe Wang
Chunyun Chen
Qing Xu
Kaixin Xu
...
Zhenghua Chen
M. Aly
Jie Lin
Min-man Wu
Xiaoli Li
33
1
0
09 May 2024
Enhancing User Experience in On-Device Machine Learning with Gated
  Compression Layers
Enhancing User Experience in On-Device Machine Learning with Gated Compression Layers
Haiguang Li
Usama Pervaiz
Joseph Antognini
Michal Matuszak
Lawrence Au
Gilles Roux
T. Thormundsson
36
0
0
02 May 2024
Towards Green AI: Current status and future research
Towards Green AI: Current status and future research
Christian Clemm
Lutz Stobbe
Kishan Wimalawarne
Jan Druschke
45
2
0
01 May 2024
QGen: On the Ability to Generalize in Quantization Aware Training
QGen: On the Ability to Generalize in Quantization Aware Training
Mohammadhossein Askarihemmat
Ahmadreza Jeddi
Reyhane Askari Hemmat
Ivan Lazarevich
Alexander Hoffman
Sudhakar Sah
Ehsan Saboori
Yvon Savaria
Jean-Pierre David
MQ
21
0
0
17 Apr 2024
Dynamic Switch Layers For Unsupervised Learning
Dynamic Switch Layers For Unsupervised Learning
Haiguang Li
Usama Pervaiz
Michal Matuszak
Robert Kamara
Gilles Roux
T. Thormundsson
Joseph Antognini
50
1
0
05 Apr 2024
Instance-Aware Group Quantization for Vision Transformers
Instance-Aware Group Quantization for Vision Transformers
Jaehyeon Moon
Dohyung Kim
Junyong Cheon
Bumsub Ham
MQ
ViT
27
6
0
01 Apr 2024
A&B BNN: Add&Bit-Operation-Only Hardware-Friendly Binary Neural Network
A&B BNN: Add&Bit-Operation-Only Hardware-Friendly Binary Neural Network
Ruichen Ma
G. Qiao
Yián Liu
L. Meng
N. Ning
Yang Liu
Shaogang Hu
AAML
MQ
28
3
0
06 Mar 2024
Compression Repair for Feedforward Neural Networks Based on Model
  Equivalence Evaluation
Compression Repair for Feedforward Neural Networks Based on Model Equivalence Evaluation
Zihao Mo
Yejiang Yang
Shuaizheng Lu
Weiming Xiang
32
1
0
18 Feb 2024
Effect of Weight Quantization on Learning Models by Typical Case
  Analysis
Effect of Weight Quantization on Learning Models by Typical Case Analysis
Shuhei Kashiwamura
Ayaka Sakata
Masaaki Imaizumi
MQ
22
1
0
30 Jan 2024
Model Compression Techniques in Biometrics Applications: A Survey
Model Compression Techniques in Biometrics Applications: A Survey
Eduarda Caldeira
Pedro C. Neto
Marco Huber
Naser Damer
Ana F. Sequeira
32
11
0
18 Jan 2024
Convolutional Neural Network Compression via Dynamic Parameter Rank
  Pruning
Convolutional Neural Network Compression via Dynamic Parameter Rank Pruning
Manish Sharma
Jamison Heard
Eli Saber
Panos P. Markopoulos
23
1
0
15 Jan 2024
RACE-IT: A Reconfigurable Analog CAM-Crossbar Engine for In-Memory
  Transformer Acceleration
RACE-IT: A Reconfigurable Analog CAM-Crossbar Engine for In-Memory Transformer Acceleration
Lei Zhao
Luca Buonanno
Ron M. Roth
Sergey Serebryakov
Archit Gajjar
John Moon
Jim Ignowski
Giacomo Pedretti
23
3
0
29 Nov 2023
Relationship between Model Compression and Adversarial Robustness: A
  Review of Current Evidence
Relationship between Model Compression and Adversarial Robustness: A Review of Current Evidence
Svetlana Pavlitska
Hannes Grolig
J. Marius Zöllner
AAML
16
3
0
27 Nov 2023
Efficient Neural Networks for Tiny Machine Learning: A Comprehensive
  Review
Efficient Neural Networks for Tiny Machine Learning: A Comprehensive Review
M. Lê
Pierre Wolinski
Julyan Arbel
32
8
0
20 Nov 2023
Reducing the Side-Effects of Oscillations in Training of Quantized YOLO
  Networks
Reducing the Side-Effects of Oscillations in Training of Quantized YOLO Networks
Kartik Gupta
Akshay Asthana
MQ
24
8
0
09 Nov 2023
FedAIoT: A Federated Learning Benchmark for Artificial Intelligence of
  Things
FedAIoT: A Federated Learning Benchmark for Artificial Intelligence of Things
Samiul Alam
Tuo Zhang
Tiantian Feng
Hui Shen
Zhichao Cao
...
JeongGil Ko
Kiran Somasundaram
Shrikanth S. Narayanan
Salman Avestimehr
Mi Zhang
25
11
0
29 Sep 2023
Efficient Post-training Quantization with FP8 Formats
Efficient Post-training Quantization with FP8 Formats
Haihao Shen
Naveen Mellempudi
Xin He
Q. Gao
Chang‐Bao Wang
Mengni Wang
MQ
23
19
0
26 Sep 2023
On Calibration of Modern Quantized Efficient Neural Networks
On Calibration of Modern Quantized Efficient Neural Networks
Joe-Hwa Kuang
Alexander Wong
UQCV
MQ
16
1
0
25 Sep 2023
Probabilistic Weight Fixing: Large-scale training of neural network
  weight uncertainties for quantization
Probabilistic Weight Fixing: Large-scale training of neural network weight uncertainties for quantization
Christopher Subia-Waud
S. Dasmahapatra
UQCV
MQ
13
0
0
24 Sep 2023
SPFQ: A Stochastic Algorithm and Its Error Analysis for Neural Network
  Quantization
SPFQ: A Stochastic Algorithm and Its Error Analysis for Neural Network Quantization
Jinjie Zhang
Rayan Saab
16
0
0
20 Sep 2023
Towards Artificial General Intelligence (AGI) in the Internet of Things
  (IoT): Opportunities and Challenges
Towards Artificial General Intelligence (AGI) in the Internet of Things (IoT): Opportunities and Challenges
Fei Dou
Jin Ye
Geng Yuan
Qin Lu
Wei Niu
...
Hongyue Sun
Yunli Shao
Changying Li
Tianming Liu
Wenzhan Song
AI4CE
23
29
0
14 Sep 2023
QD-BEV : Quantization-aware View-guided Distillation for Multi-view 3D
  Object Detection
QD-BEV : Quantization-aware View-guided Distillation for Multi-view 3D Object Detection
Yifan Zhang
Zhen Dong
Huanrui Yang
Ming Lu
Cheng-Ching Tseng
Yuan Du
Kurt Keutzer
Li Du
Shanghang Zhang
MQ
32
9
0
21 Aug 2023
NUPES : Non-Uniform Post-Training Quantization via Power Exponent Search
NUPES : Non-Uniform Post-Training Quantization via Power Exponent Search
Edouard Yvinec
Arnaud Dapogny
Kévin Bailly
MQ
24
6
0
10 Aug 2023
Self-Distilled Quantization: Achieving High Compression Rates in
  Transformer-Based Language Models
Self-Distilled Quantization: Achieving High Compression Rates in Transformer-Based Language Models
James OÑeill
Sourav Dutta
VLM
MQ
32
1
0
12 Jul 2023
Free Bits: Latency Optimization of Mixed-Precision Quantized Neural
  Networks on the Edge
Free Bits: Latency Optimization of Mixed-Precision Quantized Neural Networks on the Edge
Georg Rutishauser
Francesco Conti
Luca Benini
MQ
20
5
0
06 Jul 2023
Minimizing Energy Consumption of Deep Learning Models by Energy-Aware
  Training
Minimizing Energy Consumption of Deep Learning Models by Energy-Aware Training
Dario Lazzaro
Antonio Emanuele Cinà
Maura Pintor
Ambra Demontis
Battista Biggio
Fabio Roli
Marcello Pelillo
27
6
0
01 Jul 2023
Explainable Lifelong Stream Learning Based on "Glocal" Pairwise Fusion
Explainable Lifelong Stream Learning Based on "Glocal" Pairwise Fusion
C. K. Loo
W. S. Liew
S. Wermter
CLL
11
0
0
23 Jun 2023
Training Transformers with 4-bit Integers
Training Transformers with 4-bit Integers
Haocheng Xi
Changhao Li
Jianfei Chen
Jun Zhu
MQ
25
47
0
21 Jun 2023
1234...8910
Next