ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1612.01064
  4. Cited By
Trained Ternary Quantization

Trained Ternary Quantization

4 December 2016
Chenzhuo Zhu
Song Han
Huizi Mao
W. Dally
    MQ
ArXivPDFHTML

Papers citing "Trained Ternary Quantization"

50 / 509 papers shown
Title
Data Quality-aware Mixed-precision Quantization via Hybrid Reinforcement
  Learning
Data Quality-aware Mixed-precision Quantization via Hybrid Reinforcement Learning
Yingchun Wang
Jingcai Guo
Song Guo
Weizhan Zhang
MQ
37
21
0
09 Feb 2023
Learning Discretized Neural Networks under Ricci Flow
Learning Discretized Neural Networks under Ricci Flow
Jun Chen
Han Chen
Mengmeng Wang
Guang Dai
Ivor W. Tsang
Yong-Jin Liu
35
2
0
07 Feb 2023
Efficient and Effective Methods for Mixed Precision Neural Network
  Quantization for Faster, Energy-efficient Inference
Efficient and Effective Methods for Mixed Precision Neural Network Quantization for Faster, Energy-efficient Inference
Deepika Bablani
J. McKinstry
S. K. Esser
R. Appuswamy
D. Modha
MQ
23
4
0
30 Jan 2023
RedBit: An End-to-End Flexible Framework for Evaluating the Accuracy of
  Quantized CNNs
RedBit: An End-to-End Flexible Framework for Evaluating the Accuracy of Quantized CNNs
A. M. Ribeiro-dos-Santos
João Dinis Ferreira
O. Mutlu
G. Falcão
MQ
21
1
0
15 Jan 2023
Holistic Network Virtualization and Pervasive Network Intelligence for
  6G
Holistic Network Virtualization and Pervasive Network Intelligence for 6G
Xuemin Shen
Shen
Jie Gao
Wen Wu
Mushu Li
Conghao Zhou
W. Zhuang
35
234
0
02 Jan 2023
Hyperspherical Quantization: Toward Smaller and More Accurate Models
Hyperspherical Quantization: Toward Smaller and More Accurate Models
Dan Liu
X. Chen
Chen Ma
Xue Liu
MQ
35
3
0
24 Dec 2022
Hyperspherical Loss-Aware Ternary Quantization
Hyperspherical Loss-Aware Ternary Quantization
Dan Liu
Xue Liu
MQ
27
0
0
24 Dec 2022
Masked Wavelet Representation for Compact Neural Radiance Fields
Masked Wavelet Representation for Compact Neural Radiance Fields
Daniel Rho
Byeonghyeon Lee
Seungtae Nam
J. Lee
J. Ko
Eunbyung Park
49
52
0
18 Dec 2022
Towards Hardware-Specific Automatic Compression of Neural Networks
Towards Hardware-Specific Automatic Compression of Neural Networks
Torben Krieger
Bernhard Klein
Holger Fröning
MQ
32
2
0
15 Dec 2022
Vertical Layering of Quantized Neural Networks for Heterogeneous
  Inference
Vertical Layering of Quantized Neural Networks for Heterogeneous Inference
Hai Wu
Ruifei He
Hao Hao Tan
Xiaojuan Qi
Kaibin Huang
MQ
37
2
0
10 Dec 2022
BiFSMNv2: Pushing Binary Neural Networks for Keyword Spotting to
  Real-Network Performance
BiFSMNv2: Pushing Binary Neural Networks for Keyword Spotting to Real-Network Performance
Haotong Qin
Xudong Ma
Yifu Ding
Xiaochen Li
Yang Zhang
Zejun Ma
Jiakai Wang
Jie Luo
Xianglong Liu
MQ
40
20
0
13 Nov 2022
AskewSGD : An Annealed interval-constrained Optimisation method to train
  Quantized Neural Networks
AskewSGD : An Annealed interval-constrained Optimisation method to train Quantized Neural Networks
Louis Leconte
S. Schechtman
Eric Moulines
31
4
0
07 Nov 2022
Collaborative Multi-Teacher Knowledge Distillation for Learning Low
  Bit-width Deep Neural Networks
Collaborative Multi-Teacher Knowledge Distillation for Learning Low Bit-width Deep Neural Networks
Cuong Pham
Tuan Hoang
Thanh-Toan Do
FedML
MQ
40
14
0
27 Oct 2022
Q-ViT: Accurate and Fully Quantized Low-bit Vision Transformer
Q-ViT: Accurate and Fully Quantized Low-bit Vision Transformer
Yanjing Li
Sheng Xu
Baochang Zhang
Xianbin Cao
Penglei Gao
Guodong Guo
MQ
ViT
34
89
0
13 Oct 2022
Structural Pruning via Latency-Saliency Knapsack
Structural Pruning via Latency-Saliency Knapsack
Maying Shen
Hongxu Yin
Pavlo Molchanov
Lei Mao
Jianna Liu
J. Álvarez
37
48
0
13 Oct 2022
SeKron: A Decomposition Method Supporting Many Factorization Structures
SeKron: A Decomposition Method Supporting Many Factorization Structures
Marawan Gamal Abdel Hameed
A. Mosleh
Marzieh S. Tahaei
V. Nia
29
1
0
12 Oct 2022
Seeking Interpretability and Explainability in Binary Activated Neural
  Networks
Seeking Interpretability and Explainability in Binary Activated Neural Networks
Benjamin Leblanc
Pascal Germain
FAtt
42
1
0
07 Sep 2022
DenseShift: Towards Accurate and Efficient Low-Bit Power-of-Two
  Quantization
DenseShift: Towards Accurate and Efficient Low-Bit Power-of-Two Quantization
Xinlin Li
Bangya Liu
Ruizhi Yang
Vanessa Courville
Chao Xing
V. Nia
MQ
34
2
0
20 Aug 2022
LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale
LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale
Tim Dettmers
M. Lewis
Younes Belkada
Luke Zettlemoyer
MQ
43
637
0
15 Aug 2022
Mixed-Precision Neural Networks: A Survey
Mixed-Precision Neural Networks: A Survey
M. Rakka
M. Fouda
Pramod P. Khargonekar
Fadi J. Kurdahi
MQ
30
11
0
11 Aug 2022
Model Blending for Text Classification
Model Blending for Text Classification
Ramit Pahwa
26
0
0
05 Aug 2022
PalQuant: Accelerating High-precision Networks on Low-precision
  Accelerators
PalQuant: Accelerating High-precision Networks on Low-precision Accelerators
Qinghao Hu
Gang Li
Qiman Wu
Jian Cheng
MQ
31
2
0
03 Aug 2022
CoNLoCNN: Exploiting Correlation and Non-Uniform Quantization for
  Energy-Efficient Low-precision Deep Convolutional Neural Networks
CoNLoCNN: Exploiting Correlation and Non-Uniform Quantization for Energy-Efficient Low-precision Deep Convolutional Neural Networks
Muhammad Abdullah Hanif
G. M. Sarda
Alberto Marchisio
Guido Masera
Maurizio Martina
Mohamed Bennai
MQ
35
4
0
31 Jul 2022
Quantized Sparse Weight Decomposition for Neural Network Compression
Quantized Sparse Weight Decomposition for Neural Network Compression
Andrey Kuzmin
M. V. Baalen
Markus Nagel
Arash Behboodi
MQ
19
3
0
22 Jul 2022
Communication Acceleration of Local Gradient Methods via an Accelerated
  Primal-Dual Algorithm with Inexact Prox
Communication Acceleration of Local Gradient Methods via an Accelerated Primal-Dual Algorithm with Inexact Prox
Abdurakhmon Sadiev
D. Kovalev
Peter Richtárik
35
20
0
08 Jul 2022
Compilation and Optimizations for Efficient Machine Learning on Embedded
  Systems
Compilation and Optimizations for Efficient Machine Learning on Embedded Systems
Xiaofan Zhang
Yao Chen
Cong Hao
Sitao Huang
Yuhong Li
Deming Chen
42
1
0
06 Jun 2022
DepthShrinker: A New Compression Paradigm Towards Boosting Real-Hardware Efficiency of Compact Neural Networks
DepthShrinker: A New Compression Paradigm Towards Boosting Real-Hardware Efficiency of Compact Neural Networks
Y. Fu
Haichuan Yang
Jiayi Yuan
Meng Li
Cheng Wan
Raghuraman Krishnamoorthi
Vikas Chandra
Yingyan Lin
38
19
0
02 Jun 2022
Gator: Customizable Channel Pruning of Neural Networks with Gating
Gator: Customizable Channel Pruning of Neural Networks with Gating
E. Passov
E. David
N. Netanyahu
AAML
45
0
0
30 May 2022
A Comprehensive Survey on Model Quantization for Deep Neural Networks in
  Image Classification
A Comprehensive Survey on Model Quantization for Deep Neural Networks in Image Classification
Babak Rokh
A. Azarpeyvand
Alireza Khanteymoori
MQ
40
86
0
14 May 2022
Revisiting Random Channel Pruning for Neural Network Compression
Revisiting Random Channel Pruning for Neural Network Compression
Yawei Li
Kamil Adamczewski
Wen Li
Shuhang Gu
Radu Timofte
Luc Van Gool
37
84
0
11 May 2022
Lite Pose: Efficient Architecture Design for 2D Human Pose Estimation
Lite Pose: Efficient Architecture Design for 2D Human Pose Estimation
Yihan Wang
Zhekai Zhang
Han Cai
Wei-Ming Chen
Song Han
3DH
24
72
0
03 May 2022
Enable Deep Learning on Mobile Devices: Methods, Systems, and
  Applications
Enable Deep Learning on Mobile Devices: Methods, Systems, and Applications
Han Cai
Ji Lin
Chengyue Wu
Zhijian Liu
Haotian Tang
Hanrui Wang
Ligeng Zhu
Song Han
27
108
0
25 Apr 2022
HCFL: A High Compression Approach for Communication-Efficient Federated
  Learning in Very Large Scale IoT Networks
HCFL: A High Compression Approach for Communication-Efficient Federated Learning in Very Large Scale IoT Networks
Minh-Duong Nguyen
Sangmin Lee
Viet Quoc Pham
D. Hoang
Diep N. Nguyen
W. Hwang
28
28
0
14 Apr 2022
LilNetX: Lightweight Networks with EXtreme Model Compression and
  Structured Sparsification
LilNetX: Lightweight Networks with EXtreme Model Compression and Structured Sparsification
Sharath Girish
Kamal Gupta
Saurabh Singh
Abhinav Shrivastava
38
11
0
06 Apr 2022
Soft Threshold Ternary Networks
Soft Threshold Ternary Networks
Weixiang Xu
Xiangyu He
Tianli Zhao
Qinghao Hu
Peisong Wang
Jian Cheng
MQ
22
7
0
04 Apr 2022
FxP-QNet: A Post-Training Quantizer for the Design of Mixed
  Low-Precision DNNs with Dynamic Fixed-Point Representation
FxP-QNet: A Post-Training Quantizer for the Design of Mixed Low-Precision DNNs with Dynamic Fixed-Point Representation
Ahmad Shawahna
S. M. Sait
A. El-Maleh
Irfan Ahmad
MQ
20
7
0
22 Mar 2022
Learning Compressed Embeddings for On-Device Inference
Learning Compressed Embeddings for On-Device Inference
Niketan Pansare
J. Katukuri
Aditya Arora
F. Cipollone
R. Shaik
Noyan Tokgozoglu
Chandru Venkataraman
37
14
0
18 Mar 2022
Hardware Approximate Techniques for Deep Neural Network Accelerators: A
  Survey
Hardware Approximate Techniques for Deep Neural Network Accelerators: A Survey
Giorgos Armeniakos
Georgios Zervakis
Dimitrios Soudris
J. Henkel
217
94
0
16 Mar 2022
YONO: Modeling Multiple Heterogeneous Neural Networks on
  Microcontrollers
YONO: Modeling Multiple Heterogeneous Neural Networks on Microcontrollers
Young D. Kwon
Jagmohan Chauhan
Cecilia Mascolo
24
13
0
08 Mar 2022
Distilled Neural Networks for Efficient Learning to Rank
Distilled Neural Networks for Efficient Learning to Rank
F. M. Nardini
Cosimo Rulli
Salvatore Trani
Rossano Venturini
FedML
29
16
0
22 Feb 2022
Bit-wise Training of Neural Network Weights
Bit-wise Training of Neural Network Weights
Cristian Ivan
MQ
18
0
0
19 Feb 2022
Vau da muntanialas: Energy-efficient multi-die scalable acceleration of
  RNN inference
Vau da muntanialas: Energy-efficient multi-die scalable acceleration of RNN inference
G. Paulin
Francesco Conti
Lukas Cavigelli
Luca Benini
29
8
0
14 Feb 2022
F8Net: Fixed-Point 8-bit Only Multiplication for Network Quantization
F8Net: Fixed-Point 8-bit Only Multiplication for Network Quantization
Qing Jin
Jian Ren
Richard Zhuang
Sumant Hanumante
Zhengang Li
Zhiyu Chen
Yanzhi Wang
Kai-Min Yang
Sergey Tulyakov
MQ
32
50
0
10 Feb 2022
Lightweight Jet Reconstruction and Identification as an Object Detection
  Task
Lightweight Jet Reconstruction and Identification as an Object Detection Task
Adrian Alan Pol
T. Aarrestad
E. Govorkova
Roi Halily
Anat Klempner
...
Vladimir Loncar
J. Ngadiuba
M. Pierini
Olya Sirkin
S. Summers
32
2
0
09 Feb 2022
FAT: An In-Memory Accelerator with Fast Addition for Ternary Weight
  Neural Networks
FAT: An In-Memory Accelerator with Fast Addition for Ternary Weight Neural Networks
Shien Zhu
Luan H. K. Duong
Hui Chen
Di Liu
Weichen Liu
MQ
24
5
0
19 Jan 2022
PocketNN: Integer-only Training and Inference of Neural Networks via
  Direct Feedback Alignment and Pocket Activations in Pure C++
PocketNN: Integer-only Training and Inference of Neural Networks via Direct Feedback Alignment and Pocket Activations in Pure C++
Jae-Su Song
Fangzhen Lin
MQ
7
7
0
08 Jan 2022
Finding the Task-Optimal Low-Bit Sub-Distribution in Deep Neural
  Networks
Finding the Task-Optimal Low-Bit Sub-Distribution in Deep Neural Networks
Runpei Dong
Zhanhong Tan
Mengdi Wu
Linfeng Zhang
Kaisheng Ma
MQ
41
11
0
30 Dec 2021
Resource-Efficient Deep Learning: A Survey on Model-, Arithmetic-, and
  Implementation-Level Techniques
Resource-Efficient Deep Learning: A Survey on Model-, Arithmetic-, and Implementation-Level Techniques
JunKyu Lee
L. Mukhanov
A. S. Molahosseini
U. Minhas
Yang Hua
Jesus Martinez del Rincon
K. Dichev
Cheol-Ho Hong
Hans Vandierendonck
44
29
0
30 Dec 2021
BMPQ: Bit-Gradient Sensitivity Driven Mixed-Precision Quantization of
  DNNs from Scratch
BMPQ: Bit-Gradient Sensitivity Driven Mixed-Precision Quantization of DNNs from Scratch
Souvik Kundu
Shikai Wang
Qirui Sun
P. Beerel
Massoud Pedram
MQ
29
18
0
24 Dec 2021
Elastic-Link for Binarized Neural Network
Elastic-Link for Binarized Neural Network
Jie Hu
Ziheng Wu
Vince Tan
Zhilin Lu
Mengze Zeng
Enhua Wu
MQ
30
6
0
19 Dec 2021
Previous
12345...91011
Next