ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1612.01064
  4. Cited By
Trained Ternary Quantization

Trained Ternary Quantization

4 December 2016
Chenzhuo Zhu
Song Han
Huizi Mao
W. Dally
    MQ
ArXivPDFHTML

Papers citing "Trained Ternary Quantization"

50 / 509 papers shown
Title
Neural Network Quantization for Efficient Inference: A Survey
Neural Network Quantization for Efficient Inference: A Survey
Olivia Weng
MQ
28
23
0
08 Dec 2021
Low-bit Quantization of Recurrent Neural Network Language Models Using
  Alternating Direction Methods of Multipliers
Low-bit Quantization of Recurrent Neural Network Language Models Using Alternating Direction Methods of Multipliers
Junhao Xu
Xie Chen
Shoukang Hu
Jianwei Yu
Xunying Liu
Helen Meng
MQ
30
9
0
29 Nov 2021
AUTOKD: Automatic Knowledge Distillation Into A Student Architecture
  Family
AUTOKD: Automatic Knowledge Distillation Into A Student Architecture Family
Roy Henha Eyono
Fabio Maria Carlucci
P. Esperança
Binxin Ru
Phillip Torr
24
3
0
05 Nov 2021
Automatic Sleep Staging of EEG Signals: Recent Development, Challenges,
  and Future Directions
Automatic Sleep Staging of EEG Signals: Recent Development, Challenges, and Future Directions
Huy P Phan
Kaare B. Mikkelsen
19
94
0
03 Nov 2021
Arch-Net: Model Distillation for Architecture Agnostic Model Deployment
Arch-Net: Model Distillation for Architecture Agnostic Model Deployment
Weixin Xu
Zipeng Feng
Shuangkang Fang
Song Yuan
Yi Yang
Shuchang Zhou
MQ
30
1
0
01 Nov 2021
RMSMP: A Novel Deep Neural Network Quantization Framework with Row-wise
  Mixed Schemes and Multiple Precisions
RMSMP: A Novel Deep Neural Network Quantization Framework with Row-wise Mixed Schemes and Multiple Precisions
Sung-En Chang
Yanyu Li
Mengshu Sun
Weiwen Jiang
Sijia Liu
Yanzhi Wang
Xue Lin
MQ
31
10
0
30 Oct 2021
FAST: DNN Training Under Variable Precision Block Floating Point with
  Stochastic Rounding
FAST: DNN Training Under Variable Precision Block Floating Point with Stochastic Rounding
Shanghang Zhang
Bradley McDanel
H. T. Kung
MQ
21
65
0
28 Oct 2021
MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning
MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning
Ji Lin
Wei-Ming Chen
Han Cai
Chuang Gan
Song Han
45
156
0
28 Oct 2021
MERCURY: Accelerating DNN Training By Exploiting Input Similarity
MERCURY: Accelerating DNN Training By Exploiting Input Similarity
Vahid Janfaza
Kevin Weston
Moein Razavi
Shantanu Mandal
Farabi Mahmud
Alex Hilty
A. Muzahid
41
5
0
28 Oct 2021
Demystifying and Generalizing BinaryConnect
Demystifying and Generalizing BinaryConnect
Abhishek Sharma
Yaoliang Yu
Eyyub Sari
Mahdi Zolnouri
V. Nia
MQ
22
8
0
25 Oct 2021
Applications and Techniques for Fast Machine Learning in Science
Applications and Techniques for Fast Machine Learning in Science
A. Deiana
Nhan Tran
Joshua C. Agar
Michaela Blott
G. D. Guglielmo
...
Ashish Sharma
S. Summers
Pietro Vischia
J. Vlimant
Olivia Weng
14
71
0
25 Oct 2021
When to Prune? A Policy towards Early Structural Pruning
When to Prune? A Policy towards Early Structural Pruning
Maying Shen
Pavlo Molchanov
Hongxu Yin
J. Álvarez
VLM
28
53
0
22 Oct 2021
HALP: Hardware-Aware Latency Pruning
HALP: Hardware-Aware Latency Pruning
Maying Shen
Hongxu Yin
Pavlo Molchanov
Lei Mao
Jianna Liu
J. Álvarez
VLM
46
13
0
20 Oct 2021
Sub-bit Neural Networks: Learning to Compress and Accelerate Binary
  Neural Networks
Sub-bit Neural Networks: Learning to Compress and Accelerate Binary Neural Networks
Yikai Wang
Yi Yang
Gang Hua
Anbang Yao
MQ
29
15
0
18 Oct 2021
Finding Everything within Random Binary Networks
Finding Everything within Random Binary Networks
Kartik K. Sreenivasan
Shashank Rajput
Jy-yong Sohn
Dimitris Papailiopoulos
22
10
0
18 Oct 2021
Network Augmentation for Tiny Deep Learning
Network Augmentation for Tiny Deep Learning
Han Cai
Chuang Gan
Ji Lin
Song Han
25
29
0
17 Oct 2021
BNAS v2: Learning Architectures for Binary Networks with Empirical
  Improvements
BNAS v2: Learning Architectures for Binary Networks with Empirical Improvements
Dahyun Kim
Kunal Pratap Singh
Jonghyun Choi
MQ
46
7
0
16 Oct 2021
Towards Mixed-Precision Quantization of Neural Networks via Constrained
  Optimization
Towards Mixed-Precision Quantization of Neural Networks via Constrained Optimization
Weihan Chen
Peisong Wang
Jian Cheng
MQ
46
62
0
13 Oct 2021
8-bit Optimizers via Block-wise Quantization
8-bit Optimizers via Block-wise Quantization
Tim Dettmers
M. Lewis
Sam Shleifer
Luke Zettlemoyer
MQ
34
276
0
06 Oct 2021
CBP: Backpropagation with constraint on weight precision using a
  pseudo-Lagrange multiplier method
CBP: Backpropagation with constraint on weight precision using a pseudo-Lagrange multiplier method
Guhyun Kim
D. Jeong
MQ
55
2
0
06 Oct 2021
Random matrices in service of ML footprint: ternary random features with
  no performance loss
Random matrices in service of ML footprint: ternary random features with no performance loss
Hafiz Tiomoko Ali
Zhenyu Liao
Romain Couillet
44
7
0
05 Oct 2021
Convolutional Neural Network Compression through Generalized Kronecker
  Product Decomposition
Convolutional Neural Network Compression through Generalized Kronecker Product Decomposition
Marawan Gamal Abdel Hameed
Marzieh S. Tahaei
A. Mosleh
V. Nia
47
25
0
29 Sep 2021
TSM: Temporal Shift Module for Efficient and Scalable Video
  Understanding on Edge Device
TSM: Temporal Shift Module for Efficient and Scalable Video Understanding on Edge Device
Ji Lin
Chuang Gan
Kuan-Chieh Jackson Wang
Song Han
42
64
0
27 Sep 2021
Distribution-sensitive Information Retention for Accurate Binary Neural
  Network
Distribution-sensitive Information Retention for Accurate Binary Neural Network
Haotong Qin
Xiangguo Zhang
Ruihao Gong
Yifu Ding
Yi Xu
Xianglong Liu
MQ
25
87
0
25 Sep 2021
ECQ$^{\text{x}}$: Explainability-Driven Quantization for Low-Bit and
  Sparse DNNs
ECQx^{\text{x}}x: Explainability-Driven Quantization for Low-Bit and Sparse DNNs
Daniel Becking
Maximilian Dreyer
Wojciech Samek
Karsten Müller
Sebastian Lapuschkin
MQ
193
13
0
09 Sep 2021
BioNetExplorer: Architecture-Space Exploration of Bio-Signal Processing
  Deep Neural Networks for Wearables
BioNetExplorer: Architecture-Space Exploration of Bio-Signal Processing Deep Neural Networks for Wearables
B. Prabakaran
Asima Akhtar
Semeen Rehman
Osman Hasan
Mohamed Bennai
19
9
0
07 Sep 2021
Efficient Visual Recognition with Deep Neural Networks: A Survey on
  Recent Advances and New Directions
Efficient Visual Recognition with Deep Neural Networks: A Survey on Recent Advances and New Directions
Yang Wu
Dingheng Wang
Xiaotong Lu
Fan Yang
Guoqi Li
W. Dong
Jianbo Shi
29
18
0
30 Aug 2021
Distance-aware Quantization
Distance-aware Quantization
Dohyung Kim
Junghyup Lee
Bumsub Ham
MQ
15
28
0
16 Aug 2021
Generalizable Mixed-Precision Quantization via Attribution Rank
  Preservation
Generalizable Mixed-Precision Quantization via Attribution Rank Preservation
Ziwei Wang
Han Xiao
Jiwen Lu
Jie Zhou
MQ
22
32
0
05 Aug 2021
Pruning Ternary Quantization
Danyang Liu
Xiangshan Chen
Jie Fu
Chen Ma
Xue Liu
MQ
47
0
0
23 Jul 2021
CREW: Computation Reuse and Efficient Weight Storage for
  Hardware-accelerated MLPs and RNNs
CREW: Computation Reuse and Efficient Weight Storage for Hardware-accelerated MLPs and RNNs
Marc Riera
J. Arnau
Antonio González
15
5
0
20 Jul 2021
A High-Performance Adaptive Quantization Approach for Edge CNN
  Applications
A High-Performance Adaptive Quantization Approach for Edge CNN Applications
Hsu-Hsun Chin
R. Tsay
Hsin-I Wu
MQ
24
5
0
18 Jul 2021
Training Compact CNNs for Image Classification using Dynamic-coded
  Filter Fusion
Training Compact CNNs for Image Classification using Dynamic-coded Filter Fusion
Mingbao Lin
Bohong Chen
Rongrong Ji
Rongrong Ji
VLM
35
23
0
14 Jul 2021
LANA: Latency Aware Network Acceleration
LANA: Latency Aware Network Acceleration
Pavlo Molchanov
Jimmy Hall
Hongxu Yin
Jan Kautz
Nicolò Fusi
Arash Vahdat
25
11
0
12 Jul 2021
Model compression as constrained optimization, with application to
  neural nets. Part V: combining compressions
Model compression as constrained optimization, with application to neural nets. Part V: combining compressions
Miguel Á. Carreira-Perpiñán
Yerlan Idelbayev
30
6
0
09 Jul 2021
$S^3$: Sign-Sparse-Shift Reparametrization for Effective Training of
  Low-bit Shift Networks
S3S^3S3: Sign-Sparse-Shift Reparametrization for Effective Training of Low-bit Shift Networks
Xinlin Li
Bang Liu
Yaoliang Yu
Wulong Liu
Chunjing Xu
V. Nia
MQ
38
5
0
07 Jul 2021
Content-Aware Convolutional Neural Networks
Content-Aware Convolutional Neural Networks
Yong Guo
Yaofo Chen
Mingkui Tan
Kui Jia
Jian Chen
Jingdong Wang
36
8
0
30 Jun 2021
Post-Training Quantization for Vision Transformer
Post-Training Quantization for Vision Transformer
Zhenhua Liu
Yunhe Wang
Kai Han
Siwei Ma
Wen Gao
ViT
MQ
56
327
0
27 Jun 2021
Boggart: Towards General-Purpose Acceleration of Retrospective Video
  Analytics
Boggart: Towards General-Purpose Acceleration of Retrospective Video Analytics
Neil Agarwal
Ravi Netravali
32
14
0
21 Jun 2021
An Empirical Investigation into Deep and Shallow Rule Learning
An Empirical Investigation into Deep and Shallow Rule Learning
Florian Beck
Johannes Furnkranz
NAI
18
7
0
18 Jun 2021
Quantized Neural Networks via {-1, +1} Encoding Decomposition and
  Acceleration
Quantized Neural Networks via {-1, +1} Encoding Decomposition and Acceleration
Qigong Sun
Xiufang Li
Fanhua Shang
Hongying Liu
Kan Yang
L. Jiao
Zhouchen Lin
MQ
31
0
0
18 Jun 2021
Integer-Only Neural Network Quantization Scheme Based on
  Shift-Batch-Normalization
Integer-Only Neural Network Quantization Scheme Based on Shift-Batch-Normalization
Qingyu Guo
Yuan Wang
Xiaoxin Cui
MQ
11
2
0
28 May 2021
Is In-Domain Data Really Needed? A Pilot Study on Cross-Domain
  Calibration for Network Quantization
Is In-Domain Data Really Needed? A Pilot Study on Cross-Domain Calibration for Network Quantization
Haichao Yu
Linjie Yang
Humphrey Shi
OOD
MQ
34
5
0
16 May 2021
Model Pruning Based on Quantified Similarity of Feature Maps
Model Pruning Based on Quantified Similarity of Feature Maps
Zidu Wang
Xue-jun Liu
Long Huang
Yuxiang Chen
Yufei Zhang
Zhikang Lin
Rui Wang
21
16
0
13 May 2021
Binarized Weight Error Networks With a Transition Regularization Term
Binarized Weight Error Networks With a Transition Regularization Term
Savas Ozkan
G. Akar
MQ
26
1
0
09 May 2021
Modulating Regularization Frequency for Efficient Compression-Aware
  Model Training
Modulating Regularization Frequency for Efficient Compression-Aware Model Training
Dongsoo Lee
S. Kwon
Byeongwook Kim
Jeongin Yun
Baeseong Park
Yongkweon Jeon
21
0
0
05 May 2021
Q-Rater: Non-Convex Optimization for Post-Training Uniform Quantization
Q-Rater: Non-Convex Optimization for Post-Training Uniform Quantization
Byeongwook Kim
Dongsoo Lee
Yeonju Ro
Yongkweon Jeon
S. Kwon
Baeseong Park
Daehwan Oh
MQ
18
1
0
05 May 2021
Training Quantized Neural Networks to Global Optimality via Semidefinite
  Programming
Training Quantized Neural Networks to Global Optimality via Semidefinite Programming
Burak Bartan
Mert Pilanci
24
10
0
04 May 2021
Quantization of Deep Neural Networks for Accurate Edge Computing
Quantization of Deep Neural Networks for Accurate Edge Computing
Wentao Chen
Hailong Qiu
Zhuang Jian
Chutong Zhang
Yu Hu
Qing Lu
Tianchen Wang
Yiyu Shi
Meiping Huang
Xiaowe Xu
52
21
0
25 Apr 2021
Learning on Hardware: A Tutorial on Neural Network Accelerators and
  Co-Processors
Learning on Hardware: A Tutorial on Neural Network Accelerators and Co-Processors
Lukas Baischer
M. Wess
N. Taherinejad
20
12
0
19 Apr 2021
Previous
123456...91011
Next