ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1605.04711
  4. Cited By
Ternary Weight Networks

Ternary Weight Networks

16 May 2016
Fengfu Li
Bin Liu
Xiaoxing Wang
Bo-Wen Zhang
Junchi Yan
    MQ
ArXivPDFHTML

Papers citing "Ternary Weight Networks"

50 / 207 papers shown
Title
Evaluation Metrics for DNNs Compression
Evaluation Metrics for DNNs Compression
Abanoub Ghobrial
S. Budgett
Dieter Balemans
Hamid Asgari
Philippe Reiter
Kerstin Eder
27
1
0
18 May 2023
AutoQNN: An End-to-End Framework for Automatically Quantizing Neural
  Networks
AutoQNN: An End-to-End Framework for Automatically Quantizing Neural Networks
Cheng Gong
Ye Lu
Surong Dai
Deng Qian
Chenkun Du
Tao Li
MQ
29
0
0
07 Apr 2023
Ternary Quantization: A Survey
Ternary Quantization: A Survey
Danyang Liu
Xue Liu
MQ
18
3
0
02 Mar 2023
A Comprehensive Review and a Taxonomy of Edge Machine Learning:
  Requirements, Paradigms, and Techniques
A Comprehensive Review and a Taxonomy of Edge Machine Learning: Requirements, Paradigms, and Techniques
Wenbin Li
Hakim Hacid
Ebtesam Almazrouei
Merouane Debbah
34
13
0
16 Feb 2023
Learning Discretized Neural Networks under Ricci Flow
Learning Discretized Neural Networks under Ricci Flow
Jun Chen
Han Chen
Mengmeng Wang
Guang Dai
Ivor W. Tsang
Y. Liu
25
2
0
07 Feb 2023
Self-Compressing Neural Networks
Self-Compressing Neural Networks
Szabolcs Cséfalvay
J. Imber
16
2
0
30 Jan 2023
Understanding INT4 Quantization for Transformer Models: Latency Speedup,
  Composability, and Failure Cases
Understanding INT4 Quantization for Transformer Models: Latency Speedup, Composability, and Failure Cases
Xiaoxia Wu
Cheng-rong Li
Reza Yazdani Aminabadi
Z. Yao
Yuxiong He
MQ
11
19
0
27 Jan 2023
RedBit: An End-to-End Flexible Framework for Evaluating the Accuracy of
  Quantized CNNs
RedBit: An End-to-End Flexible Framework for Evaluating the Accuracy of Quantized CNNs
A. M. Ribeiro-dos-Santos
João Dinis Ferreira
O. Mutlu
G. Falcão
MQ
18
1
0
15 Jan 2023
Hyperspherical Quantization: Toward Smaller and More Accurate Models
Hyperspherical Quantization: Toward Smaller and More Accurate Models
Dan Liu
X. Chen
Chen-li Ma
Xue Liu
MQ
27
3
0
24 Dec 2022
Pruning On-the-Fly: A Recoverable Pruning Method without Fine-tuning
Pruning On-the-Fly: A Recoverable Pruning Method without Fine-tuning
Danyang Liu
Xue Liu
20
0
0
24 Dec 2022
Hyperspherical Loss-Aware Ternary Quantization
Hyperspherical Loss-Aware Ternary Quantization
Dan Liu
Xue Liu
MQ
21
0
0
24 Dec 2022
CSQ: Growing Mixed-Precision Quantization Scheme with Bi-level
  Continuous Sparsification
CSQ: Growing Mixed-Precision Quantization Scheme with Bi-level Continuous Sparsification
Lirui Xiao
Huanrui Yang
Zhen Dong
Kurt Keutzer
Li Du
Shanghang Zhang
MQ
27
10
0
06 Dec 2022
Are Straight-Through gradients and Soft-Thresholding all you need for
  Sparse Training?
Are Straight-Through gradients and Soft-Thresholding all you need for Sparse Training?
A. Vanderschueren
Christophe De Vleeschouwer
MQ
25
9
0
02 Dec 2022
Boosted Dynamic Neural Networks
Boosted Dynamic Neural Networks
Haichao Yu
Haoxiang Li
G. Hua
Gao Huang
Humphrey Shi
35
7
0
30 Nov 2022
Signed Binary Weight Networks
Sachit Kuhar
Alexey Tumanov
Judy Hoffman
MQ
18
1
0
25 Nov 2022
AskewSGD : An Annealed interval-constrained Optimisation method to train
  Quantized Neural Networks
AskewSGD : An Annealed interval-constrained Optimisation method to train Quantized Neural Networks
Louis Leconte
S. Schechtman
Eric Moulines
29
4
0
07 Nov 2022
Collaborative Multi-Teacher Knowledge Distillation for Learning Low
  Bit-width Deep Neural Networks
Collaborative Multi-Teacher Knowledge Distillation for Learning Low Bit-width Deep Neural Networks
Cuong Pham
Tuan Hoang
Thanh-Toan Do
FedML
MQ
21
14
0
27 Oct 2022
Q-ViT: Accurate and Fully Quantized Low-bit Vision Transformer
Q-ViT: Accurate and Fully Quantized Low-bit Vision Transformer
Yanjing Li
Sheng Xu
Baochang Zhang
Xianbin Cao
Penglei Gao
Guodong Guo
MQ
ViT
26
89
0
13 Oct 2022
SeKron: A Decomposition Method Supporting Many Factorization Structures
SeKron: A Decomposition Method Supporting Many Factorization Structures
Marawan Gamal Abdel Hameed
A. Mosleh
Marzieh S. Tahaei
V. Nia
21
1
0
12 Oct 2022
Seeking Interpretability and Explainability in Binary Activated Neural
  Networks
Seeking Interpretability and Explainability in Binary Activated Neural Networks
Benjamin Leblanc
Pascal Germain
FAtt
37
1
0
07 Sep 2022
DenseShift: Towards Accurate and Efficient Low-Bit Power-of-Two
  Quantization
DenseShift: Towards Accurate and Efficient Low-Bit Power-of-Two Quantization
Xinlin Li
Bangya Liu
Ruizhi Yang
Vanessa Courville
Chao Xing
V. Nia
MQ
34
2
0
20 Aug 2022
AdaBin: Improving Binary Neural Networks with Adaptive Binary Sets
AdaBin: Improving Binary Neural Networks with Adaptive Binary Sets
Zhaopeng Tu
Xinghao Chen
Pengju Ren
Yunhe Wang
MQ
36
54
0
17 Aug 2022
Mixed-Precision Neural Networks: A Survey
Mixed-Precision Neural Networks: A Survey
M. Rakka
M. Fouda
Pramod P. Khargonekar
Fadi J. Kurdahi
MQ
21
11
0
11 Aug 2022
Safety and Performance, Why not Both? Bi-Objective Optimized Model
  Compression toward AI Software Deployment
Safety and Performance, Why not Both? Bi-Objective Optimized Model Compression toward AI Software Deployment
Jie Zhu
Leye Wang
Xiao Han
23
9
0
11 Aug 2022
Quantized Adaptive Subgradient Algorithms and Their Applications
Quantized Adaptive Subgradient Algorithms and Their Applications
Ke Xu
Jianqiao Wangni
Yifan Zhang
Deheng Ye
Jiaxiang Wu
P. Zhao
31
0
0
11 Aug 2022
PalQuant: Accelerating High-precision Networks on Low-precision
  Accelerators
PalQuant: Accelerating High-precision Networks on Low-precision Accelerators
Qinghao Hu
Gang Li
Qiman Wu
Jian Cheng
MQ
23
2
0
03 Aug 2022
Low-bit Shift Network for End-to-End Spoken Language Understanding
Low-bit Shift Network for End-to-End Spoken Language Understanding
Anderson R. Avila
Khalil Bibi
Ruizhi Yang
Xinlin Li
Chao Xing
Xiao Chen
MQ
31
4
0
15 Jul 2022
Expressive power of binary and ternary neural networks
Expressive power of binary and ternary neural networks
A. Beknazaryan
MQ
11
0
0
27 Jun 2022
Towards Green ASR: Lossless 4-bit Quantization of a Hybrid TDNN System
  on the 300-hr Switchboard Corpus
Towards Green ASR: Lossless 4-bit Quantization of a Hybrid TDNN System on the 300-hr Switchboard Corpus
Junhao Xu
Shoukang Hu
Xunying Liu
Helen M. Meng
MQ
14
5
0
23 Jun 2022
Compilation and Optimizations for Efficient Machine Learning on Embedded
  Systems
Compilation and Optimizations for Efficient Machine Learning on Embedded Systems
Xiaofan Zhang
Yao Chen
Cong Hao
Sitao Huang
Yuhong Li
Deming Chen
39
1
0
06 Jun 2022
Combinatorial optimization for low bit-width neural networks
Combinatorial optimization for low bit-width neural networks
Hanxu Zhou
Aida Ashrafi
Matthew B. Blaschko
MQ
24
0
0
04 Jun 2022
ZeroQuant: Efficient and Affordable Post-Training Quantization for
  Large-Scale Transformers
ZeroQuant: Efficient and Affordable Post-Training Quantization for Large-Scale Transformers
Z. Yao
Reza Yazdani Aminabadi
Minjia Zhang
Xiaoxia Wu
Conglong Li
Yuxiong He
VLM
MQ
45
441
0
04 Jun 2022
Extreme Compression for Pre-trained Transformers Made Simple and
  Efficient
Extreme Compression for Pre-trained Transformers Made Simple and Efficient
Xiaoxia Wu
Z. Yao
Minjia Zhang
Conglong Li
Yuxiong He
MQ
19
31
0
04 Jun 2022
A Low Memory Footprint Quantized Neural Network for Depth Completion of
  Very Sparse Time-of-Flight Depth Maps
A Low Memory Footprint Quantized Neural Network for Depth Completion of Very Sparse Time-of-Flight Depth Maps
Xiao-Yan Jiang
V. Cambareri
Gianluca Agresti
C. Ugwu
Adriano Simonetto
Fabien Cardinaux
Pietro Zanuttigh
3DV
MQ
30
9
0
25 May 2022
On Algebraic Constructions of Neural Networks with Small Weights
On Algebraic Constructions of Neural Networks with Small Weights
Kordag Mehmet Kilic
Jin Sima
J. Bruck
8
2
0
17 May 2022
A Comprehensive Survey on Model Quantization for Deep Neural Networks in
  Image Classification
A Comprehensive Survey on Model Quantization for Deep Neural Networks in Image Classification
Babak Rokh
A. Azarpeyvand
Alireza Khanteymoori
MQ
30
82
0
14 May 2022
Hardware-Robust In-RRAM-Computing for Object Detection
Hardware-Robust In-RRAM-Computing for Object Detection
Yu-Hsiang Chiang
Cheng-En Ni
Yun Sung
T. Hou
Tian-Sheuan Chang
S. Jou
6
5
0
09 May 2022
Enable Deep Learning on Mobile Devices: Methods, Systems, and
  Applications
Enable Deep Learning on Mobile Devices: Methods, Systems, and Applications
Han Cai
Ji Lin
Yujun Lin
Zhijian Liu
Haotian Tang
Hanrui Wang
Ligeng Zhu
Song Han
21
107
0
25 Apr 2022
HCFL: A High Compression Approach for Communication-Efficient Federated
  Learning in Very Large Scale IoT Networks
HCFL: A High Compression Approach for Communication-Efficient Federated Learning in Very Large Scale IoT Networks
Minh-Duong Nguyen
Sangmin Lee
Viet Quoc Pham
D. Hoang
Diep N. Nguyen
W. Hwang
20
28
0
14 Apr 2022
LilNetX: Lightweight Networks with EXtreme Model Compression and
  Structured Sparsification
LilNetX: Lightweight Networks with EXtreme Model Compression and Structured Sparsification
Sharath Girish
Kamal Gupta
Saurabh Singh
Abhinav Shrivastava
28
11
0
06 Apr 2022
FxP-QNet: A Post-Training Quantizer for the Design of Mixed
  Low-Precision DNNs with Dynamic Fixed-Point Representation
FxP-QNet: A Post-Training Quantizer for the Design of Mixed Low-Precision DNNs with Dynamic Fixed-Point Representation
Ahmad Shawahna
S. M. Sait
A. El-Maleh
Irfan Ahmad
MQ
18
6
0
22 Mar 2022
DQ-BART: Efficient Sequence-to-Sequence Model via Joint Distillation and
  Quantization
DQ-BART: Efficient Sequence-to-Sequence Model via Joint Distillation and Quantization
Zheng Li
Zijian Wang
Ming Tan
Ramesh Nallapati
Parminder Bhatia
Andrew O. Arnold
Bing Xiang
Dan Roth
MQ
18
40
0
21 Mar 2022
EAutoDet: Efficient Architecture Search for Object Detection
EAutoDet: Efficient Architecture Search for Object Detection
Xiaoxing Wang
Jiale Lin
Junchi Yan
Juanping Zhao
Xiaokang Yang
32
23
0
21 Mar 2022
FlexBlock: A Flexible DNN Training Accelerator with Multi-Mode Block
  Floating Point Support
FlexBlock: A Flexible DNN Training Accelerator with Multi-Mode Block Floating Point Support
Seock-Hwan Noh
Jahyun Koo
Seunghyun Lee
Jongse Park
Jaeha Kung
AI4CE
29
17
0
13 Mar 2022
YONO: Modeling Multiple Heterogeneous Neural Networks on
  Microcontrollers
YONO: Modeling Multiple Heterogeneous Neural Networks on Microcontrollers
Young D. Kwon
Jagmohan Chauhan
Cecilia Mascolo
16
13
0
08 Mar 2022
Standard Deviation-Based Quantization for Deep Neural Networks
Standard Deviation-Based Quantization for Deep Neural Networks
Amir Ardakani
A. Ardakani
B. Meyer
J. Clark
W. Gross
MQ
46
1
0
24 Feb 2022
Bitwidth Heterogeneous Federated Learning with Progressive Weight
  Dequantization
Bitwidth Heterogeneous Federated Learning with Progressive Weight Dequantization
Jaehong Yoon
Geondo Park
Wonyong Jeong
Sung Ju Hwang
FedML
21
19
0
23 Feb 2022
Distilled Neural Networks for Efficient Learning to Rank
Distilled Neural Networks for Efficient Learning to Rank
F. M. Nardini
Cosimo Rulli
Salvatore Trani
Rossano Venturini
FedML
29
16
0
22 Feb 2022
Bit-wise Training of Neural Network Weights
Bit-wise Training of Neural Network Weights
Cristian Ivan
MQ
18
0
0
19 Feb 2022
Energy awareness in low precision neural networks
Energy awareness in low precision neural networks
Nurit Spingarn-Eliezer
Ron Banner
Elad Hoffer
Hilla Ben-Yaacov
T. Michaeli
38
0
0
06 Feb 2022
Previous
12345
Next