ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1802.02615
  4. Cited By
Effective Quantization Approaches for Recurrent Neural Networks

Effective Quantization Approaches for Recurrent Neural Networks

7 February 2018
Md. Zahangir Alom
A. Moody
N. Maruyama
B. Van Essen
T. Taha
    MQ
ArXiv (abs)PDFHTML

Papers citing "Effective Quantization Approaches for Recurrent Neural Networks"

18 / 18 papers shown
Title
HadamRNN: Binary and Sparse Ternary Orthogonal RNNs
HadamRNN: Binary and Sparse Ternary Orthogonal RNNs
Armand Foucault
Franck Mamalet
François Malgouyres
MQ
329
0
0
28 Jan 2025
BILLNET: A Binarized Conv3D-LSTM Network with Logic-gated residual architecture for hardware-efficient video inference
BILLNET: A Binarized Conv3D-LSTM Network with Logic-gated residual architecture for hardware-efficient video inference
Van Thien Nguyen
William Guicquero
Gilles Sicard
3DVMQ
149
2
0
24 Jan 2025
Histogram-Equalized Quantization for logic-gated Residual Neural Networks
Histogram-Equalized Quantization for logic-gated Residual Neural Networks
Van Thien Nguyen
William Guicquero
Gilles Sicard
MQ
130
2
0
10 Jan 2025
Towards a tailored mixed-precision sub-8-bit quantization scheme for
  Gated Recurrent Units using Genetic Algorithms
Towards a tailored mixed-precision sub-8-bit quantization scheme for Gated Recurrent Units using Genetic Algorithms
Riccardo Miccini
Alessandro Cerioli
Clément Laroche
Tobias Piechowiak
J. Sparsø
Luca Pezzarossa
MQ
63
2
0
19 Feb 2024
Quantized Approximately Orthogonal Recurrent Neural Networks
Quantized Approximately Orthogonal Recurrent Neural Networks
Armand Foucault
Franck Mamalet
Franccois Malgouyres
MQ
104
1
0
05 Feb 2024
NASH: A Simple Unified Framework of Structured Pruning for Accelerating
  Encoder-Decoder Language Models
NASH: A Simple Unified Framework of Structured Pruning for Accelerating Encoder-Decoder Language Models
Jongwoo Ko
Seungjoon Park
Yujin Kim
Sumyeong Ahn
Du-Seong Chang
Euijai Ahn
SeYoung Yun
115
6
0
16 Oct 2023
FullPack: Full Vector Utilization for Sub-Byte Quantized Inference on
  General Purpose CPUs
FullPack: Full Vector Utilization for Sub-Byte Quantized Inference on General Purpose CPUs
Hossein Katebi
Navidreza Asadi
M. Goudarzi
MQ
65
0
0
13 Nov 2022
Vau da muntanialas: Energy-efficient multi-die scalable acceleration of
  RNN inference
Vau da muntanialas: Energy-efficient multi-die scalable acceleration of RNN inference
G. Paulin
Francesco Conti
Lukas Cavigelli
Luca Benini
64
8
0
14 Feb 2022
Spectral Pruning for Recurrent Neural Networks
Spectral Pruning for Recurrent Neural Networks
Takashi Furuya
Kazuma Suetake
K. Taniguchi
Hiroyuki Kusumoto
Ryuji Saiin
Tomohiro Daimon
64
4
0
23 May 2021
A Variational Information Bottleneck Based Method to Compress Sequential
  Networks for Human Action Recognition
A Variational Information Bottleneck Based Method to Compress Sequential Networks for Human Action Recognition
Ayush Srivastava
Oshin Dutta
A. Prathosh
Sumeet Agarwal
Jigyasa Gupta
61
8
0
03 Oct 2020
Compression of Deep Learning Models for Text: A Survey
Compression of Deep Learning Models for Text: A Survey
Manish Gupta
Puneet Agrawal
VLMMedImAI4CE
79
119
0
12 Aug 2020
Training with reduced precision of a support vector machine model for
  text classification
Training with reduced precision of a support vector machine model for text classification
Dominik Zurek
Marcin Pietroñ
MQ
13
0
0
17 Jul 2020
Tensor train decompositions on recurrent networks
Tensor train decompositions on recurrent networks
A. Murua
R. Ramakrishnan
Xinlin Li
Rui Heng Yang
V. Nia
51
0
0
09 Jun 2020
Non-Volatile Memory Array Based Quantization- and Noise-Resilient LSTM
  Neural Networks
Non-Volatile Memory Array Based Quantization- and Noise-Resilient LSTM Neural Networks
Wen Ma
P. Chiu
Won Ho Choi
Minghai Qin
D. Bedau
Martin Lueker-Boden
MQ
23
4
0
25 Feb 2020
AdaptivFloat: A Floating-point based Data Type for Resilient Deep
  Learning Inference
AdaptivFloat: A Floating-point based Data Type for Resilient Deep Learning Inference
Thierry Tambe
En-Yu Yang
Zishen Wan
Yuntian Deng
Vijay Janapa Reddi
Alexander M. Rush
David Brooks
Gu-Yeon Wei
MQ
69
21
0
29 Sep 2019
Recurrent Neural Networks: An Embedded Computing Perspective
Recurrent Neural Networks: An Embedded Computing Perspective
Nesma M. Rezk
M. Purnaprajna
Tomas Nordstrom
Z. Ul-Abdin
129
82
0
23 Jul 2019
Efficient 8-Bit Quantization of Transformer Neural Machine Language
  Translation Model
Efficient 8-Bit Quantization of Transformer Neural Machine Language Translation Model
Aishwarya Bhandare
Vamsi Sripathi
Deepthi Karkada
Vivek V. Menon
Sun Choi
Kushal Datta
V. Saletore
MQ
95
132
0
03 Jun 2019
The History Began from AlexNet: A Comprehensive Survey on Deep Learning
  Approaches
The History Began from AlexNet: A Comprehensive Survey on Deep Learning Approaches
Md. Zahangir Alom
T. Taha
C. Yakopcic
Stefan Westberg
P. Sidike
Mst Shamima Nasrin
B. Van Essen
A. Awwal
V. Asari
VLM
133
883
0
03 Mar 2018
1