ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1705.08922
  4. Cited By
Exploring the Regularity of Sparse Structure in Convolutional Neural
  Networks

Exploring the Regularity of Sparse Structure in Convolutional Neural Networks

24 May 2017
Huizi Mao
Song Han
Jeff Pool
Wenshuo Li
Xingyu Liu
Yu Wang
W. Dally
ArXivPDFHTML

Papers citing "Exploring the Regularity of Sparse Structure in Convolutional Neural Networks"

45 / 45 papers shown
Title
Training Acceleration of Low-Rank Decomposed Networks using Sequential
  Freezing and Rank Quantization
Training Acceleration of Low-Rank Decomposed Networks using Sequential Freezing and Rank Quantization
Habib Hajimolahoseini
Walid Ahmed
Yang Liu
OffRL
MQ
21
6
0
07 Sep 2023
Efficient Quantized Sparse Matrix Operations on Tensor Cores
Efficient Quantized Sparse Matrix Operations on Tensor Cores
Shigang Li
Kazuki Osawa
Torsten Hoefler
82
31
0
14 Sep 2022
Towards Sparsification of Graph Neural Networks
Towards Sparsification of Graph Neural Networks
Hongwu Peng
Deniz Gurevin
Shaoyi Huang
Tong Geng
Weiwen Jiang
O. Khan
Caiwen Ding
GNN
30
24
0
11 Sep 2022
Entropy Induced Pruning Framework for Convolutional Neural Networks
Entropy Induced Pruning Framework for Convolutional Neural Networks
Yihe Lu
Ziyu Guan
Yaming Yang
Maoguo Gong
Wei Zhao
Kaiyuan Feng
38
2
0
13 Aug 2022
Mixed-Precision Neural Networks: A Survey
Mixed-Precision Neural Networks: A Survey
M. Rakka
M. Fouda
Pramod P. Khargonekar
Fadi J. Kurdahi
MQ
30
11
0
11 Aug 2022
Compiler-Aware Neural Architecture Search for On-Mobile Real-time
  Super-Resolution
Compiler-Aware Neural Architecture Search for On-Mobile Real-time Super-Resolution
Yushu Wu
Yifan Gong
Pu Zhao
Yanyu Li
Zheng Zhan
Wei Niu
Hao Tang
Minghai Qin
Bin Ren
Yanzhi Wang
SupR
MQ
40
23
0
25 Jul 2022
ZeroQuant: Efficient and Affordable Post-Training Quantization for
  Large-Scale Transformers
ZeroQuant: Efficient and Affordable Post-Training Quantization for Large-Scale Transformers
Z. Yao
Reza Yazdani Aminabadi
Minjia Zhang
Xiaoxia Wu
Conglong Li
Yuxiong He
VLM
MQ
73
448
0
04 Jun 2022
LilNetX: Lightweight Networks with EXtreme Model Compression and
  Structured Sparsification
LilNetX: Lightweight Networks with EXtreme Model Compression and Structured Sparsification
Sharath Girish
Kamal Gupta
Saurabh Singh
Abhinav Shrivastava
40
11
0
06 Apr 2022
Quantization in Layer's Input is Matter
Quantization in Layer's Input is Matter
Daning Cheng
Wenguang Chen
MQ
11
0
0
10 Feb 2022
EcoFlow: Efficient Convolutional Dataflows for Low-Power Neural Network
  Accelerators
EcoFlow: Efficient Convolutional Dataflows for Low-Power Neural Network Accelerators
Lois Orosa
Skanda Koppula
Yaman Umuroglu
Konstantinos Kanellopoulos
Juan Gómez Luna
Michaela Blott
K. Vissers
O. Mutlu
46
4
0
04 Feb 2022
Load-balanced Gather-scatter Patterns for Sparse Deep Neural Networks
Load-balanced Gather-scatter Patterns for Sparse Deep Neural Networks
Fei Sun
Minghai Qin
Tianyun Zhang
Xiaolong Ma
Haoran Li
Junwen Luo
Zihao Zhao
Yen-kuang Chen
Yuan Xie
27
1
0
20 Dec 2021
Mixed Precision Low-bit Quantization of Neural Network Language Models
  for Speech Recognition
Mixed Precision Low-bit Quantization of Neural Network Language Models for Speech Recognition
Junhao Xu
Jianwei Yu
Shoukang Hu
Xunying Liu
Helen Meng
MQ
32
13
0
29 Nov 2021
Improving the Accuracy of Early Exits in Multi-Exit Architectures via
  Curriculum Learning
Improving the Accuracy of Early Exits in Multi-Exit Architectures via Curriculum Learning
Arian Bakhtiarnia
Qi Zhang
Alexandros Iosifidis
38
12
0
21 Apr 2021
RingCNN: Exploiting Algebraically-Sparse Ring Tensors for
  Energy-Efficient CNN-Based Computational Imaging
RingCNN: Exploiting Algebraically-Sparse Ring Tensors for Energy-Efficient CNN-Based Computational Imaging
Chao-Tsung Huang
45
10
0
19 Apr 2021
Lottery Jackpots Exist in Pre-trained Models
Lottery Jackpots Exist in Pre-trained Models
Yuxin Zhang
Mingbao Lin
Yan Wang
Rongrong Ji
Rongrong Ji
35
15
0
18 Apr 2021
Dancing along Battery: Enabling Transformer with Run-time
  Reconfigurability on Mobile Devices
Dancing along Battery: Enabling Transformer with Run-time Reconfigurability on Mobile Devices
Yuhong Song
Weiwen Jiang
Bingbing Li
Panjie Qi
Qingfeng Zhuge
E. Sha
Sakyasingha Dasgupta
Yiyu Shi
Caiwen Ding
18
18
0
12 Feb 2021
BRDS: An FPGA-based LSTM Accelerator with Row-Balanced Dual-Ratio
  Sparsification
BRDS: An FPGA-based LSTM Accelerator with Row-Balanced Dual-Ratio Sparsification
Seyed Abolfazl Ghasemzadeh
E. Tavakoli
M. Kamal
A. Afzali-Kusha
Massoud Pedram
24
13
0
07 Jan 2021
I-BERT: Integer-only BERT Quantization
I-BERT: Integer-only BERT Quantization
Sehoon Kim
A. Gholami
Z. Yao
Michael W. Mahoney
Kurt Keutzer
MQ
107
345
0
05 Jan 2021
Parallel Blockwise Knowledge Distillation for Deep Neural Network
  Compression
Parallel Blockwise Knowledge Distillation for Deep Neural Network Compression
Cody Blakeney
Xiaomin Li
Yan Yan
Ziliang Zong
53
40
0
05 Dec 2020
Depthwise Multiception Convolution for Reducing Network Parameters
  without Sacrificing Accuracy
Depthwise Multiception Convolution for Reducing Network Parameters without Sacrificing Accuracy
Guoqing Bao
M. Graeber
Xiuying Wang
13
5
0
07 Nov 2020
Self-grouping Convolutional Neural Networks
Self-grouping Convolutional Neural Networks
Qingbei Guo
Xiaojun Wu
J. Kittler
Zhiquan Feng
25
22
0
29 Sep 2020
Standing on the Shoulders of Giants: Hardware and Neural Architecture
  Co-Search with Hot Start
Standing on the Shoulders of Giants: Hardware and Neural Architecture Co-Search with Hot Start
Weiwen Jiang
Lei Yang
Sakyasingha Dasgupta
Jiaxi Hu
Yiyu Shi
27
59
0
17 Jul 2020
CSB-RNN: A Faster-than-Realtime RNN Acceleration Framework with
  Compressed Structured Blocks
CSB-RNN: A Faster-than-Realtime RNN Acceleration Framework with Compressed Structured Blocks
Runbin Shi
Peiyan Dong
Tong Geng
Yuhao Ding
Xiaolong Ma
Hayden Kwok-Hay So
Martin C. Herbordt
Ang Li
Yanzhi Wang
MQ
21
13
0
11 May 2020
A Survey of Convolutional Neural Networks: Analysis, Applications, and
  Prospects
A Survey of Convolutional Neural Networks: Analysis, Applications, and Prospects
Zewen Li
Wenjie Yang
Shouheng Peng
Fan Liu
HAI
3DV
64
2,608
0
01 Apr 2020
CoCoPIE: Making Mobile AI Sweet As PIE --Compression-Compilation
  Co-Design Goes a Long Way
CoCoPIE: Making Mobile AI Sweet As PIE --Compression-Compilation Co-Design Goes a Long Way
Shaoshan Liu
Bin Ren
Xipeng Shen
Yanzhi Wang
17
18
0
14 Mar 2020
How Does BN Increase Collapsed Neural Network Filters?
How Does BN Increase Collapsed Neural Network Filters?
Sheng Zhou
Xinjiang Wang
Ping Luo
Xue Jiang
Wenjie Li
Wei Zhang
21
1
0
30 Jan 2020
An Image Enhancing Pattern-based Sparsity for Real-time Inference on
  Mobile Devices
An Image Enhancing Pattern-based Sparsity for Real-time Inference on Mobile Devices
Xiaolong Ma
Wei Niu
Tianyun Zhang
Sijia Liu
Sheng Lin
...
Xiang Chen
Jian Tang
Kaisheng Ma
Bin Ren
Yanzhi Wang
46
27
0
20 Jan 2020
Sparse Weight Activation Training
Sparse Weight Activation Training
Md Aamir Raihan
Tor M. Aamodt
34
73
0
07 Jan 2020
ZeroQ: A Novel Zero Shot Quantization Framework
ZeroQ: A Novel Zero Shot Quantization Framework
Yaohui Cai
Z. Yao
Zhen Dong
A. Gholami
Michael W. Mahoney
Kurt Keutzer
MQ
43
389
0
01 Jan 2020
PatDNN: Achieving Real-Time DNN Execution on Mobile Devices with
  Pattern-based Weight Pruning
PatDNN: Achieving Real-Time DNN Execution on Mobile Devices with Pattern-based Weight Pruning
Wei Niu
Xiaolong Ma
Sheng Lin
Shihao Wang
Xuehai Qian
Xinyu Lin
Yanzhi Wang
Bin Ren
MQ
35
227
0
01 Jan 2020
A Pre-defined Sparse Kernel Based Convolution for Deep CNNs
A Pre-defined Sparse Kernel Based Convolution for Deep CNNs
Souvik Kundu
Saurav Prakash
H. Akrami
Peter A. Beerel
K. Chugg
36
12
0
02 Oct 2019
PCONV: The Missing but Desirable Sparsity in DNN Weight Pruning for
  Real-time Execution on Mobile Devices
PCONV: The Missing but Desirable Sparsity in DNN Weight Pruning for Real-time Execution on Mobile Devices
Xiaolong Ma
Fu-Ming Guo
Wei Niu
Xue Lin
Jian Tang
Kaisheng Ma
Bin Ren
Yanzhi Wang
CVBM
27
173
0
06 Sep 2019
Parameterized Structured Pruning for Deep Neural Networks
Parameterized Structured Pruning for Deep Neural Networks
Günther Schindler
Wolfgang Roth
Franz Pernkopf
Holger Froening
26
6
0
12 Jun 2019
Structured Compression by Weight Encryption for Unstructured Pruning and
  Quantization
Structured Compression by Weight Encryption for Unstructured Pruning and Quantization
S. Kwon
Dongsoo Lee
Byeongwook Kim
Parichay Kapoor
Baeseong Park
Gu-Yeon Wei
MQ
35
48
0
24 May 2019
Towards Efficient Model Compression via Learned Global Ranking
Towards Efficient Model Compression via Learned Global Ranking
Ting-Wu Chin
Ruizhou Ding
Cha Zhang
Diana Marculescu
16
170
0
28 Apr 2019
Progressive DNN Compression: A Key to Achieve Ultra-High Weight Pruning
  and Quantization Rates using ADMM
Progressive DNN Compression: A Key to Achieve Ultra-High Weight Pruning and Quantization Rates using ADMM
Shaokai Ye
Xiaoyu Feng
Tianyun Zhang
Xiaolong Ma
Sheng Lin
...
Jian Tang
M. Fardad
Xinyu Lin
Yongpan Liu
Yanzhi Wang
MQ
38
38
0
23 Mar 2019
ADMM-NN: An Algorithm-Hardware Co-Design Framework of DNNs Using
  Alternating Direction Method of Multipliers
ADMM-NN: An Algorithm-Hardware Co-Design Framework of DNNs Using Alternating Direction Method of Multipliers
Ao Ren
Tianyun Zhang
Shaokai Ye
Jiayu Li
Wenyao Xu
Xuehai Qian
Xinyu Lin
Yanzhi Wang
MQ
40
161
0
31 Dec 2018
Exploiting Kernel Sparsity and Entropy for Interpretable CNN Compression
Exploiting Kernel Sparsity and Entropy for Interpretable CNN Compression
Yuchao Li
Shaohui Lin
Baochang Zhang
Jianzhuang Liu
David Doermann
Yongjian Wu
Feiyue Huang
Rongrong Ji
43
130
0
11 Dec 2018
Efficient Structured Pruning and Architecture Searching for Group
  Convolution
Efficient Structured Pruning and Architecture Searching for Group Convolution
Ruizhe Zhao
Wayne Luk
45
16
0
23 Nov 2018
Fast On-the-fly Retraining-free Sparsification of Convolutional Neural
  Networks
Fast On-the-fly Retraining-free Sparsification of Convolutional Neural Networks
Amir H. Ashouri
T. Abdelrahman
Alwyn Dos Remedios
MQ
16
12
0
10 Nov 2018
Progressive Weight Pruning of Deep Neural Networks using ADMM
Progressive Weight Pruning of Deep Neural Networks using ADMM
Shaokai Ye
Tianyun Zhang
Kaiqi Zhang
Jiayu Li
Kaidi Xu
...
M. Fardad
Sijia Liu
Xiang Chen
Xinyu Lin
Yanzhi Wang
AI4CE
37
38
0
17 Oct 2018
Accelerator-Aware Pruning for Convolutional Neural Networks
Accelerator-Aware Pruning for Convolutional Neural Networks
Hyeong-Ju Kang
13
88
0
26 Apr 2018
IGCV$2$: Interleaved Structured Sparse Convolutional Neural Networks
IGCV222: Interleaved Structured Sparse Convolutional Neural Networks
Guotian Xie
Jingdong Wang
Ting Zhang
Jianhuang Lai
Richang Hong
Guo-Jun Qi
19
105
0
17 Apr 2018
Efficient Hardware Realization of Convolutional Neural Networks using
  Intra-Kernel Regular Pruning
Efficient Hardware Realization of Convolutional Neural Networks using Intra-Kernel Regular Pruning
Maurice Yang
Mahmoud Faraj
Assem Hussein
V. Gaudet
CVBM
22
12
0
15 Mar 2018
Escoin: Efficient Sparse Convolutional Neural Network Inference on GPUs
Escoin: Efficient Sparse Convolutional Neural Network Inference on GPUs
Xuhao Chen
18
25
0
28 Feb 2018
1