Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2405.06038
Cited By
From Algorithm to Hardware: A Survey on Efficient and Safe Deployment of Deep Neural Networks
9 May 2024
Xue Geng
Zhe Wang
Chunyun Chen
Qing Xu
Kaixin Xu
Chao Jin
Manas Gupta
Xulei Yang
Zhenghua Chen
M. Aly
Jie Lin
Min-man Wu
Xiaoli Li
Re-assign community
ArXiv
PDF
HTML
Papers citing
"From Algorithm to Hardware: A Survey on Efficient and Safe Deployment of Deep Neural Networks"
50 / 103 papers shown
Title
Work-Efficient Parallel Non-Maximum Suppression Kernels
David Oro
Carles Fernández
Xavier Martorell
Javier Hernando
110
7
0
01 Feb 2025
REED: Chiplet-Based Accelerator for Fully Homomorphic Encryption
Aikata Aikata
A. Mert
Sunmin Kwon
M. Deryabin
S. Roy
127
2
0
05 Aug 2023
Towards Accurate Post-Training Quantization for Vision Transformer
Yifu Ding
Haotong Qin
Qing-Yu Yan
Z. Chai
Junjie Liu
Xiaolin K. Wei
Xianglong Liu
MQ
94
69
0
25 Mar 2023
Scaling Instruction-Finetuned Language Models
Hyung Won Chung
Le Hou
Shayne Longpre
Barret Zoph
Yi Tay
...
Jacob Devlin
Adam Roberts
Denny Zhou
Quoc V. Le
Jason W. Wei
ReLM
LRM
167
3,110
0
20 Oct 2022
Structural Pruning via Latency-Saliency Knapsack
Maying Shen
Hongxu Yin
Pavlo Molchanov
Lei Mao
Jianna Liu
J. Álvarez
69
50
0
13 Oct 2022
Is Complexity Required for Neural Network Pruning? A Case Study on Global Magnitude Pruning
Manas Gupta
Efe Camci
Vishandi Rudy Keneta
Abhishek Vaidyanathan
Ritwik Kanodia
Chuan-Sheng Foo
Wu Min
Lin Jie
46
14
0
29 Sep 2022
Optimal Brain Compression: A Framework for Accurate Post-Training Quantization and Pruning
Elias Frantar
Sidak Pal Singh
Dan Alistarh
MQ
76
236
0
24 Aug 2022
ZeroQuant: Efficient and Affordable Post-Training Quantization for Large-Scale Transformers
Z. Yao
Reza Yazdani Aminabadi
Minjia Zhang
Xiaoxia Wu
Conglong Li
Yuxiong He
VLM
MQ
102
476
0
04 Jun 2022
Decoupled Knowledge Distillation
Borui Zhao
Quan Cui
Renjie Song
Yiyu Qiu
Jiajun Liang
56
539
0
16 Mar 2022
A Survey on Model Compression and Acceleration for Pretrained Language Models
Canwen Xu
Julian McAuley
73
60
0
15 Feb 2022
Exploring Inter-Channel Correlation for Diversity-preserved KnowledgeDistillation
Li Liu
Qingle Huang
Sihao Lin
Hongwei Xie
Bing Wang
Xiaojun Chang
Xiao-Xue Liang
72
103
0
08 Feb 2022
BTS: An Accelerator for Bootstrappable Fully Homomorphic Encryption
Sangpyo Kim
Jongmin Kim
M. Kim
Wonkyung Jung
Minsoo Rhu
John Kim
Jung Ho Ahn
44
145
0
31 Dec 2021
MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning
Ji Lin
Wei-Ming Chen
Han Cai
Chuang Gan
Song Han
77
157
0
28 Oct 2021
Revisiting Knowledge Distillation: An Inheritance and Exploration Framework
Zhen Huang
Xu Shen
Jun Xing
Tongliang Liu
Xinmei Tian
Houqiang Li
Bing Deng
Jianqiang Huang
Xiansheng Hua
55
27
0
01 Jul 2021
Post-Training Quantization for Vision Transformer
Zhenhua Liu
Yunhe Wang
Kai Han
Siwei Ma
Wen Gao
ViT
MQ
89
337
0
27 Jun 2021
PSRR-MaxpoolNMS: Pyramid Shifted MaxpoolNMS with Relationship Recovery
Tianyi Zhang
Jie Lin
Peng Hu
Bin Zhao
M. Aly
32
4
0
27 May 2021
Complementary Relation Contrastive Distillation
Jinguo Zhu
Shixiang Tang
Dapeng Chen
Shijie Yu
Yakun Liu
A. Yang
M. Rong
Xiaohua Wang
57
80
0
29 Mar 2021
Dynamic Slimmable Network
Changlin Li
Guangrun Wang
Bing Wang
Xiaodan Liang
Zhihui Li
Xiaojun Chang
58
144
0
24 Mar 2021
BossNAS: Exploring Hybrid CNN-transformers with Block-wisely Self-supervised Neural Architecture Search
Changlin Li
Tao Tang
Guangrun Wang
Jiefeng Peng
Bing Wang
Xiaodan Liang
Xiaojun Chang
ViT
97
107
0
23 Mar 2021
Refine Myself by Teaching Myself: Feature Refinement via Self-Knowledge Distillation
Mingi Ji
Seungjae Shin
Seunghyun Hwang
Gibeom Park
Il-Chul Moon
32
123
0
15 Mar 2021
Learning Student-Friendly Teacher Networks for Knowledge Distillation
D. Park
Moonsu Cha
C. Jeong
Daesin Kim
Bohyung Han
158
101
0
12 Feb 2021
Learning N:M Fine-grained Structured Sparse Neural Networks From Scratch
Aojun Zhou
Yukun Ma
Junnan Zhu
Jianbo Liu
Zhijie Zhang
Kun Yuan
Wenxiu Sun
Hongsheng Li
188
247
0
08 Feb 2021
Show, Attend and Distill:Knowledge Distillation via Attention-based Feature Matching
Mingi Ji
Byeongho Heo
Sungrae Park
93
147
0
05 Feb 2021
Neural Pruning via Growing Regularization
Huan Wang
Can Qin
Yulun Zhang
Y. Fu
74
145
0
16 Dec 2020
SCOP: Scientific Control for Reliable Neural Network Pruning
Yehui Tang
Yunhe Wang
Yixing Xu
Dacheng Tao
Chunjing Xu
Chao Xu
Chang Xu
AAML
76
166
0
21 Oct 2020
Accelerate CNNs from Three Dimensions: A Comprehensive Pruning Framework
Wenxiao Wang
Minghao Chen
Shuai Zhao
Long Chen
Jinming Hu
Haifeng Liu
Deng Cai
Xiaofei He
Wei Liu
54
60
0
10 Oct 2020
Transform Quantization for CNN (Convolutional Neural Network) Compression
Sean I. Young
Wang Zhe
David S. Taubman
B. Girod
MQ
49
72
0
02 Sep 2020
Semantics-aware Adaptive Knowledge Distillation for Sensor-to-Vision Action Recognition
Yang Liu
Keze Wang
Guanbin Li
Liang Lin
79
89
0
01 Sep 2020
FracBits: Mixed Precision Quantization via Fractional Bit-Widths
Linjie Yang
Qing Jin
MQ
49
74
0
04 Jul 2020
Hardware Acceleration of Sparse and Irregular Tensor Computations of ML Models: A Survey and Insights
Shail Dave
Riyadh Baghdadi
Tony Nowatzki
Sasikanth Avancha
Aviral Shrivastava
Baoxin Li
88
82
0
02 Jul 2020
Progressive Skeletonization: Trimming more fat from a network at initialization
Pau de Jorge
Amartya Sanyal
Harkirat Singh Behl
Philip Torr
Grégory Rogez
P. Dokania
57
96
0
16 Jun 2020
Dynamic Sparse Training: Find Efficient Sparse Network From Scratch With Trainable Masked Layers
Junjie Liu
Zhe Xu
Runbin Shi
R. Cheung
Hayden Kwok-Hay So
46
121
0
14 May 2020
HRank: Filter Pruning using High-Rank Feature Map
Mingbao Lin
Rongrong Ji
Yan Wang
Yichen Zhang
Baochang Zhang
Yonghong Tian
Ling Shao
64
724
0
24 Feb 2020
A
3
^3
3
: Accelerating Attention Mechanisms in Neural Networks with Approximation
Tae Jun Ham
Sungjun Jung
Seonghak Kim
Young H. Oh
Yeonhong Park
...
Jung-Hun Park
Sanghee Lee
Kyoung Park
Jae W. Lee
D. Jeong
50
218
0
22 Feb 2020
Lookahead: A Far-Sighted Alternative of Magnitude-based Pruning
Sejun Park
Jaeho Lee
Sangwoo Mo
Jinwoo Shin
44
93
0
12 Feb 2020
Soft Threshold Weight Reparameterization for Learnable Sparsity
Aditya Kusupati
Vivek Ramanujan
Raghav Somani
Mitchell Wortsman
Prateek Jain
Sham Kakade
Ali Farhadi
140
246
0
08 Feb 2020
Feature-map-level Online Adversarial Knowledge Distillation
Inseop Chung
Seonguk Park
Jangho Kim
Nojun Kwak
GAN
75
128
0
05 Feb 2020
Adaptive Loss-aware Quantization for Multi-bit Networks
Zhongnan Qu
Zimu Zhou
Yun Cheng
Lothar Thiele
MQ
118
55
0
18 Dec 2019
Dreaming to Distill: Data-free Knowledge Transfer via DeepInversion
Hongxu Yin
Pavlo Molchanov
Zhizhong Li
J. Álvarez
Arun Mallya
Derek Hoiem
N. Jha
Jan Kautz
62
563
0
18 Dec 2019
Winning the Lottery with Continuous Sparsification
Pedro H. P. Savarese
Hugo Silva
Michael Maire
61
135
0
10 Dec 2019
Survey of Attacks and Defenses on Edge-Deployed Neural Networks
Mihailo Isakov
V. Gadepally
K. Gettings
Michel A. Kinsy
AAML
39
31
0
27 Nov 2019
Rigging the Lottery: Making All Tickets Winners
Utku Evci
Trevor Gale
Jacob Menick
Pablo Samuel Castro
Erich Elsen
178
600
0
25 Nov 2019
Contrastive Representation Distillation
Yonglong Tian
Dilip Krishnan
Phillip Isola
141
1,045
0
23 Oct 2019
Additive Powers-of-Two Quantization: An Efficient Non-uniform Discretization for Neural Networks
Yuhang Li
Xin Dong
Wei Wang
MQ
60
258
0
28 Sep 2019
Once-for-All: Train One Network and Specialize it for Efficient Deployment
Han Cai
Chuang Gan
Tianzhe Wang
Zhekai Zhang
Song Han
OOD
100
1,277
0
26 Aug 2019
Learning Filter Basis for Convolutional Neural Network Compression
Yawei Li
Shuhang Gu
Luc Van Gool
Radu Timofte
SupR
54
99
0
23 Aug 2019
Differentiable Soft Quantization: Bridging Full-Precision and Low-Bit Neural Networks
Ruihao Gong
Xianglong Liu
Shenghu Jiang
Tian-Hao Li
Peng Hu
Jiazhen Lin
F. Yu
Junjie Yan
MQ
58
457
0
14 Aug 2019
Sampled Softmax with Random Fourier Features
A. S. Rawat
Jiecao Chen
Felix X. Yu
A. Suresh
Sanjiv Kumar
62
55
0
24 Jul 2019
And the Bit Goes Down: Revisiting the Quantization of Neural Networks
Pierre Stock
Armand Joulin
Rémi Gribonval
Benjamin Graham
Hervé Jégou
MQ
70
150
0
12 Jul 2019
Sparse Networks from Scratch: Faster Training without Losing Performance
Tim Dettmers
Luke Zettlemoyer
130
339
0
10 Jul 2019
1
2
3
Next