Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2308.04269
Cited By
Lossy and Lossless (L
2
^2
2
) Post-training Model Size Compression
8 August 2023
Yumeng Shi
Shihao Bai
Xiuying Wei
Ruihao Gong
Jianlei Yang
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Lossy and Lossless (L$^2$) Post-training Model Size Compression"
21 / 21 papers shown
Title
CSQ: Growing Mixed-Precision Quantization Scheme with Bi-level Continuous Sparsification
Lirui Xiao
Huanrui Yang
Zhen Dong
Kurt Keutzer
Li Du
Shanghang Zhang
MQ
47
10
0
06 Dec 2022
Quantized Sparse Weight Decomposition for Neural Network Compression
Andrey Kuzmin
M. V. Baalen
Markus Nagel
Arash Behboodi
MQ
35
3
0
22 Jul 2022
QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quantization
Xiuying Wei
Ruihao Gong
Yuhang Li
Xianglong Liu
F. Yu
MQ
VLM
74
175
0
11 Mar 2022
OMPQ: Orthogonal Mixed Precision Quantization
Yuexiao Ma
Taisong Jin
Xiawu Zheng
Yan Wang
Huixia Li
Yongjian Wu
Guannan Jiang
Wei Zhang
Rongrong Ji
MQ
96
37
0
16 Sep 2021
Post-training deep neural network pruning via layer-wise calibration
Ivan Lazarevich
Alexander Kozlov
Nikita Malinin
3DPC
49
26
0
30 Apr 2021
Topology-Aware Network Pruning using Multi-stage Graph Embedding and Reinforcement Learning
Sixing Yu
Arya Mazaheri
Ali Jannesari
75
39
0
05 Feb 2021
Up or Down? Adaptive Rounding for Post-Training Quantization
Markus Nagel
Rana Ali Amjad
M. V. Baalen
Christos Louizos
Tijmen Blankevoort
MQ
80
577
0
22 Apr 2020
Training with Quantization Noise for Extreme Model Compression
Angela Fan
Pierre Stock
Benjamin Graham
Edouard Grave
Remi Gribonval
Hervé Jégou
Armand Joulin
MQ
90
245
0
15 Apr 2020
Designing Network Design Spaces
Ilija Radosavovic
Raj Prateek Kosaraju
Ross B. Girshick
Kaiming He
Piotr Dollár
GNN
100
1,682
0
30 Mar 2020
HAWQ-V2: Hessian Aware trace-Weighted Quantization of Neural Networks
Zhen Dong
Z. Yao
Yaohui Cai
Daiyaan Arfeen
A. Gholami
Michael W. Mahoney
Kurt Keutzer
MQ
84
279
0
10 Nov 2019
DeepCABAC: A Universal Compression Algorithm for Deep Neural Networks
Simon Wiedemann
H. Kirchhoffer
Stefan Matlage
Paul Haase
Arturo Marbán
...
Ahmed Osman
D. Marpe
H. Schwarz
Thomas Wiegand
Wojciech Samek
82
95
0
27 Jul 2019
Data-Free Quantization Through Weight Equalization and Bias Correction
Markus Nagel
M. V. Baalen
Tijmen Blankevoort
Max Welling
MQ
70
512
0
11 Jun 2019
Feature Map Transform Coding for Energy-Efficient CNN Inference
Brian Chmiel
Chaim Baskin
Ron Banner
Evgenii Zheltonozhskii
Yevgeny Yermolin
Alex Karbachevsky
A. Bronstein
A. Mendelson
71
25
0
26 May 2019
DeepCABAC: Context-adaptive binary arithmetic coding for deep neural network compression
Simon Wiedemann
H. Kirchhoffer
Stefan Matlage
Paul Haase
Arturo Marbán
...
Ahmed Osman
D. Marpe
H. Schwarz
Thomas Wiegand
Wojciech Samek
MQ
46
21
0
15 May 2019
MnasNet: Platform-Aware Neural Architecture Search for Mobile
Mingxing Tan
Bo Chen
Ruoming Pang
Vijay Vasudevan
Mark Sandler
Andrew G. Howard
Quoc V. Le
MQ
120
3,009
0
31 Jul 2018
LQ-Nets: Learned Quantization for Highly Accurate and Compact Deep Neural Networks
Dongqing Zhang
Jiaolong Yang
Dongqiangzi Ye
G. Hua
MQ
62
703
0
26 Jul 2018
AMC: AutoML for Model Compression and Acceleration on Mobile Devices
Yihui He
Ji Lin
Zhijian Liu
Hanrui Wang
Li Li
Song Han
93
1,347
0
10 Feb 2018
MobileNetV2: Inverted Residuals and Linear Bottlenecks
Mark Sandler
Andrew G. Howard
Menglong Zhu
A. Zhmoginov
Liang-Chieh Chen
181
19,271
0
13 Jan 2018
Soft Weight-Sharing for Neural Network Compression
Karen Ullrich
Edward Meeds
Max Welling
167
417
0
13 Feb 2017
Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding
Song Han
Huizi Mao
W. Dally
3DGS
255
8,833
0
01 Oct 2015
Learning both Weights and Connections for Efficient Neural Networks
Song Han
Jeff Pool
J. Tran
W. Dally
CVBM
310
6,672
0
08 Jun 2015
1