Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2306.02379
Cited By
Modular Transformers: Compressing Transformers into Modularized Layers for Flexible Efficient Inference
4 June 2023
Wangchunshu Zhou
Ronan Le Bras
Yejin Choi
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Modular Transformers: Compressing Transformers into Modularized Layers for Flexible Efficient Inference"
4 / 4 papers shown
Title
A Survey on Green Deep Learning
Jingjing Xu
Wangchunshu Zhou
Zhiyi Fu
Hao Zhou
Lei Li
VLM
81
83
0
08 Nov 2021
BERT-of-Theseus: Compressing BERT by Progressive Module Replacing
Canwen Xu
Wangchunshu Zhou
Tao Ge
Furu Wei
Ming Zhou
229
198
0
07 Feb 2020
Q-BERT: Hessian Based Ultra Low Precision Quantization of BERT
Sheng Shen
Zhen Dong
Jiayu Ye
Linjian Ma
Z. Yao
A. Gholami
Michael W. Mahoney
Kurt Keutzer
MQ
236
576
0
12 Sep 2019
Teaching Machines to Read and Comprehend
Karl Moritz Hermann
Tomás Kociský
Edward Grefenstette
L. Espeholt
W. Kay
Mustafa Suleyman
Phil Blunsom
211
3,513
0
10 Jun 2015
1