Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2412.11242
Cited By
v1
v2 (latest)
TrimLLM: Progressive Layer Dropping for Domain-Specific LLMs
15 December 2024
Lanxiang Hu
Tajana Rosing
Hao Zhang
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"TrimLLM: Progressive Layer Dropping for Domain-Specific LLMs"
10 / 10 papers shown
Title
The Internal State of an LLM Knows When It's Lying
A. Azaria
Tom Michael Mitchell
HILM
269
340
0
26 Apr 2023
GPTQ: Accurate Post-Training Quantization for Generative Pre-trained Transformers
Elias Frantar
Saleh Ashkboos
Torsten Hoefler
Dan Alistarh
MQ
129
989
0
31 Oct 2022
Mass-Editing Memory in a Transformer
Kevin Meng
Arnab Sen Sharma
A. Andonian
Yonatan Belinkov
David Bau
KELM
VLM
133
584
0
13 Oct 2022
LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale
Tim Dettmers
M. Lewis
Younes Belkada
Luke Zettlemoyer
MQ
98
653
0
15 Aug 2022
Locating and Editing Factual Associations in GPT
Kevin Meng
David Bau
A. Andonian
Yonatan Belinkov
KELM
248
1,357
0
10 Feb 2022
LexGLUE: A Benchmark Dataset for Legal Language Understanding in English
Ilias Chalkidis
Abhik Jana
D. Hartung
M. Bommarito
Ion Androutsopoulos
Daniel Martin Katz
Nikolaos Aletras
AILaw
ELM
223
264
0
03 Oct 2021
Transformer Feed-Forward Layers Are Key-Value Memories
Mor Geva
R. Schuster
Jonathan Berant
Omer Levy
KELM
158
828
0
29 Dec 2020
Intrinsic Dimensionality Explains the Effectiveness of Language Model Fine-Tuning
Armen Aghajanyan
Luke Zettlemoyer
Sonal Gupta
101
568
1
22 Dec 2020
Accelerating Training of Transformer-Based Language Models with Progressive Layer Dropping
Minjia Zhang
Yuxiong He
AI4CE
42
104
0
26 Oct 2020
PubMedQA: A Dataset for Biomedical Research Question Answering
Qiao Jin
Bhuwan Dhingra
Zhengping Liu
William W. Cohen
Xinghua Lu
382
896
0
13 Sep 2019
1