ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2211.00635
  4. Cited By
Two-stage LLM Fine-tuning with Less Specialization and More
  Generalization

Two-stage LLM Fine-tuning with Less Specialization and More Generalization

1 November 2022
Yihan Wang
Si Si
Daliang Li
Michal Lukasik
Felix X. Yu
Cho-Jui Hsieh
Inderjit S Dhillon
Sanjiv Kumar
ArXivPDFHTML

Papers citing "Two-stage LLM Fine-tuning with Less Specialization and More Generalization"

22 / 22 papers shown
Title
HSplitLoRA: A Heterogeneous Split Parameter-Efficient Fine-Tuning Framework for Large Language Models
HSplitLoRA: A Heterogeneous Split Parameter-Efficient Fine-Tuning Framework for Large Language Models
Zheng Lin
Yuxin Zhang
Zhe Chen
Zihan Fang
Xianhao Chen
Praneeth Vepakomma
Wei Ni
Jun Luo
Yue Gao
MoE
41
2
0
05 May 2025
ML For Hardware Design Interpretability: Challenges and Opportunities
ML For Hardware Design Interpretability: Challenges and Opportunities
Raymond Baartmans
Andrew Ensinger
Victor Agostinelli
Lizhong Chen
29
0
0
11 Apr 2025
Effective Text Adaptation for LLM-based ASR through Soft Prompt
  Fine-Tuning
Effective Text Adaptation for LLM-based ASR through Soft Prompt Fine-Tuning
Yingyi Ma
Zhe Liu
Ozlem Kalinli
70
0
0
09 Dec 2024
Preserving Pre-trained Representation Space: On Effectiveness of
  Prefix-tuning for Large Multi-modal Models
Preserving Pre-trained Representation Space: On Effectiveness of Prefix-tuning for Large Multi-modal Models
Donghoon Kim
Gusang Lee
Kyuhong Shim
B. Shim
56
1
0
29 Oct 2024
Semantic Image Inversion and Editing using Rectified Stochastic
  Differential Equations
Semantic Image Inversion and Editing using Rectified Stochastic Differential Equations
Litu Rout
Yujia Chen
Nataniel Ruiz
C. Caramanis
Sanjay Shakkottai
Wen-Sheng Chu
DiffM
59
23
0
14 Oct 2024
DELIA: Diversity-Enhanced Learning for Instruction Adaptation in Large
  Language Models
DELIA: Diversity-Enhanced Learning for Instruction Adaptation in Large Language Models
Yuanhao Zeng
Fei Ren
Xinpeng Zhou
Yihang Wang
Yingxia Shao
ALM
31
0
0
19 Aug 2024
PromptIntern: Saving Inference Costs by Internalizing Recurrent Prompt
  during Large Language Model Fine-tuning
PromptIntern: Saving Inference Costs by Internalizing Recurrent Prompt during Large Language Model Fine-tuning
Jiaru Zou
Mengyu Zhou
Tao Li
Shi Han
Dongmei Zhang
46
6
0
02 Jul 2024
Large Scale Transfer Learning for Tabular Data via Language Modeling
Large Scale Transfer Learning for Tabular Data via Language Modeling
Josh Gardner
Juan C. Perdomo
Ludwig Schmidt
LMTD
36
13
0
17 Jun 2024
Mistral-C2F: Coarse to Fine Actor for Analytical and Reasoning
  Enhancement in RLHF and Effective-Merged LLMs
Mistral-C2F: Coarse to Fine Actor for Analytical and Reasoning Enhancement in RLHF and Effective-Merged LLMs
Chen Zheng
Ke Sun
Xun Zhou
MoE
49
0
0
12 Jun 2024
TAIA: Large Language Models are Out-of-Distribution Data Learners
TAIA: Large Language Models are Out-of-Distribution Data Learners
Shuyang Jiang
Yusheng Liao
Ya-Qin Zhang
Yu Wang
Yanfeng Wang
29
3
0
30 May 2024
Intuition-aware Mixture-of-Rank-1-Experts for Parameter Efficient
  Finetuning
Intuition-aware Mixture-of-Rank-1-Experts for Parameter Efficient Finetuning
Yijiang Liu
Rongyu Zhang
Huanrui Yang
Kurt Keutzer
Yuan Du
Li Du
Shanghang Zhang
MoE
41
6
0
13 Apr 2024
Unveiling the Generalization Power of Fine-Tuned Large Language Models
Unveiling the Generalization Power of Fine-Tuned Large Language Models
Haoran Yang
Yumeng Zhang
Jiaqi Xu
Hongyuan Lu
Pheng Ann Heng
Wai Lam
45
28
0
14 Mar 2024
Balancing Enhancement, Harmlessness, and General Capabilities: Enhancing
  Conversational LLMs with Direct RLHF
Balancing Enhancement, Harmlessness, and General Capabilities: Enhancing Conversational LLMs with Direct RLHF
Chen Zheng
Ke Sun
Hang Wu
Chenguang Xi
Xun Zhou
52
12
0
04 Mar 2024
Mitigating Catastrophic Forgetting in Large Language Models with
  Self-Synthesized Rehearsal
Mitigating Catastrophic Forgetting in Large Language Models with Self-Synthesized Rehearsal
Jianheng Huang
Leyang Cui
Ante Wang
Chengyi Yang
Xinting Liao
Linfeng Song
Junfeng Yao
Jinsong Su
KELM
CLL
32
32
0
02 Mar 2024
Aalap: AI Assistant for Legal & Paralegal Functions in India
Aalap: AI Assistant for Legal & Paralegal Functions in India
Aman Tiwari
Prathamesh Kalamkar
Atreyo Banerjee
S. Karn
V. Hemachandran
Smita Gupta
AILaw
ELM
VLM
26
1
0
30 Jan 2024
The ART of LLM Refinement: Ask, Refine, and Trust
The ART of LLM Refinement: Ask, Refine, and Trust
Kumar Shridhar
Koustuv Sinha
Andrew Cohen
Tianlu Wang
Ping Yu
Ramakanth Pasunuru
Mrinmaya Sachan
Jason Weston
Asli Celikyilmaz
LLMAG
ReLM
LRM
27
24
0
14 Nov 2023
Investigating the Catastrophic Forgetting in Multimodal Large Language
  Models
Investigating the Catastrophic Forgetting in Multimodal Large Language Models
Yuexiang Zhai
Shengbang Tong
Xiao Li
Mu Cai
Qing Qu
Yong Jae Lee
Y. Ma
VLM
MLLM
CLL
77
77
0
19 Sep 2023
Domain Specialization as the Key to Make Large Language Models
  Disruptive: A Comprehensive Survey
Domain Specialization as the Key to Make Large Language Models Disruptive: A Comprehensive Survey
Chen Ling
Xujiang Zhao
Jiaying Lu
Chengyuan Deng
Can Zheng
...
Chris White
Quanquan Gu
Jian Pei
Carl Yang
Liang Zhao
ALM
25
126
0
30 May 2023
Large Language Models are Versatile Decomposers: Decompose Evidence and
  Questions for Table-based Reasoning
Large Language Models are Versatile Decomposers: Decompose Evidence and Questions for Table-based Reasoning
Yunhu Ye
Binyuan Hui
Min Yang
Binhua Li
Fei Huang
Yongbin Li
LMTD
ReLM
LRM
44
143
0
31 Jan 2023
P-Tuning v2: Prompt Tuning Can Be Comparable to Fine-tuning Universally
  Across Scales and Tasks
P-Tuning v2: Prompt Tuning Can Be Comparable to Fine-tuning Universally Across Scales and Tasks
Xiao Liu
Kaixuan Ji
Yicheng Fu
Weng Lam Tam
Zhengxiao Du
Zhilin Yang
Jie Tang
VLM
238
806
0
14 Oct 2021
The Power of Scale for Parameter-Efficient Prompt Tuning
The Power of Scale for Parameter-Efficient Prompt Tuning
Brian Lester
Rami Al-Rfou
Noah Constant
VPVLM
280
3,848
0
18 Apr 2021
Making Pre-trained Language Models Better Few-shot Learners
Making Pre-trained Language Models Better Few-shot Learners
Tianyu Gao
Adam Fisch
Danqi Chen
241
1,919
0
31 Dec 2020
1