42

Steer2Adapt: Dynamically Composing Steering Vectors Elicits Efficient Adaptation of LLMs

Pengrui Han
Xueqiang Xu
Keyang Xuan
Peiyang Song
Siru Ouyang
Runchu Tian
Yuqing Jiang
Cheng Qian
Pengcheng Jiang
Jiashuo Sun
Junxia Cui
Ming Zhong
Ge Liu
Jiawei Han
Jiaxuan You
Main:8 Pages
8 Figures
Bibliography:5 Pages
10 Tables
Appendix:12 Pages
Abstract

Activation steering has emerged as a promising approach for efficiently adapting large language models (LLMs) to downstream behaviors. However, most existing steering methods rely on a single static direction per task or concept, making them inflexible under task variation and inadequate for complex tasks that require multiple coordinated capabilities. To address this limitation, we propose STEER2ADAPT, a lightweight framework that adapts LLMs by composing steering vectors rather than learning new ones from scratch. In many domains (e.g., reasoning or safety), tasks share a small set of underlying concept dimensions. STEER2ADAPT captures these dimensions as a reusable, low-dimensional semantic prior subspace, and adapts to new tasks by dynamically discovering a linear combination of basis vectors from only a handful of examples. Experiments across 9 tasks and 3 models in both reasoning and safety domains demonstrate the effectiveness of STEER2ADAPT, achieving an average improvement of 8.2%. Extensive analyses further show that STEER2ADAPT is a data-efficient, stable, and transparent inference-time adaptation method for LLMs.

View on arXiv
Comments on this paper