ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.18706
23
0

Steering LLM Reasoning Through Bias-Only Adaptation

24 May 2025
Viacheslav Sinii
Alexey Gorbatovski
Artem Cherepanov
Boris Shaposhnikov
Nikita Balagansky
Daniil Gavrilov
    LLMSVLRM
ArXiv (abs)PDFHTML
Main:3 Pages
Bibliography:2 Pages
4 Tables
Appendix:2 Pages
Abstract

Recent work on reasoning-oriented language models, exemplified by o1-like systems, suggests that reinforcement-learning (RL) finetuning does not create new capabilities but instead strengthens reasoning patterns already latent in the pretrained network. We test this claim by training steering vectors: layer-wise biases that additively amplify selected hidden features while leaving all original weights unchanged. Experiments on four base models across the GSM8K and MATH benchmarks show that steering vectors recover, and in several cases exceed, the accuracy of fully-tuned counterparts. This result supports the view that the required reasoning skills pre-exist in the base model. Further, logit-lens analysis reveals that the trained vectors consistently boost token groups linked to structured languages and logical connectors, providing an interpretable account that aligns with the demands of quantitative reasoning tasks.

View on arXiv
@article{sinii2025_2505.18706,
  title={ Steering LLM Reasoning Through Bias-Only Adaptation },
  author={ Viacheslav Sinii and Alexey Gorbatovski and Artem Cherepanov and Boris Shaposhnikov and Nikita Balagansky and Daniil Gavrilov },
  journal={arXiv preprint arXiv:2505.18706},
  year={ 2025 }
}
Comments on this paper