ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.15021
5
0

SFT-GO: Supervised Fine-Tuning with Group Optimization for Large Language Models

17 June 2025
Gyuhak Kim
Sumiran Singh Thakur
Su Min Park
Wei Wei
Yujia Bao
ArXiv (abs)PDFHTML
Main:9 Pages
7 Figures
Bibliography:4 Pages
10 Tables
Appendix:11 Pages
Abstract

Supervised fine-tuning (SFT) has become an essential step in tailoring large language models (LLMs) to align with human expectations and specific downstream tasks. However, existing SFT methods typically treat each training instance as a uniform sequence, giving equal importance to all tokens regardless of their relevance. This overlooks the fact that only a subset of tokens often contains critical, task-specific information. To address this limitation, we introduce Supervised Fine-Tuning with Group Optimization (SFT-GO), a novel approach that treats groups of tokens differently based on theirthis http URL-GO groups tokens in each sample based on their importance values and optimizes the LLM using a weighted combination of the worst-group loss and the standard cross-entropy loss. This mechanism adaptively emphasizes the most challenging token groups and guides the model to better handle different group distributions, thereby improving overall learning dynamics. We provide a theoretical analysis of SFT-GO's convergence rate, demonstrating its efficiency. Empirically, we apply SFT-GO with three different token grouping strategies and show that models trained with SFT-GO consistently outperform baseline approaches across popular LLM benchmarks. These improvements hold across various datasets and base models, demonstrating the robustness and the effectiveness of our method.

View on arXiv
@article{kim2025_2506.15021,
  title={ SFT-GO: Supervised Fine-Tuning with Group Optimization for Large Language Models },
  author={ Gyuhak Kim and Sumiran Singh Thakur and Su Min Park and Wei Wei and Yujia Bao },
  journal={arXiv preprint arXiv:2506.15021},
  year={ 2025 }
}
Comments on this paper