ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2412.18860
37
1

Bootstrap Your Own Context Length

25 December 2024
Liang Wang
Nan Yang
Xingxing Zhang
Xiaolong Huang
Furu Wei
    SyDa
ArXivPDFHTML
Abstract

We introduce a bootstrapping approach to train long-context language models by exploiting their short-context capabilities only. Our method utilizes a simple agent workflow to synthesize diverse long-context instruction tuning data, thereby eliminating the necessity for manual data collection and annotation. The proposed data synthesis workflow requires only a short-context language model, a text retriever, and a document collection, all of which are readily accessible within the open-source ecosystem. Subsequently, language models are fine-tuned using the synthesized data to extend their context lengths. In this manner, we effectively transfer the short-context capabilities of language models to long-context scenarios through a bootstrapping process. We conduct experiments with the open-source Llama-3 family of models and demonstrate that our method can successfully extend the context length to up to 1M tokens, achieving superior performance across various benchmarks.

View on arXiv
@article{wang2025_2412.18860,
  title={ Bootstrap Your Own Context Length },
  author={ Liang Wang and Nan Yang and Xingxing Zhang and Xiaolong Huang and Furu Wei },
  journal={arXiv preprint arXiv:2412.18860},
  year={ 2025 }
}
Comments on this paper