ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.17871
69
0
v1v2 (latest)

BLAST: Balanced Sampling Time Series Corpus for Universal Forecasting Models

23 May 2025
Zezhi Shao
Yujie Li
Fei Wang
Chengqing Yu
Yisong Fu
Tangwen Qian
Bin Xu
Boyu Diao
Yongjun Xu
Xueqi Cheng
    AI4TS
ArXiv (abs)PDFHTML
Main:9 Pages
6 Figures
Bibliography:2 Pages
12 Tables
Appendix:1 Pages
Abstract

The advent of universal time series forecasting models has revolutionized zero-shot forecasting across diverse domains, yet the critical role of data diversity in training these models remains underexplored. Existing large-scale time series datasets often suffer from inherent biases and imbalanced distributions, leading to suboptimal model performance and generalization. To address this gap, we introduce BLAST, a novel pre-training corpus designed to enhance data diversity through a balanced sampling strategy. First, BLAST incorporates 321 billion observations from publicly available datasets and employs a comprehensive suite of statistical metrics to characterize time series patterns. Then, to facilitate pattern-oriented sampling, the data is implicitly clustered using grid-based partitioning. Furthermore, by integrating grid sampling and grid mixup techniques, BLAST ensures a balanced and representative coverage of diverse patterns. Experimental results demonstrate that models pre-trained on BLAST achieve state-of-the-art performance with a fraction of the computational resources and training tokens required by existing methods. Our findings highlight the pivotal role of data diversity in improving both training efficiency and model performance for the universal forecasting task.

View on arXiv
@article{shao2025_2505.17871,
  title={ BLAST: Balanced Sampling Time Series Corpus for Universal Forecasting Models },
  author={ Zezhi Shao and Yujie Li and Fei Wang and Chengqing Yu and Yisong Fu and Tangwen Qian and Bin Xu and Boyu Diao and Yongjun Xu and Xueqi Cheng },
  journal={arXiv preprint arXiv:2505.17871},
  year={ 2025 }
}
Comments on this paper