14
0

AutoHFormer: Efficient Hierarchical Autoregressive Transformer for Time Series Prediction

Main:12 Pages
7 Figures
Bibliography:2 Pages
1 Tables
Abstract

Time series forecasting requires architectures that simultaneously achieve three competing objectives: (1) strict temporal causality for reliable predictions, (2) sub-quadratic complexity for practical scalability, and (3) multi-scale pattern recognition for accurate long-horizon forecasting. We introduce AutoHFormer, a hierarchical autoregressive transformer that addresses these challenges through three key innovations: 1) Hierarchical Temporal Modeling: Our architecture decomposes predictions into segment-level blocks processed in parallel, followed by intra-segment sequential refinement. This dual-scale approach maintains temporal coherence while enabling efficient computation. 2) Dynamic Windowed Attention: The attention mechanism employs learnable causal windows with exponential decay, reducing complexity while preserving precise temporal relationships. This design avoids both the anti-causal violations of standard transformers and the sequential bottlenecks of RNN hybrids. 3) Adaptive Temporal Encoding: a novel position encoding system is adopted to capture time patterns at multiple scales. It combines fixed oscillating patterns for short-term variations with learnable decay rates for long-term trends. Comprehensive experiments demonstrate that AutoHFormer 10.76X faster training and 6.06X memory reduction compared to PatchTST on PEMS08, while maintaining consistent accuracy across 96-720 step horizons in most of cases. These breakthroughs establish new benchmarks for efficient and precise time series modeling. Implementations of our method and all baselines in hierarchical autoregressive mechanism are available atthis https URL.

View on arXiv
@article{zhang2025_2506.16001,
  title={ AutoHFormer: Efficient Hierarchical Autoregressive Transformer for Time Series Prediction },
  author={ Qianru Zhang and Honggang Wen and Ming Li and Dong Huang and Siu-Ming Yiu and Christian S. Jensen and Pietro Liò },
  journal={arXiv preprint arXiv:2506.16001},
  year={ 2025 }
}
Comments on this paper