24
0

HouseTS: A Large-Scale, Multimodal Spatiotemporal U.S. Housing Dataset

Main:9 Pages
8 Figures
Bibliography:4 Pages
6 Tables
Appendix:4 Pages
Abstract

Accurate house-price forecasting is essential for investors, planners, and researchers. However, reproducible benchmarks with sufficient spatiotemporal depth and contextual richness for long horizon prediction remain scarce. To address this, we introduce HouseTS a large scale, multimodal dataset covering monthly house prices from March 2012 to December 2023 across 6,000 ZIP codes in 30 major U.S. metropolitan areas. The dataset includes over 890K records, enriched with points of Interest (POI), socioeconomic indicators, and detailed real estate metrics. To establish standardized performance baselines, we evaluate 14 models, spanning classical statistical approaches, deep neural networks (DNNs), and pretrained time-series foundation models. We further demonstrate the value of HouseTS in a multimodal case study, where a vision language model extracts structured textual descriptions of geographic change from time stamped satellite imagery. This enables interpretable, grounded insights into urban evolution. HouseTS is hosted on Kaggle, while all preprocessing pipelines, benchmark code, and documentation are openly maintained on GitHub to ensure full reproducibility and easy adoption.

View on arXiv
@article{wang2025_2506.00765,
  title={ HouseTS: A Large-Scale, Multimodal Spatiotemporal U.S. Housing Dataset },
  author={ Shengkun Wang and Yanshen Sun and Fanglan Chen and Linhan Wang and Naren Ramakrishnan and Chang-Tien Lu and Yinlin Chen },
  journal={arXiv preprint arXiv:2506.00765},
  year={ 2025 }
}
Comments on this paper