35
0

Hidden Representation Clustering with Multi-Task Representation Learning towards Robust Online Budget Allocation

Main:8 Pages
8 Figures
Bibliography:2 Pages
3 Tables
Appendix:3 Pages
Abstract

Marketing optimization, commonly formulated as an online budget allocation problem, has emerged as a pivotal factor in driving user growth. Most existing research addresses this problem by following the principle of 'first predict then optimize' for each individual, which presents challenges related to large-scale counterfactual prediction and solving complexity trade-offs. Note that the practical data quality is uncontrollable, and the solving scale tends to be tens of millions. Therefore, the existing approaches make the robust budget allocation non-trivial, especially in industrial scenarios with considerable data noise. To this end, this paper proposes a novel approach that solves the problem from the cluster perspective. Specifically, we propose a multi-task representation network to learn the inherent attributes of individuals and project the original features into high-dimension hidden representations through the first two layers of the trained network. Then, we divide these hidden representations into KK groups through partitioning-based clustering, thus reformulating the problem as an integer stochastic programming problem under different total budgets. Finally, we distill the representation module and clustering model into a multi-category model to facilitate online deployment. Offline experiments validate the effectiveness and superiority of our approach compared to six state-of-the-art marketing optimization algorithms. Online A/B tests on the Meituan platform indicate that the approach outperforms the online algorithm by 0.53% and 0.65%, considering order volume (OV) and gross merchandise volume (GMV), respectively.

View on arXiv
@article{wang2025_2506.00959,
  title={ Hidden Representation Clustering with Multi-Task Representation Learning towards Robust Online Budget Allocation },
  author={ Xiaohan Wang and Yu Zhang and Guibin Jiang and Bing Cheng and Wei Lin },
  journal={arXiv preprint arXiv:2506.00959},
  year={ 2025 }
}
Comments on this paper