ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2502.12507
47
0

Mixture of Attention Yields Accurate Results for Tabular Data

18 February 2025
Xuechen Li
Yupeng Li
Jian Liu
Xiaolin Jin
Tian Yang
Xin Hu
ArXivPDFHTML
Abstract

Tabular data inherently exhibits significant feature heterogeneity, but existing transformer-based methods lack specialized mechanisms to handle this property. To bridge the gap, we propose MAYA, an encoder-decoder transformer-based framework. In the encoder, we design a Mixture of Attention (MOA) that constructs multiple parallel attention branches and averages the features at each branch, effectively fusing heterogeneous features while limiting parameter growth. Additionally, we employ collaborative learning with a dynamic consistency weight constraint to produce more robust representations. In the decoder stage, cross-attention is utilized to seamlessly integrate tabular data with corresponding label features. This dual-attention mechanism effectively captures both intra-instance and inter-instance interactions. We evaluate the proposed method on a wide range of datasets and compare it with other state-of-the-art transformer-based methods. Extensive experiments demonstrate that our model achieves superior performance among transformer-based methods in both tabular classification and regression tasks.

View on arXiv
@article{li2025_2502.12507,
  title={ Mixture of Attention Yields Accurate Results for Tabular Data },
  author={ Xuechen Li and Yupeng Li and Jian Liu and Xiaolin Jin and Tian Yang and Xin Hu },
  journal={arXiv preprint arXiv:2502.12507},
  year={ 2025 }
}
Comments on this paper