149
0

Do BERT-Like Bidirectional Models Still Perform Better on Text Classification in the Era of LLMs?

Main:5 Pages
3 Figures
Bibliography:1 Pages
4 Tables
Appendix:2 Pages
Abstract

The rapid adoption of LLMs has overshadowed the potential advantages of traditional BERT-like models in text classification. This study challenges the prevailing "LLM-centric" trend by systematically comparing three category methods, i.e., BERT-like models fine-tuning, LLM internal state utilization, and zero-shot inference across six high-difficulty datasets. Our findings reveal that BERT-like models often outperform LLMs. We further categorize datasets into three types, perform PCA and probing experiments, and identify task-specific model strengths: BERT-like models excel in pattern-driven tasks, while LLMs dominate those requiring deep semantics or world knowledge. Based on this, we propose TaMAS, a fine-grained task selection strategy, advocating for a nuanced, task-driven approach over a one-size-fits-all reliance on LLMs.

View on arXiv
@article{zhang2025_2505.18215,
  title={ Do BERT-Like Bidirectional Models Still Perform Better on Text Classification in the Era of LLMs? },
  author={ Junyan Zhang and Yiming Huang and Shuliang Liu and Yubo Gao and Xuming Hu },
  journal={arXiv preprint arXiv:2505.18215},
  year={ 2025 }
}
Comments on this paper

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.