10
0

Enhancing LLMs via High-Knowledge Data Selection

Abstract

The performance of Large Language Models (LLMs) is intrinsically linked to the quality of its training data. Although several studies have proposed methods for high-quality data selection, they do not consider the importance of knowledge richness in text corpora. In this paper, we propose a novel and gradient-free High-Knowledge Scorer (HKS) to select high-quality data from the dimension of knowledge, to alleviate the problem of knowledge scarcity in the pre-trained corpus. We propose a comprehensive multi-domain knowledge element pool and introduce knowledge density and coverage as metrics to assess the knowledge content of the text. Based on this, we propose a comprehensive knowledge scorer to select data with intensive knowledge, which can also be utilized for domain-specific high-knowledge data selection by restricting knowledge elements to the specific domain. We train models on a high-knowledge bilingual dataset, and experimental results demonstrate that our scorer improves the model's performance in knowledge-intensive and general comprehension tasks, and is effective in enhancing both the generic and domain-specific capabilities of the model.

View on arXiv
@article{duan2025_2505.14070,
  title={ Enhancing LLMs via High-Knowledge Data Selection },
  author={ Feiyu Duan and Xuemiao Zhang and Sirui Wang and Haoran Que and Yuqi Liu and Wenge Rong and Xunliang Cai },
  journal={arXiv preprint arXiv:2505.14070},
  year={ 2025 }
}
Comments on this paper