InfiFusion: A Unified Framework for Enhanced Cross-Model Reasoning via LLM Fusion

We introduce InfiFusion, an efficient training pipeline designed to integrate multiple domain-specialized Large Language Models (LLMs) into a single pivot model, effectively harnessing the strengths of each source model. Traditional fusion methods either merge model parameters directly or rely on knowledge distillation with rigid assumptions, limiting their flexibility and efficiency. InfiFusion overcomes these limitations by enhancing Universal Logit Distillation (ULD) with Top-K selection and Logits Standardization. We propose two fusion strategies: Pairwise Fusion (InfiFusion), where each source model knowledge is distilled individually into the pivot model followed by merging and Unified Fusion (InfiFusion), where knowledge from all source models is distilled simultaneously into the pivot model. InfiFusion outperforms the state-of-the-art models, such as Qwen-2.5-14B-Instruct and Phi-4, across 11 widely applied benchmarks covering reasoning, coding, mathematics, and instruction-following tasks. Notably, InfiFusion achieves this superior performance while significantly reduces computational costs, completing full training with only 160 H800 GPU hours compared to the millions typically required for traditional LLM training.
View on arXiv@article{yan2025_2501.02795, title={ InfiFusion: A Unified Framework for Enhanced Cross-Model Reasoning via LLM Fusion }, author={ Zhaoyi Yan and Yiming Zhang and Baoyi He and Yuhao Fu and Qi Zhou and Zhijie Sang and Chunlin Ji and Shengyu Zhang and Fei Wu and Hongxia Yang }, journal={arXiv preprint arXiv:2501.02795}, year={ 2025 } }