Existing methods for merging LLMs are memory-intensive and prone to task interference. We propose a framework that adaptively selects and fuses knowledge from multiple LLMs, enabling more scalable, stable, and memory-efficient integration—reducing interference by up to 50% compared to prior approaches.
conda create -n fusionx python=3.9
conda activate fusionx
pip3 install torch==2.1.2 torchvision==0.16.2 torchaudio==2.1.2 --index-url https://download.pytorch.org/whl/cu121
pip install -r requirements.txt
Please follow the detailed guidelines in FuseLLM
sbatch train_deepspeed_local_dynamic.script
git clone https://github.com/EleutherAI/lm-evaluation-harness.git
cd lm-evaluation-harness
pip3 install -e .
pip3 install omegaconf pycountry sentencepiece protobuf
bash run-lm-eval.sh
git clone https://github.com/allenai/open-instruct.git
cd open-instruct
pip install -r requirements.txt
./scripts/data/prepare_eval_data.sh
bash scripts/eval/bbh.sh
If you find our work useful in your research, please consider citing:
@article{kong2025enabling,
title={Enabling Flexible Multi-LLM Integration for Scalable Knowledge Aggregation},
author={Kong, Zhenglun and Zhan, Zheng and Hou, Shiyue and Gong, Yifan and Meng, Xin and Sui, Pengwei and Dong, Peiyan and Shen, Xuan and Wang, Zifeng and Zhao, Pu and others},
journal={arXiv preprint arXiv:2505.23844},
year={2025}
}
