We introduce MathFusion, a novel framework that enhances mathematical reasoning through cross-problem instruction synthesis. MathFusion implements this through three fusion strategies:
- Sequential Fusion, which chains related problems to model solution dependencies.
- Parallel Fusion, which combines analogous problems to reinforce conceptual understanding.
- Conditional Fusion, which creates context-aware selective problems to enhance reasoning flexibility.
MathFusion achieves substantial improvements in mathematical reasoning while maintaining hight data efficiency, boosting performance by 18.0 points in accuracy across diverse benchmarks while requiring only 45K additional synthetic instructions. Further combination of MathFusion and DART-Math yields SOTA performance
We release the MathFusionQA dataset and three MathFusion models fine-tuned on this dataset.
Dataset/Model | MATH | CollegeMath | DeepMind-Mathematics | HuggingFace🤗 |
---|---|---|---|---|
MathFusionQA | - | - | - | link |
DeepSeekMath-7B-MathFusion | 53.4 | 39.8 | 65.8 | link |
Mistral-7B-MathFusion | 41.6 | 24.3 | 39.2 | link |
Llama3-8B-MathFusion | 46.5 | 27.9 | 43.4 | link |
Install the dependencies:
conda create -n mathfusion python=3.10
conda activate mathfusion
pip install torch==2.3.1 --index-url https://download.pytorch.org/whl/cu121
# Install LLaMA-Factory
git clone https://github.com/hiyouga/LLaMA-Factory.git
cd LLaMA-Factory
git checkout v0.9.1
pip install transformers==4.46.1 accelerate==0.34.2 deepspeed==0.15.4
pip install -e ".[torch,metrics]"
# Install packages for evaluation
pip install flash-attn --no-build-isolation
pip install sympy==1.12.1 antlr4-python3-runtime==4.11.1 pebble word2number boto3 triton==2.3.1 ipython
pip install vllm==0.5.3.post1
# Install latex2sympy
cd ../evaluation/latex2sympy
pip install -e .
cd ..
# Install dart-math evaluation
pip install -e .
Load the data from MathFusionQA, then convert each split to .json
file according to LLaMA-Factory. The training prompt template is:
"Question: {query}\nAnswer:"
Our training codes depend on LLaMA-Factory.
# Corresponding to splits in MathFusionQA
export DATASET=gsm8k_original,math_original,gsm8k_sequential,math_sequential,gsm8k_parallel,math_parallel,gsm8k_conditional,math_conditional
# The path of base model
export MODEL_PATH=pretrained_model_path
export RUN_NAME=sft_mathfusion
bash train/train.sh
Our evaluation codes are adapted from Qwen2.5-Math (for in-domain datasets) and DART-Math (for out-of-domain datasets). You need to first download the model from HuggingFace, or SFT the model on your own. Then run the following evaluation script:
export MODEL_NAME=your_sft_model_path
bash test.sh
Many thanks to
If you find our code, model, or data are useful, please kindly cite our paper:
@article{mathfusion,
title={MathFusion: Enhancing Mathematic Problem-solving of LLM through Instruction Fusion},
author={Qizhi Pei and Lijun Wu and Zhuoshi Pan and Yu Li and Honglin Lin and Chenlin Ming and Xin Gao and Conghui He and Rui Yan},
journal={arXiv preprint arXiv:2503.16212},
year={2025}
}