This is the repo for AAAI 2024 paper: Turning Dust into Gold: Distilling Complex Reasoning Capabilities from LLMs by Leveraging Negative Data. [Arxiv]
The repo contains:
- The synthetic data from ChatGPT and GPT4.
- The training and inference code for this work.
- The experimental results.
- Current works related to MATH dataset and math reasoning.
We provide the synthetic samples from GPT3.5-turbo/GPT4 through ICL on the MATH training set, which are saved in the data folder GPT3.5-turbo-MATH and GPT4-MATH.
For each sample, 8 samples are generated.
The demonstrations for generating rationales are in our paper.
The training and inference code are as follows:
prepare llama-7b checkpoint and store it in the code directory
prepare conda environment with requirements.txt
conda activate llm
training LoRA-neg
cd code
bash run_neg.sh
training LoRA-NAT
bash run_NAT.sh
training NCE
bash run_NCE.sh
training ASC
bash run_ASC.sh
We have also organized some work related to the MATH dataset and mathematical reasoning tasks to promote future research