Turning Dust into Gold [AAAI 2024]
January 10, 2024 · View on GitHub

This is the repo for AAAI 2024 paper: Turning Dust into Gold: Distilling Complex Reasoning Capabilities from LLMs by Leveraging Negative Data. [Arxiv]
The repo contains:
- The synthetic data from ChatGPT and GPT4.
- The training and inference code for this work.
- The experimental results.
- Current works related to MATH dataset and math reasoning.
Data
We provide the synthetic samples from GPT3.5-turbo/GPT4 through ICL on the MATH training set, which are saved in the data folder GPT3.5-turbo-MATH and GPT4-MATH.
For each sample, 8 samples are generated.
The demonstrations for generating rationales are in our paper.
Code
The training and inference code are as follows:
step1:
prepare llama-7b checkpoint and store it in the code directory
step2:
prepare conda environment with requirements.txt
step3:
conda activate llm
step4:
training LoRA-neg
cd code
bash run_neg.sh
step5:
training LoRA-NAT
bash run_NAT.sh
step6:
training NCE
bash run_NCE.sh
step7:
training ASC
bash run_ASC.sh
Results

A list of work related to MATH and math reasoning
We have also organized some work related to the MATH dataset and mathematical reasoning tasks to promote future research