Skip to content

SuperBruceJia/MetaMath-Fine-Tune-with-LoRA

Repository files navigation

Fine-tune LLaMA 2 (7B) with LoRA on meta-math/MetaMathQA Dataset

Fine-tuning and Inference codes on the MetaMath Dataset

Code License License Python 3.9+

Model Details

MetaMath-Fine-Tune-with-LoRA is trained to reason and answer mathematical problems on meta-math/MetaMathQA dataset. We used meta-llama/Llama-2-7b-hf as a base model and used LoRA to fine-tune it.

Model Description

Results on GSM8K

Epoch Accuracy on the testing set Model Link
$1$ $0.609$ 🤗 Hugging Face
$2$ $0.635$ 🤗 Hugging Face
$3$ $0.641$ 🤗 Hugging Face
$4$ $0.641$ 🤗 Hugging Face

Deployment

# Load the Pre-trained LoRA Adapter
model.load_adapter("shuyuej/metamath_lora_llama2_7b_4_epoch")
model.enable_adapters()

About

Fine-tuning and Inference codes of the MetaMath Dataset

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published