-
Notifications
You must be signed in to change notification settings - Fork 0
Home
Stanford Alpaca edited this page Mar 29, 2023
·
1 revision
Welcome to the Stanford Alpaca GPT!
Alpaca aims to build and share an instruction-following LLaMA model. The repo contains:
The 52K data used for fine-tuning the model. The code for generating the data. The code for fine-tuning the model.
The current Alpaca model is fine-tuned from a 7B LLaMA model [1] on 52K instruction-following data generated by the techniques in the Self-Instruct [2] paper, with some modifications that we discuss in the next section. In a preliminary human evaluation, we found that the Alpaca 7B model behaves similarly to the text-davinci-003 model on the Self-Instruct instruction-following evaluation suite [2].