Skip to content
Stanford Alpaca edited this page Mar 29, 2023 · 1 revision

Welcome to the Stanford Alpaca GPT!

Alpaca aims to build and share an instruction-following LLaMA model. The repo contains:

The 52K data used for fine-tuning the model. The code for generating the data. The code for fine-tuning the model.

The current Alpaca model is fine-tuned from a 7B LLaMA model [1] on 52K instruction-following data generated by the techniques in the Self-Instruct [2] paper, with some modifications that we discuss in the next section. In a preliminary human evaluation, we found that the Alpaca 7B model behaves similarly to the text-davinci-003 model on the Self-Instruct instruction-following evaluation suite [2].

Clone this wiki locally