Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[P1] Confirmation of alpaca_eval version #73

Closed
BaohaoLiao opened this issue May 1, 2024 · 4 comments
Closed

[P1] Confirmation of alpaca_eval version #73

BaohaoLiao opened this issue May 1, 2024 · 4 comments
Assignees
Labels
question Further information is requested

Comments

@BaohaoLiao
Copy link

Hi,

I want to confirm the version of alpaca_eval. In the paper and the README, you said you use v1.0. May I ask how you install it? If I install as your recommendation:
pip install alpaca-eval

It is the latest version. Should I install as:

pip install alpaca-eval==v0.1.0

@frankaging frankaging self-assigned this May 2, 2024
@frankaging frankaging added the question Further information is requested label May 2, 2024
@frankaging
Copy link
Collaborator

frankaging commented May 2, 2024

@BaohaoLiao hey thanks for your question! the version here is okay. You can install the newest version.

Alpaca-Evalv1.0 means we evaluate against text-davinci-003, not against GPT4. You can follow this to do the evaluation:
https://github.com/stanfordnlp/pyreft/tree/main/examples/loreft#offline-evaluation-with-alpaca-eval-v10

Alpaca-Evalv2.0 uses GPT-4. To be competitive with v2.0, please train with longer sequence length. We are training with max sequence length of 768, this is too limited. We have to do it to keep a fair comparison with previous works.

@frankaging frankaging changed the title Confirmation of alpaca_eval version [P1] Confirmation of alpaca_eval version May 2, 2024
@frankaging
Copy link
Collaborator

I am closing this issue; feel free to reopen if you have more questions! ty!

@BaohaoLiao
Copy link
Author

Thank you for the details. Is it possible to release the generated outputs for alpaca_eval? I.e. the output from Llama-2 7B & LoReFT (ours) in Table 3.

It would be great if you could also release the evaluation results from alpaca_eval.

@frankaging
Copy link
Collaborator

@BaohaoLiao unfortunately, these results are removed locally from my disk -- but it should be very easy to reproduce. i think you just need to follow the command we give in the README. let me know if you encounter any issues.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants