Skip to content

Commit

Permalink
adding flash attention to example config
Browse files Browse the repository at this point in the history
  • Loading branch information
benjaminye committed Apr 9, 2024
1 parent 3837366 commit a938550
Showing 1 changed file with 3 additions and 1 deletion.
4 changes: 3 additions & 1 deletion config.yml
Original file line number Diff line number Diff line change
Expand Up @@ -24,6 +24,8 @@ data:
# Model Definition -------------------
model:
hf_model_ckpt: "NousResearch/Llama-2-7b-hf"
torch_dtype: "bfloat16"
attn_implementation: "flash_attention_2"
quantize: true
bitsandbytes:
load_in_4bit: true
Expand Down Expand Up @@ -80,4 +82,4 @@ qa:
- verb_percent
- adjective_percent
- noun_percent
- summary_length
- summary_length

0 comments on commit a938550

Please sign in to comment.