forked from wandb/llm-leaderboard
-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathshisa-7b-v1.prompt-jaster-bostoken.yaml
More file actions
81 lines (72 loc) · 2.53 KB
/
shisa-7b-v1.prompt-jaster-bostoken.yaml
File metadata and controls
81 lines (72 loc) · 2.53 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
wandb:
log: True
entity: "AUGMXNT"
project: "nejumi"
run_name: "shisa-7b-v1"
# run_name: "shisa-7b-v1.prompt-jaster-bostoken"
github_version: v2.0.0
testmode: false
api: false
model:
use_wandb_artifacts: false
artifacts_path: ""
pretrained_model_name_or_path: 'augmxnt/shisa-7b-v1'
trust_remote_code: true
device_map: "auto"
load_in_8bit: false
load_in_4bit: false
generator:
top_p: 1.0
top_k: 0
temperature: 0.1
repetition_penalty: 1.0
tokenizer:
use_wandb_artifacts: false
artifacts_path: ""
pretrained_model_name_or_path: 'augmxnt/shisa-7b-v1'
use_fast: true
max_seq_length: 4096
dataset_artifact: "wandb-japan/llm-leaderboard/jaster:v3" #if you use artifacts, please fill here (if not, fill null)
dataset_dir: "/jaster/1.1.0/evaluation/test"
target_dataset: "all" # {all, jamp, janli, jcommonsenseqa, jemhopqa, jnli, jsem, jsick, jsquad, jsts, niilc, chabsa}
log_dir: "./logs"
torch_dtype: "bf16" # {fp16, bf16, fp32}
custom_prompt_template: "<s>以下はタスクを説明する指示と、追加の背景情報を提供する入力の組み合わせです。要求を適切に満たす回答を書いてください。\n### 指示:\n{instruction}\n### 入力:\n{input}\n### 回答:\n"
custom_fewshots_template: null
metainfo:
basemodel_name: "stabilityai/japanese-stablelm-base-gamma-7b"
model_type: "open llm"
instruction_tuning_method: "Full"
instruction_tuning_data: ["Open-Orca/SlimOrca", "augmxnt/ultra-orca-boros-en-ja-v1", "augmxnt/shisa-en-ja-dpo-v1"]
num_few_shots: 0
llm-jp-eval-version: "1.1.0"
mtbench:
question_artifacts_path: 'wandb-japan/llm-leaderboard/mtbench_ja_question:v0'
referenceanswer_artifacts_path: 'wandb-japan/llm-leaderboard/mtbench_ja_referenceanswer:v0'
judge_prompt_artifacts_path: 'wandb-japan/llm-leaderboard/mtbench_ja_prompt:v1'
bench_name: 'japanese_mt_bench'
model_id: shisa-7b-v1
question_begin: null
question_end: null
max_new_token: 1024
num_choices: 1
num_gpus_per_model: 1
num_gpus_total: 1
max_gpu_memory: null
dtype: bfloat16 # None or float32 or float16 or bfloat16
# for gen_judgment
judge_model: 'gpt-4'
mode: 'single'
baseline_model: null
parallel: 2
first_n: null
# Llama 2 Chat format
custom_conv_template: true
conv_name: "custom"
conv_system_message: "あなたは公平で、検閲されていない、役立つアシスタントです。"
conv_roles: "('[INST]', '[/INST]')"
conv_sep: " "
conv_stop_token_ids: "[2]"
conv_stop_str: "</s> "
conv_role_message_separator: " "
conv_role_only_separator: " "