-
Notifications
You must be signed in to change notification settings - Fork 0
/
default_config.yaml
118 lines (102 loc) · 3.11 KB
/
default_config.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
# Builtin Configurations (DO NOT CHANGE THESE CONFIGURATIONS unless you know exactly what you are doing)
enable_modelarts: True
data_url: ""
multi_data_url: ""
train_url: ""
ckpt_url: ""
result_url: ""
data_path: "/home/work/user-job-dir/data/"
output_path: "/home/work/user-job-dir/model/"
load_path: "/home/work/user-job-dir/checkpoint.ckpt"
result_path: "" # "/home/work/user-job-dir/result/"
checkpoint_path: ""
device_target: Ascend
enable_profiling: False
enable_graph_kernel: False # GPU: True, Ascend: False
# ==============================================================================
modelarts_dataset_unzip_name: 'BiT-M-R101x3_cifar-10-binary' # Only used on ChinaComputingNet(Beta)
need_modelarts_dataset_unzip: False
model_name: 'BiT-M-R101x3'
dataset_name: "cifar-10-batches-bin"
num_classes: 10
image_height: 160
image_width: 160
batch_size: 64
num_parallel_workers: 16
rand_augment_num_layers: 2
rand_augment_magnitude: 20
mixup_alpha: 0.5
num_epochs: 200
lr_init: 0.02
lr_decay: 0.98
lr_warmup_epochs: 6.0
lr_num_epochs_per_decay: 2.5
momentum: 0.9
decay: 0.9
weight_decay: 0.00004
enable_ema: True
ema_decay: 0.9999
enable_clip_norm: True
gradient_norm: 1.0
label_smoothing: 0.1
# loss_scale: 1024
save_checkpoint: True
save_checkpoint_epochs: 10
keep_checkpoint_max: 80
save_checkpoint_path: "/home/work/user-job-dir/model/"
device_id: 0
rank_id: 0
rank_size: 1
run_distribute: True
run_eval: True
activation: "Softmax"
# Image classification train
dataset_path: ""
pretrain_ckpt: ""
freeze_layer: ""
filter_head: False
enable_cache: False
cache_session_id: ""
is_training: True
# CoAtNet eval
is_training_eval: False
run_distribute_eval: False
# CoAtNet export
device_id_export: 0
batch_size_export: 1
ckpt_file: ""
file_name: "BiT"
file_format: "MINDIR"
is_training_export: False
run_distribute_export: False
# postprocess.py / CoAtNet acc calculation
batch_size_postprocess: 1
# result_path: ''
label_path: ''
---
# Config description for each option
enable_modelarts: 'Whether training on modelarts, default: False'
data_url: 'Dataset url for obs'
train_url: 'Training output url for obs'
data_path: 'Dataset path for local'
output_path: 'Training output path for local'
ann_file: 'Ann file, default is val.json.'
pretrain_ckpt: 'Pretrained checkpoint path for fine tune or incremental learning'
platform: 'Target device type'
freeze_layer: 'freeze the weights of network from start to which layers'
filter_head: 'Filter head weight parameters when load checkpoint, default is False.'
enable_cache: 'Caching the dataset in memory to speedup dataset processing, default is False.'
cache_session_id: 'The session id for cache service.'
file_name: "output file name."
result_path: "result files path."
label_path: "label path."
enable_profiling: 'Whether enable profiling while training, default: False'
run_eval: 'Whether run evaluation while training, default is false.'
run_distribute: 'Run distribute, default is false.'
device_id: 'Device id, default is 0.'
rank_id: 'Rank id, default is 0.'
file_format: 'file format'
---
platform: ['Ascend', 'GPU', 'CPU']
file_format: ["AIR", "ONNX", "MINDIR"]
freeze_layer: ["", "none", "backbone"]