forked from ShivamShrirao/diffusers
-
Notifications
You must be signed in to change notification settings - Fork 1
/
dreambooth.yaml
95 lines (85 loc) · 2.17 KB
/
dreambooth.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
data:
resolution: 512
center_crop: false
concepts:
# You can add more concepts
- instance_set:
path: 'example/data/instance'
prompt: 'sks 1girl'
combine_prompt_from_txt: false
prompt_combine_template: '{PROMPT}, {TXT_PROMPT}'
class_set:
path: 'example/data/class'
prompt: '1girl'
combine_prompt_from_txt: false
prompt_combine_template: '{PROMPT}, {TXT_PROMPT}'
auto_generate:
enabled: true
negative_prompt: 'lowres, bad anatomy, bad hands, text, error, missing fingers, extra digit, fewer digits, cropped, worst quality, low quality, normal quality, jpeg artifacts, signature, watermark, username, blurry'
steps: 28
cfg_scale: 11
num_target: 100
batch_size: 1
sampling:
interval_steps: 50
batch_size: 1
concepts:
- prompt: 'sks 1girl, sitting'
negative_prompt: 'lowres, bad anatomy, bad hands, text, error, missing fingers, extra digit, fewer digits, cropped, worst quality, low quality, normal quality, jpeg artifacts, signature, watermark, username, blurry'
steps: 28
cfg_scale: 11
num_samples: 8
seed: 114514
batch_size: 2
seed: 114514
prior_preservation:
enabled: true
prior_loss_weight: 1.0
train_text_encoder: false
aspect_ratio_bucket:
enabled: false
debug: false
gradient_checkpointing: true
gradient_accumulation_steps: 1
gradient_clipping:
enabled: false
max_grad_norm: 1.0
mixed_precision: 'fp16'
cache_latents: true
clip_skip: 2
pad_tokens: true
saving:
#interval_steps: 1000
interval_epochs: 10
min_steps: 100
unet_half: false
monitoring:
monitors:
- 'wandb'
wandb:
sample: true
artifact: false
remove_ckpt_after_upload: false
optimizer:
name: bitsandbytes.optim.AdamW8bit
params:
lr: 5e-6
beta1: 0.9
beta2: 0.999
weight_decay: 1e-2
eps: 1e-8
lr_scale:
enabled: true
method: 'sqrt'
lr_scheduler:
name: torch.optim.lr_scheduler.CosineAnnealingWarmRestarts
params:
T_0: 10
T_mult: 1
eta_min: 7e-8
last_epoch: -1
warmup:
enabled: true
init_lr: 7e-8
steps: 100
strategy: 'linear'