Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

guide on how to run it on windows #56

Open
mhnoni opened this issue Oct 22, 2022 · 12 comments
Open

guide on how to run it on windows #56

mhnoni opened this issue Oct 22, 2022 · 12 comments

Comments

@mhnoni
Copy link

mhnoni commented Oct 22, 2022

can't make it run on windows natively, getting this error after running python train_dreambooth.py

train_dreambooth.py: error: the following arguments are required: --pretrained_m
odel_name_or_path, --pretrained_vae_name_or_path, --instance_data_dir

maybe I'm doing something wrong? the only difference I made is to run python train_dreambooth.py instead of launch.sh for Linux.

@NeoAnthropocene
Copy link

can't make it run on windows natively, getting this error after running python train_dreambooth.py

train_dreambooth.py: error: the following arguments are required: --pretrained_m odel_name_or_path, --pretrained_vae_name_or_path, --instance_data_dir

maybe I'm doing something wrong? the only difference I made is to run python train_dreambooth.py instead of launch.sh for Linux.

@mhnoni
I would suggest watching the video on the below link and read carefully the updates on the video description. You can run it locally by following the steps carefully in that video.

(https://youtu.be/w6PTviOCYQY)

@mhnoni
Copy link
Author

mhnoni commented Oct 23, 2022

can't make it run on windows natively, getting this error after running python train_dreambooth.py
train_dreambooth.py: error: the following arguments are required: --pretrained_m odel_name_or_path, --pretrained_vae_name_or_path, --instance_data_dir
maybe I'm doing something wrong? the only difference I made is to run python train_dreambooth.py instead of launch.sh for Linux.

@mhnoni I would suggest watching the video on the below link and read carefully the updates on the video description. You can run it locally by following the steps carefully in that video.

(https://youtu.be/w6PTviOCYQY)

That one uses wsl to install Linux. I would like to use it natively on windows without Linux which seems possible but no guide on how to run the last command on windows, I mean instead of launch.sh we should have a python file to run train_dreambooth.py but not idea how to do that.

@ThereforeGames
Copy link

Here you go: https://www.reddit.com/r/StableDiffusion/comments/ydip3s/guide_dreambooth_training_with_shivamshriraos/

@mhnoni
Copy link
Author

mhnoni commented Oct 26, 2022

Here you go: https://www.reddit.com/r/StableDiffusion/comments/ydip3s/guide_dreambooth_training_with_shivamshriraos/

oh wow, thank you so much!

@InB4DevOps
Copy link

@ShivamShrirao can we please have a Wiki in this repo where we could document things like this?

@ShivamShrirao
Copy link
Owner

@InB4DevOps yeah I wanted to but I haven't been able to verify them myself. Like the one linked here doesn't install xformers. Also the devs for xformers and bitsandbytes have been working on getting them officially working on windows.
Will add a section where I can add these unverified resources until then.

@InB4DevOps
Copy link

@ShivamShrirao a wiki would be nice to have nonetheless. For all the other knowledge surrounding DreamBooth with SD..

@mhnoni
Copy link
Author

mhnoni commented Oct 27, 2022

regarding xformers
from auto111 code:
pip install https://github.com/C43H66N12O12S2/stable-diffusion-webui/releases/download/c/xformers-0.0.14.dev0-cp310-cp310-win_amd64.whl

needs python 10.
I was able to install it, but I can't confirm if it will work or not since I only have 8GBVram. waiting for deepspeed.

@patrickgalbraith
Copy link

patrickgalbraith commented Nov 7, 2022

This is what I did to get it working.

environment.yml

name: dbshiv
channels:
  - conda-forge
  - pytorch
  - nvidia
  - defaults
dependencies:
  - python>=3.10
  - pip
  - pytorch==1.12.1
  - cudatoolkit=11.6.0
  - numpy
  - pip:
    - accelerate==0.12.0
    - torchvision==0.13.1
    - torchaudio==0.12.1
    - ftfy
    - tensorboard
    - modelcards
    - bitsandbytes
    - transformers>=4.21.0
    - pyre-extensions==0.0.23
    - -e git+https://github.com/ShivamShrirao/diffusers.git@main#egg=diffusers

train.bat

set HUGGINGFACE_TOKEN="PUT YOUR TOKEN HERE"
set INSTANCE_NAME="dboperson"
set CLASS_NAME="person"
 
set INSTANCE_DIR="./dataset/source_images/dboperson"
set OUTPUT_DIR="./output/dboperson"
set CLASS_DIR="./dataset/class_images/person"
 
set NUM_INSTANCE_IMAGES=10
set /a NUM_CLASS_IMAGES=%NUM_INSTANCE_IMAGES%*12
set /a MAX_NUM_STEPS = %NUM_INSTANCE_IMAGES%*80
set LR_SCHEDULE="polynomial"
set /a LR_WARMUP_STEPS=%MAX_NUM_STEPS%/10
 
set INSTANCE_PROMPT="%INSTANCE_NAME% %CLASS_NAME%"
 
call T:/programs/anaconda3/Scripts/activate.bat
call conda activate dbshiv
 
accelerate launch train_dreambooth.py ^
  --pretrained_model_name_or_path="runwayml/stable-diffusion-v1-5" ^
  --pretrained_vae_name_or_path="stabilityai/sd-vae-ft-mse" ^
  --tokenizer_name=%TOKENIZER_NAME% ^
  --instance_data_dir=%INSTANCE_DIR% ^
  --class_data_dir=%CLASS_DIR% ^
  --output_dir=%OUTPUT_DIR% ^
  --with_prior_preservation --prior_loss_weight=1.0 ^
  --instance_prompt=%INSTANCE_PROMPT% ^
  --class_prompt=%CLASS_NAME% ^
  --seed=1337 ^
  --resolution=512 ^
  --train_batch_size=1 ^
  --train_text_encoder ^
  --mixed_precision="fp16" ^
  --gradient_accumulation_steps=1 ^
  --learning_rate=1e-6 ^
  --lr_scheduler=%LR_SCHEDULE% ^
  --lr_warmup_steps=%LR_WARMUP_STEPS% ^
  --num_class_images=%NUM_CLASS_IMAGES% ^
  --sample_batch_size=4 ^
  --max_train_steps=%MAX_NUM_STEPS% ^
  --not_cache_latents ^
  --save_interval=250
pause

Setup notes

git clone https://github.com/ShivamShrirao/diffusers ShivamShriraoDiffusers
cd ShivamShriraoDiffusers/examples/dreambooth

# Note the pytorch version used is important since newer versions don't work right now
conda env create -f environment.yml
conda activate dbshiv

# Download https://github.com/C43H66N12O12S2/stable-diffusion-webui/releases/download/f/xformers-0.0.14.dev0-cp310-cp310-win_amd64.whl
pip install -U -I --no-deps xformers-0.0.14.dev0-cp310-cp310-win_amd64.whl

# Run accelerate config
#  In which compute environment are you running? ([0] This machine, [1] AWS (Amazon SageMaker)): 0
#  Which type of machine are you using? ([0] No distributed training, [1] multi-CPU, [2] multi-GPU, [3] TPU [4] MPS): 0
#  Do you want to run your training on CPU only (even if a GPU is available)? [yes/NO]:no
#  Do you want to use DeepSpeed? [yes/NO]: no
#  Do you wish to use FP16 or BF16 (mixed precision)? [NO/fp16/bf16]: FP16
accelerate config

# Make sure GPU is enabled in the follow output, if not then pytorch install is not correct
accelerate env

# To train update and run
train.bat

# To convert to checkpoint
python convert_diffusers_to_original_stable_diffusion.py --model_path="./output/out" --checkpoint_path="./output/checkpoint.ckpt"

@NeoAnthropocene
Copy link

@patrickgalbraith , thanks for sharing this clean code.
I wonder why you are using the LR Scheduler with the "polynomial". Did you get better results compared with "constant" parameter?

I don't have deep knowledge about it but I'm trying to find out the best results by trying different parameters that I saw from other examples around the internet. I just wanted to know if it worth to spend time on it or not.

Thanks.

@InB4DevOps
Copy link

@patrickgalbraith , thanks for sharing this clean code. I wonder why you are using the LR Scheduler with the "polynomial". Did you get better results compared with "constant" parameter?

I don't have deep knowledge about it but I'm trying to find out the best results by trying different parameters that I saw from other examples around the internet. I just wanted to know if it worth to spend time on it or not.

Thanks.

Hey,

I did a comparison of all learning rate schedulers.
Read about it here:
https://www.reddit.com/r/StableDiffusion/comments/yd56cy/dreambooth_i_compared_all_learning_rate/

TL;DR: use constant

@NeoAnthropocene
Copy link

NeoAnthropocene commented Nov 10, 2022

@InB4DevOps >

Thank you so much. This will save so much time :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

6 participants