Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Slow LoRA Loader with Block Weight #121

Closed
ai-waifu-life opened this issue Jul 12, 2024 · 6 comments
Closed

Slow LoRA Loader with Block Weight #121

ai-waifu-life opened this issue Jul 12, 2024 · 6 comments

Comments

@ai-waifu-life
Copy link

Hello! After running updates for ComfyUI and no other changes to any other drivers or requirements that I am aware of, the LoRA loaders are extremely slow. Usually it'll be slow for the first run, like a second or two per loader, but then blast through them for following generations. Now it is taking several seconds per loader and it does not seem to cache for following generations, so every one has to go through that. I replaced them with the standard loaders and those work normally, so seems to be something with Inspire's loaders. I set highvram and that didn't make any difference. Running on a 4070 Super and this is my first time encountering an issue like this.

@ltdrdata
Copy link
Owner

I'll check this later.

@ai-waifu-life
Copy link
Author

I'll check this later.

Just checking if you've had a sec to take a look. Lemme know if you need anything from my end!

@ltdrdata
Copy link
Owner

lbw-simple

Is this issue also being reproduced in this workflow? In my tests, while there is a slowdown when model patches are applied block by block compared to the default LoRA, even this process is instantaneous, so it's not to the extent that it would be problematic.

@ai-waifu-life
Copy link
Author

Is this issue also being reproduced in this workflow? In my tests, while there is a slowdown when model patches are applied block by block compared to the default LoRA, even this process is instantaneous, so it's not to the extent that it would be problematic.

I noticed that in that you have the seed set to fixed. If no changes are made to the LORA Loader at all, then it is quick after the first run. If I randomize the seed, like I always have, then it will treat it like a first process every time.

@ltdrdata
Copy link
Owner

Is this issue also being reproduced in this workflow? In my tests, while there is a slowdown when model patches are applied block by block compared to the default LoRA, even this process is instantaneous, so it's not to the extent that it would be problematic.

I noticed that in that you have the seed set to fixed. If no changes are made to the LORA Loader at all, then it is quick after the first run. If I randomize the seed, like I always have, then it will treat it like a first process every time.

It is correct to always re-execute when randomizing the seed. This is because the patch weights applied to the model change. That seed is a feature to randomize the 'r' when it is used in block weight.

@ai-waifu-life
Copy link
Author

It is correct to always re-execute when randomizing the seed. This is because the patch weights applied to the model change. That seed is a feature to randomize the 'r' when it is used in block weight.

Oh! Wait, so that could mean my workflow wasn't working correctly previously? 😂 Well that's great to know lol.
Appreciate your time on this!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants