Kohya SS FLUX LoRA training is way faster on Linux than Windows any ideas to debug? Same settings, libraries and GPU #9259
Unanswered
FurkanGozukara
asked this question in
Q&A
Replies: 1 comment 1 reply
-
This should be a discussion not an issue. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Describe the bug
I am using Kohya SS to train FLUX LoRA
On Linux RTX 3090 gets like 5.5 second / it - batch size 1 and 1024x1024 px resolution
On Windows RTX 3090 TI gets 7.7 second / it - has the most powerful CPU 13900 K
This speed dispercany is huge between Windows and Linux for some reason
Torch upgrade from 2.1 to 2.4 on Linux caused huge speed up and VRAM usage reduction but on Windows only VRAM usage dropped - speed same
Any ideas for how to fix? Using SDPA Cross Attention
I am sharing venv pip freeze of both Windows and Linux
Both has Python 3.10.11
Windows pip freeze
Ubuntu pip freeze
Who can help?
Stable Diffusion 3 / FLUX : @yiyixuxu @sayakpaul @DN6 @asomoza
Beta Was this translation helpful? Give feedback.
All reactions