Replies: 1 comment 6 replies
-
no it should be more like 7 seconds / step on an A6000 but 4 gradient accumulations is probably too many. set it to just 1 or 2. it linearly increases the runtime. |
Beta Was this translation helpful? Give feedback.
6 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi all,
I'm just wondering how many seconds/iteration is somewhat reasonable to see for FLUX LoRA training. I am on two RTX A6000s, and using batch size 8 with 4 steps of gradient accumulation for resolution 1024x1024, I can get 30 seconds/iter. Is this somewhat reasonable?
I think A6000s should support bf16 well, but a bit scared because of my failures at SD3 LoRA training here (even though I upgraded from RTX8000s to A6000s).
Thanks in advance.
Beta Was this translation helpful? Give feedback.
All reactions