Replies: 1 comment 2 replies
-
Even with your 80GB of memory on an H100, you will loos cuda memory. The model size, resolution, and the memory needed for training like storing activations and gradients are too high for the available memory. Could you try with turn on gradient chckpt and mixed precision? |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I try to fine tune full-parameter of flux on H100 (80G cuda memory), with batch size=1 and resolution = 512.
But it crashed with CUDA Out of Memory. So I'm wondering how much memory it need.
I'm using the dream booth code and didn't change anything of it.
Beta Was this translation helpful? Give feedback.
All reactions