Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

VQGAN training details #61

Closed
Andrew-Brown1 opened this issue Jun 25, 2021 · 3 comments
Closed

VQGAN training details #61

Andrew-Brown1 opened this issue Jun 25, 2021 · 3 comments

Comments

@Andrew-Brown1
Copy link

Hi,

Thanks for the great repo! Could I ask some questions about training VQGAN?

What batch size did you train it with, and for how long?
Also I see here that you wait until you add the discriminator loss https://www.youtube.com/watch?v=fy153-yXSQk

Do you wait until the model has converged without it before adding it?

Thanks!

@rromb
Copy link
Collaborator

rromb commented Sep 28, 2021

Hi, great question :) Most of our published VQGAN models are trained on a single 40GB VRAM GPU with a batch size of ~12 (bs=14 for the f16 model), depending on the hyperparameters. Regarding your second question, yes, it makes sense to monitor the perceptual loss and then add the discriminator to the training loop.

@rromb rromb closed this as completed Sep 28, 2021
@Andrew-Brown1
Copy link
Author

Hey - thanks!

@gombru
Copy link

gombru commented Oct 20, 2022

Hi! Can we get any details about training costs? I.e. how many epochs and gpu days took to train the OpenImages model on a single 40GB VRAM GPU?

Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants