-
Notifications
You must be signed in to change notification settings - Fork 23
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Question about the training parameters #1
Comments
Yes, we trained 100 epochs. In fact, I have tried the tf implementation, it dit not work well and the batch size could only be set to a small value and the training speed is slow. So i guess some details are different from the author. We used cosine lr in the experiment. |
@feymanpriv , thank you for your reply. Your code is very clear to read, thanks for your work. I still have some questions:
|
@zjcs thank you for your attention
|
Your reply is very helpful to me, thank you very much, I will try later. Have a nice day~ |
Hello, @feymanpriv , thank you very much for your work.
I am trying to review your code, here I want to know some key important parameters in your work:
As claimed in delg paper, only 15M step(about 25epoch) is needed in GLDv2Clean-Train-80percent dataset, while the max epoch in your config file is 100, do you train 100 epochs finnally to get your result?
Is there any modification in your implemention which is different to origin implemention by tensorflow ?
The text was updated successfully, but these errors were encountered: