Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

You do not use NativeScaler and ModelEma for training? #1

Open
wangpichao opened this issue Apr 28, 2021 · 3 comments
Open

You do not use NativeScaler and ModelEma for training? #1

wangpichao opened this issue Apr 28, 2021 · 3 comments

Comments

@wangpichao
Copy link

You do not use NativeScaler and ModelEma for training?

@danczs
Copy link
Owner

danczs commented Apr 29, 2021

Yeah, we do not use amp, beacause it causes NAN values sometimes. Thus we removed amp and Scaler in our code. See also in deit issue: facebookresearch/deit#29. Without this acceleration,our code takes more time during training than Deit. But amp can be freely used after training, so the the inference is as efficient.
For ModelEma, we do not use it as in Deit.
Thanks for your question.

@m-aliabbas
Copy link

Hello, I am getting the error.
NameError: name 'model_ema' is not defined
How can I train the model without it. Please help

@danczs
Copy link
Owner

danczs commented Apr 29, 2021

Hello, I am getting the error.
NameError: name 'model_ema' is not defined
How can I train the model without it. Please help

Sorry about this error. The code of Scaler and model_ema is not cleared completely . We have fixed it by removing the code in main.py(ling 287,288)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants