-
Notifications
You must be signed in to change notification settings - Fork 27
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Question for your paper MSD #5
Comments
MSD is a bi-KD or mutual learning framework while the 'be your own teacher' is a one-way knowledge distillation method |
MSD is a bi-KD or mutual learning framework while the 'be your own teacher' is a one-way knowledge distillation method |
hi, |
Hi,
Thanks for sharing this code and it's really helpful.
Recently I read your paper:"MSD: Multi-Self-Distillation Learning via Multi-classifiers within Deep Neural Networks".It's a very interesting work and the results are much better than the paper "be your own teacher" which you reimplement here.
However,after reading your paper,I could just find some slight differences between this two papers:
1.the differences of bottleneck in the model.
2.some changes of hyper-parameter.
Is there some important details that I missed?And could you please tell me about the key difference between the two papers that lead to such a significant improvement?
The text was updated successfully, but these errors were encountered: