Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

transformer results #2

Closed
homelifes opened this issue Apr 21, 2020 · 3 comments
Closed

transformer results #2

homelifes opened this issue Apr 21, 2020 · 3 comments

Comments

@homelifes
Copy link

homelifes commented Apr 21, 2020

Hello. Thanks for your work and for sharing the code.
May you please tell me the details of the pure Transformer model you implemented which achieves 128.3 cider? To best of my knowledge, all implementations could achieve a maximum of around 126.6, according to all papers which utilized the transformer model. In your paper, you don't provide details on Transformer, and there is no supplementary material. So may I kindly know the details for your re-implementation of the pure transformer which achieves 128.3?

@Panda-Peter
Copy link
Collaborator

The implementation of Transformer can be referred from https://github.com/ruotianluo/self-critical.pytorch, which can achieve ~128 CIDEr score.

@homelifes
Copy link
Author

homelifes commented Apr 21, 2020

Hi @Panda-Peter. Thanks for your reply. I am actually following his code, but according to the results here, it can achieve 1.266 with self-critical. It can achieve 1.295 but with the new self-critical proposed, which you do not use. So his reported score for transformer is 1.266 (Transformer+self_critical | 1.266), may I know how you achieved 1.283 and what changes you made in his code? Thanks a lot for your kind help

@Panda-Peter
Copy link
Collaborator

We also implement the baseline of Transformer based on this code. However, we found the primary hyper-parameters are not optimal. You can tune the parameters and obtain ~1.283 cider.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants