Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Results are not reproducible on Qwen1.5-1.8B #92

Open
Fantasy1120 opened this issue Jul 2, 2024 · 0 comments
Open

Results are not reproducible on Qwen1.5-1.8B #92

Fantasy1120 opened this issue Jul 2, 2024 · 0 comments

Comments

@Fantasy1120
Copy link

Hi, I tried training with Qwen1.5-1.8, and I found that the results varied greatly each time.
For example, I trained three times and the corresponding evaluations are 46,61,55 (GQA); 37,53,43 (TextVQA).
I have followed your default training settings, i.e, global batch size, lr and conv_version, and I'd like to know what caused such a big difference?

Additionally, I would like to ask if I want to add a new LLM, then how can I find the template for this LLM? For example the template for Qwen2.

Thanks for your answer.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant