Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Remove falcon style rope #34

Closed
wants to merge 1 commit into from
Closed

Conversation

magician-blue
Copy link
Contributor

All HF llama model are falcon style ROPE and we can convert them to original llama style ROPE with a permutation.
This pull request solve the bug when converting HF GQA to gguf format.
I learned idea from it and fix the similar bug in the llama2.c's exports.py.
Now I successfully convert Tinyllama-1.1B-chat to llama style ROPE. So, we can remove the falcon ROPE part.
I have upload the new export.py and llama2.mojo.

Details:
python export.py tl-chat.bin --hf PY007/TinyLlama-1.1B-Chat-v0.2 --version 0 to conver the model

@magician-blue
Copy link
Contributor Author

Why there is a difference is readme??

@magician-blue
Copy link
Contributor Author

I have update my model on huggingface.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants