Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Exllamav2 lora support #4229

Merged
merged 9 commits into from
Oct 14, 2023
Merged

Exllamav2 lora support #4229

merged 9 commits into from
Oct 14, 2023

Conversation

Ph0rk0z
Copy link
Contributor

@Ph0rk0z Ph0rk0z commented Oct 8, 2023

Doesn't work for multi GPU because of a kernel bug but it does work on single card models. Tested it for llama 7b.

Somehow multiple loras are possible but I'm not sure how to make that work.

@Ph0rk0z
Copy link
Contributor Author

Ph0rk0z commented Oct 9, 2023

Its working for 70b now.

@oobabooga
Copy link
Owner

Nice. I'm waiting for a 0.0.6 release to merge this.

@oobabooga
Copy link
Owner

It should be working for multiple LoRAs now (if you find any issue, please let me know)

@oobabooga oobabooga merged commit 8cce1f1 into oobabooga:main Oct 14, 2023
@noiraku
Copy link

noiraku commented Oct 14, 2023

It should be working for multiple LoRAs now (if you find any issue, please let me know)

Mistral + Lora works with exllamav2. This fixes #4272 Thank you!
(Exllamav1 still same error but no need when v2 faster.)

@noiraku noiraku mentioned this pull request Oct 14, 2023
1 task
@Ph0rk0z Ph0rk0z deleted the exllamav2_lora branch October 16, 2023 12:15
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants