-
-
Notifications
You must be signed in to change notification settings - Fork 5.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
error loading model: create_tensor: tensor 'blk.0.ffn_gate.weight' not found #4881
Comments
Does llama.cpp even support mixtral yet? |
llama.cpp has Mixtral support in the works but it's not part of the master branch yet. You need to wait for it to be merged into the master branch and for llama.cpp python bindings to get updated before it can be added to ooba. |
came to say I get the same error when trying to work with llama cpp directly
I find it funny that https://huggingface.co/TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF says mixtral works with llama.cpp august 27 onward's, but I guess that's a generic message that applies only to gguf |
today everyone is saying it merged, but all Ooba users finding this thread should know Ooba hasn't merged it yet to the script update, you need to manually refresh the llamma.cpp install |
Can you tell me how I do that? |
Finally managed to run it on windows, got to install https://developer.nvidia.com/cuda-12-1-0-download-archive (getting the correct version is important) with the other dependencies, install the webui manually with conda, (without the one click installer), then just follow this https://old.reddit.com/r/Oobabooga/comments/18gijyx/simple_tutorial_using_mixtral_8x7b_gguf_in_ooba/ |
Je fais marcher mes IA sur windows 11 avec les CPU uniquement, donc cuda ne me concerne pas, mais je vais quand même regarder ce lien et tenter les mises a jour. Au moins on sait que le fichier gguf n'est pas en cause. |
I don't think these workarounds are require any more. According to abetlen/llama-cpp-python#1000, it's available in v0.2.23 this python library. I needed to fix moving of my virtual environment, so I wiped the entire environment and rebuilt from scratch (which I'm using pip, so it was just removing |
This issue has been closed due to inactivity for 6 weeks. If you believe it is still relevant, please leave a comment below. You can tag a developer in your comment. |
Describe the bug
I get an error when running mixtral-8x7b via text-generation-web ui. This is the mistake:
Is there an existing issue for this?
Reproduction
I used this google colab - link
it performs the following steps:
Screenshot
Logs
System Info
The text was updated successfully, but these errors were encountered: