Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Are Qwen1.5 MOE models supported? #6415

Closed
l3utterfly opened this issue Apr 1, 2024 · 6 comments
Closed

Are Qwen1.5 MOE models supported? #6415

l3utterfly opened this issue Apr 1, 2024 · 6 comments

Comments

@l3utterfly
Copy link
Contributor

I tried to convert the Qwen 1.5 MOE: https://huggingface.co/Qwen/Qwen1.5-MoE-A2.7B

It gives me an error message:

 python convert-hf-to-gguf.py /home/layla/src/text-generation-webui/models/Qwen1.5-MoE-A2.7B
Loading model: Qwen1.5-MoE-A2.7B
Traceback (most recent call last):
  File "/home/layla/src/llama.cpp/convert-hf-to-gguf.py", line 2296, in <module>
    main()
  File "/home/layla/src/llama.cpp/convert-hf-to-gguf.py", line 2276, in main
    model_class = Model.from_model_architecture(hparams["architectures"][0])
                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/layla/src/llama.cpp/convert-hf-to-gguf.py", line 215, in from_model_architecture
    raise NotImplementedError(f'Architecture {arch!r} not supported!') from None
NotImplementedError: Architecture 'Qwen2MoeForCausalLM' not supported!

I see others have uploaded GGUFs of the Qwen MOE models on HF, so it leads me to think I'm doing something wrong

@l3utterfly
Copy link
Contributor Author

I see this issue: huggingface/transformers#29377

Which adds support to Qwen2, do I need that PR to convert Qwen1.5 moe?

@l3utterfly l3utterfly reopened this Apr 1, 2024
@l3utterfly l3utterfly changed the title Are Qwen MOE models supported? Are Qwen1.5 MOE models supported? Apr 1, 2024
@gswsqffsapd3056
Copy link

Same needs, waiting to be solved

@Jeximo
Copy link
Contributor

Jeximo commented Apr 1, 2024

I see there's only 2 people making Qwen MOE 1.5 GGUFs on Huggingface(not sure how). llama.cpp shows support for Qwen, but doesn't specify Qwen Moe(yet).

Here's a PR discussing converting/running it, though it appears there's more work to be done before it'll run.

@Lyzin
Copy link

Lyzin commented Apr 3, 2024

I don't know when qwen1.5-MoE will support llama.cpp, I can only wait for qwen official update!

@maziyarpanahi
Copy link

Speaking of MoE and Qwen, #6453

@github-actions github-actions bot added the stale label May 4, 2024
Copy link
Contributor

This issue was closed because it has been inactive for 14 days since being marked as stale.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

5 participants