Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

customise number of experts in mixtral #1553

Open
scienlabs opened this issue Dec 15, 2023 · 4 comments
Open

customise number of experts in mixtral #1553

scienlabs opened this issue Dec 15, 2023 · 4 comments

Comments

@scienlabs
Copy link

Could you someone provide guidance or documentation on how to adjust the number of experts in mixtral? I'm particularly interested in understanding if there's a way to dynamically adjust this number based on the requirements of different tasks or scenarios.

@scienlabs scienlabs changed the title cutomise number of experts in mixtral customise number of experts in mixtral Dec 15, 2023
@RafaAguilar
Copy link

I'm not sure what Ollama uses, but for the llama.cpp backend you can override a Key in the model with:

--override-kv KEY=TYPE:VALUE
                        advanced option to override model metadata by key. may be specified multiple times.
                        types: int, float, bool. example: --override-kv tokenizer.ggml.add_bos_token=bool:false

For example I override them using:

--override-kv llama.expert_used_count=int:3

But I think this is not yet supported by MODELFILE.

@scienlabs
Copy link
Author

how can i do it with ollama? wondering if anyone can help

@PLK2
Copy link

PLK2 commented May 8, 2024

Figured it out yet?

@ColumbusAI
Copy link

Any update on this?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants