Incorrect Backend Selected #569
Replies: 4 comments 1 reply
-
In order to specify a backend you have to set The available backends are listed here https://localai.io/model-compatibility/index.html |
Beta Was this translation helpful? Give feedback.
-
Thanks. It would be nice to have updated documentation with examples for the models that are not covered in the existing documentation. |
Beta Was this translation helpful? Give feedback.
-
I think the documents at https://localai.io/advanced/index.html will help |
Beta Was this translation helpful? Give feedback.
-
I found the documentation needed. |
Beta Was this translation helpful? Give feedback.
-
When I select a model it does not seem to pickup the correct backend in this case.
How do I get t to associate the model with the correct backend?
In this case I am trying to use a model that is for use with starcoder.
Starting LocalAI using 4 threads, with models path: /home/andrew/Downloads/LocalAI/models
┌───────────────────────────────────────────────────┐
│ Fiber v2.46.0 │
│ http://127.0.0.1:8080 │
│ (bound on host 0.0.0.0 and port 8080) │
│ │
│ Handlers ............ 23 Processes ........... 1 │
│ Prefork ....... Disabled PID ............. 27409 │
└───────────────────────────────────────────────────┘
llama.cpp: loading model from /home/andrew/Downloads/LocalAI/models/gpt_bigcode-santacoder-ggml-q4_1.bin
error loading model: missing tok_embeddings.weight
llama_init_from_file: failed to load model
gptj_model_load: loading model from '/home/andrew/Downloads/LocalAI/models/gpt_bigcode-santacoder-ggml-q4_1.bin' - please wait ...
gptj_model_load: n_vocab = 49280
gptj_model_load: n_ctx = 2048
gptj_model_load: n_embd = 2048
gptj_model_load: n_head = 16
gptj_model_load: n_layer = 24
gptj_model_load: n_rot = 2003
gptj_model_load: f16 = 49280
gptj_model_load: invalid model file '/home/andrew/Downloads/LocalAI/models/gpt_bigcode-santacoder-ggml-q4_1.bin' (bad vocab size 1 != 49280)
GPT-J ERROR: failed to load model from /home/andrew/Downloads/LocalAI/models/gpt_bigcode-santacoder-ggml-q4_1.bingpt_neox_model_load: loading model from '/home/andrew/Downloads/LocalAI/models/gpt_bigcode-santacoder-ggml-q4_1.bin' - please wait ...
gpt_neox_model_load: n_vocab = 49280
gpt_neox_model_load: n_ctx = 2048
gpt_neox_model_load: n_embd = 2048
gpt_neox_model_load: n_head = 16
gpt_neox_model_load: n_layer = 24
gpt_neox_model_load: n_rot = 2003
gpt_neox_model_load: par_res = 49280
gpt_neox_model_load: ftype = 1
gpt_neox_model_load: qntvr = 0
Beta Was this translation helpful? Give feedback.
All reactions