You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
am fallowing the intructions to install but i hade to chnage the numpy version to 1.19.0 to work and also i install vicuna.bin to run my model since there is no bulided model with the repo so when i download the vicuna.bin from fastChat repo and create 7B folder in the models folder and i put the ggml-vocab.bin there i run this command from the intructions: ./main -m models/7B/ggml-vocab.bin -p "the sky is" i get this:
command: ./main -m models/7B/ggml-vocab.bin -p "the sky is"
error: main: build = 526 (e6a46b0) main: seed = 1683697939 llama.cpp: loading model from models/7B/ggml-vocab.bin error loading model: missing tok_embeddings.weight llama_init_from_file: failed to load model llama_init_from_gpt_params: error: failed to load model 'models/7B/ggml-vocab.bin' main: error: unable to load mode
can i skip this part and how can i go forward if there is not model bulided in with this repo?
plz help if you can also if i try to host the backend as a API how is that possible since am just using localhost::8080 as a backend endpoint
The text was updated successfully, but these errors were encountered:
you need an actual model – ggml-vocab.bin isn't a model. you need to download it online u can find plenty on hugging face, just make sure it's labelled ggml
you need an actual model – ggml-vocab.bin isn't a model. you need to download it online u can find plenty on hugging face, just make sure it's labelled ggml
am getting error npm:
npm install ─╯
[email protected] postinstall
npm run updateengines && cd InferenceEngine/embeddings/all-mpnet-base-v2 && python -m pip install -r requirements.txt
am fallowing the intructions to install but i hade to chnage the numpy version to 1.19.0 to work and also i install vicuna.bin to run my model since there is no bulided model with the repo so when i download the vicuna.bin from fastChat repo and create 7B folder in the models folder and i put the ggml-vocab.bin there i run this command from the intructions:
./main -m models/7B/ggml-vocab.bin -p "the sky is"
i get this:command:
./main -m models/7B/ggml-vocab.bin -p "the sky is"
error:
main: build = 526 (e6a46b0) main: seed = 1683697939 llama.cpp: loading model from models/7B/ggml-vocab.bin error loading model: missing tok_embeddings.weight llama_init_from_file: failed to load model llama_init_from_gpt_params: error: failed to load model 'models/7B/ggml-vocab.bin' main: error: unable to load mode
can i skip this part and how can i go forward if there is not model bulided in with this repo?
plz help if you can also if i try to host the backend as a API how is that possible since am just using localhost::8080 as a backend endpoint
The text was updated successfully, but these errors were encountered: