You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi @ggerganov , I saw the 'good first issue' tag on this. I'm new to this project and thought to give this a try. I manage to build the llama-server, but I'm getting this strange error when get the server up and running.
I'm following the README here : https://github.com/ggerganov/llama.cpp/tree/master/examples/server and getting the following error. llama_model_load: error loading model: missing tensor 'token_embd.weight' llama_load_model_from_file: failed to load model common_init_from_params: failed to load model '.\models\ggml-vocab-deepseek-llm.gguf' srv load_model: failed to load model, '.\models\ggml-vocab-deepseek-llm.gguf' main: exiting due to model loading error
Any pointers on how I should proceed? Building with MSVC 19.42.34435.0 x64 Windows machine.
Name and Version
version: 4291 (ce8784b)
built with MSVC 19.29.30157.0 for x64
Operating systems
Windows
Which llama.cpp modules do you know to be affected?
llama-server
Problem description & steps to reproduce
After #10722, there are still some issues when using the native
/completion
endpoint:has_new_line
property is set to1
instead oftrue
like it used to be.model
property is set togpt-3.5-turbo-0613
even though this isn't the OpenAI-compatible API.First Bad Commit
No response
Relevant log output
No response
The text was updated successfully, but these errors were encountered: