-
Notifications
You must be signed in to change notification settings - Fork 3.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Failed to load llama model #702
Comments
I am facing the same issue. I will experiment replacing the llama files here with ones from the latest llama repo to see if that makes a difference. |
The reason I believe is due to the ggml format has changed in llama.cpp, see ggerganov/llama.cpp#613. The changes have not back ported to whisper.cpp yet. So to use talk-llama, after you have replaced the |
Did it work? I've redownloaded the model, but the issue still persists. |
you are correct. i used all the above source code to update the whisper version and re compile everything. it worked. |
If someone is trying to just get this working I managed to do it. So don't do that and use convert script from #324 to convert ggml-model-q4_0.bin (not ...q4_1.bin as it is newer ggml format) and it worked for me. Note that I am trying to use alpaca model here (as linked) (Seems like 30B variant fails with: |
This should work now and the performance is much better compared to what we had before. |
Hi,
I followed the instruction here to create this ggml-model-q4_0.bin file.
https://github.com/ggerganov/llama.cpp
Then try to run the talk-llama with following command:
./talk-llama -mw ./models/ggml-model-whisper-base.en.bin -ml ./models/ggml-model-q4_0.bin -p “Myname” -t 8
why got the following feedback:
Obviously the whisper model launched successfully. but the llama model didn't. Not sure what I did wrong.
I am 100% sure the model file folder path is correct.
The text was updated successfully, but these errors were encountered: