Porting LeoLM instruct models to llama.cpp #3935
Unanswered
sorgfresser
asked this question in
Q&A
Replies: 1 comment 2 replies
-
I think what you might need to do is add pad tokens rather than trying to change the model's vocab size. You can try #3743 with the |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hey,
I'm very impressed by the speed and ease at which llama.cpp can deploy many models. I tried converting a German & English only model named LeoLM but did only manage to get it to work for the non-instruct finetuned variants which seems a bit odd to me.
First of all, if I just try to convert the
LeoLM/leo-hessianai-7b-chat
(available on hf) I getThe vocab size of the instruction ones does actually exceed
32000
(there are some special tokens above id31999
in the range[32000, 32006]
) but by simply modifying the config.json to vocab_size: 32000 similar to #3900 I at least managed to get the conversion itself to run through.But if I run
./main -m "models/leo-hessianai-7b-chat/ggml-model-f16.gguf"
it fails withwhich is fair since I edited the vocab size from
32128
to32000
but confuses me since I initially got the impression that the tokenizer only had 32000 tokens due to the error.Is there any param I could pass to convert.py that I'm overlooking? All the ones I get using -h do not seem helpful to me.
Thanks a lot for any advice!
Beta Was this translation helpful? Give feedback.
All reactions