Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
All HF llama model are falcon style ROPE and we can convert them to original llama style ROPE with a permutation.
This pull request solve the bug when converting HF GQA to gguf format.
I learned idea from it and fix the similar bug in the llama2.c's exports.py.
Now I successfully convert Tinyllama-1.1B-chat to llama style ROPE. So, we can remove the falcon ROPE part.
I have upload the new export.py and llama2.mojo.
Details:
python export.py tl-chat.bin --hf PY007/TinyLlama-1.1B-Chat-v0.2 --version 0
to conver the model