-
Notifications
You must be signed in to change notification settings - Fork 10.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add support for InternLM 2.5 1M context. Should be as good as command r+ #8285
Comments
@mirek190 hi, can you share an example of outputs? |
Is just terrible
I think gguf is not correct and llamacpp maybe as well . Such questions are easy for llama 3 or gemma 2. |
@mirek190 it seems not so good. Can you share the command you use to convert to |
It uses that template
So my command looks like that:
|
Looking on open-llm-leaderboard it has level of command r+ or Qwen1.5-110B so implementation under llamacpp is broken. |
Here are more (reddit) complaints, just FYI Bug exists 100% |
@mirek190 Just a heads up, your command has prefix and suffix defined twice and there's a typo in the system prompt: You don't have to do all of that anymore with Repeat temperature should be Are you using the latest version? |
Withe the newest build is even worse and corrected command
loop and answer has no sense at all
So bad ....
just bad bad
still soo bad ... even llama 2 was better ;) |
Is this related to the model itself or the inference engine? |
@mirek190 hi, something wrong with converting. Could try again and use this pr #8321 to convert to gguf. |
I have tested the apple problem with chatllm.cpp (Q8_0, greedy sampling). InternLM 2.5 1M
2.5
|
I have also tried LlaMA-3 8B, Gemma-2 9B, Qwen2 7B, Phi-3 medium 128k, Phi-3 mini Jun update. Both LlaMA-3 8B and Gemma-2 9B say there are 36 apples. As far as I can tell, this is correct. Qwen2 7B and Phi-3's results are wrong. So, I don't think InternLM 2.5 can answer this apple problem correctly. But the generated text looks worse than mine. There might be somethings wrong on RoPE scaling. |
Clarification about Phi-3-medium: the 4k one can solve it while the 128k one can't. (greedy sampling) |
This issue was closed because it has been inactive for 14 days since being marked as stale. |
Prerequisites
Feature Description
Can you add support for InternLM 2.5 1M because is working but responses are terrible.
https://huggingface.co/internlm/internlm2_5-7b-chat-1m
Motivation
to work properly
Possible Implementation
No response
The text was updated successfully, but these errors were encountered: