-
Notifications
You must be signed in to change notification settings - Fork 24
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Ollama] certain models are not loaded with correct n_ctx #58
Comments
Aider can config context. but this plugin doesn't give this to user. so I think this may cause this issue. does this influence on your result? |
Yes I think this is the cause cause I just noticed I had the same issue with aider alone.
This correctly results in all my models being loaded with a 16384 token window. So yeah, we should have a way to set it up in aider-composer too to have better control. |
And thanks for such a quick response! |
this is a good advise. but currently I am working on other features. and how to design the settings may be a big problem. |
Are you foreseeing problems because of the fact there is 2 separated models? |
thanks, since this may not be urgent, It has more time. maybe others have a better solution. |
Maybe Ive been a bit hasty closing this, I'll let this up to you |
@lee88688 btw are you reachable somewhere through PM? Ive a little something I wanna share related to your project |
@blakkd you can contact me with my email in my profile. |
Sorry mate didn't think about the email way lol
I was just wondering if you saw that Aider now have a native IDE integration?
I'm you asking this because I just stumbled on it yesterday!
It appears to me like a new feature but I didn't check how long it's been implemented for.
So I was thinking maybe you were already aware of this but had further goals and deliberately decided to keep your project going, but afterward worried you weren't and in this case it I thought was fair to inform you just in case.
That said, I've haven't been trying your project for more than 5min and so couldn't really compare the feature of your project to their in term of functionalities.
I'm only reaching you in good will. I just felt like a duty to relay the info.I hope you already knew it not to bring a bad news!
All the best!
…On Thursday, January 9th, 2025 at 12:10 PM, Jimmie Lee ***@***.***> wrote:
***@***.***(https://github.com/blakkd) you can contact me with my email in my profile.
—
Reply to this email directly, [view it on GitHub](#58 (comment)), or [unsubscribe](https://github.com/notifications/unsubscribe-auth/BJEKR2YN2327M64MTVE5W632JZKMBAVCNFSM6AAAAABUUJIEZ6VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDKNZZHA3DQNRTG4).
You are receiving this because you were mentioned.Message ID: ***@***.***>
|
I have 2 daily driver models: qwen2.5-coder:32b-instruct-q4_K_M and qwq:32b-preview-q4_K_M.
Both are used in other applications.
The issue is that the Ollama num_ctx parameter is not always respected.
I don't know what triggers this behavior.
Loads correctly in 16384 (ollama serve logs)
BUT
Loads in 8192 instead (ollama serve logs)
The text was updated successfully, but these errors were encountered: