Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

offline chat generates garbage output #516

Closed
mtoniott opened this issue Oct 25, 2023 · 3 comments
Closed

offline chat generates garbage output #516

mtoniott opened this issue Oct 25, 2023 · 3 comments

Comments

@mtoniott
Copy link

Hello,

Thank you for this software getting better everyday.

So recently I updated khoj and the offline chat that was giving me normal answers before is now outputing garbage with Mississipi in it each time for some reasons.
image

I tried reinstalling khoj in a new venv. Did not work.
I tried turning off the offline chat, removing the model in my .cache file then redownload it. Same result.

I guess it is linked to the fact that it wants to use my intel integrated graphics to accelerate the queries. But I did not find a way to turn it off. I get the following line in the terminal;

llama.cpp: using Vulkan on Intel(R) Iris(R) Plus Graphics 655 (CFL GT3)

Any idea on how to fix this?

@debanjum
Copy link
Member

debanjum commented Oct 25, 2023

Ah that's unfortunate that you're seeing a regression in behavior. The chat
model being loaded into your GPU could be the reason for this regression
for sure. Let me look into testing this out on my end.

Can you share details of your machine specs. Specifically the RAM, Processor and GPU on
your machine?

Details

We'd started using an upgraded default model for offline chat (Mistral
instead of Llama 2) and try to use GPU when a (Vulcan) supported GPU is
available on the users machine. Using Intel or AMD GPUs hasn't been tested
as we dont have such machines ourselves.

@debanjum
Copy link
Member

debanjum commented Oct 26, 2023

Vulkan support in our upstream dependency (GPT4All) still needing some ironing out. Until then I've exposed a CLI flag to allow users to disable using GPU for offline chat.

To use this fix:

  1. Upgrade to the latest pre-release version of Khoj: pip install --upgrade --pre khoj-assistant
  2. Start Khoj server with --disable-chat-on-gpu flag: khoj --disable-chat-on-gpu

@mtoniott: Let me know if this mitigates the issue with offline chat generating gibberish output?

@mtoniott
Copy link
Author

It works now. Ty !

image

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants