Replies: 1 comment 7 replies
-
@SeanMcTex yeah - can currently use ollama for that (which is really openai API to it - and it works ok as long as can run a suitable model). Can gpt4all run large enough models to be worthwhile? (you need fairly hefty ones with some local GPU support to be worth it for goose from my experience). I have used llamafile and llama.cpp as well (antthing that offers a openai like http interface, and has to have tool calling). Would ollama be interesting? https://github.com/square/exchange/blob/main/src/exchange/providers/ollama.py |
Beta Was this translation helpful? Give feedback.
7 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
gpt4all makes it pretty easy to download and run LLMs on your own machine. It also has an OpenAI-compatible API.
It would be great to be able to point Goose to a local LLM this way, thereby saving costs and mitigating concerns about what's getting sent to one's LLM provider.
Beta Was this translation helpful? Give feedback.
All reactions