Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Custom inline completion providers #18490

Open
zerocorebeta opened this issue Sep 29, 2024 · 5 comments
Open

Custom inline completion providers #18490

zerocorebeta opened this issue Sep 29, 2024 · 5 comments
Labels
ai Improvement related to Assistant, Copilot, or other AI features enhancement [core label] inline completion Umbrella label for Copilot, Supermaven, etc. completions

Comments

@zerocorebeta
Copy link

zerocorebeta commented Sep 29, 2024

Summary: Custom inline completion providers for local models or other platforms

--

After going through: https://zed.dev/docs/completions

Zed currently supports completions via external LLM APIs like GitHub Copilot and Supermaven, but this is restrictive. Many users, for privacy or performance reasons, might prefer alternatives like Gemini Flash or local models via Ollama.

There are several advanced LLMs that support the Fill-in-the-Middle (FIM) objective, such as CodeGemma. Additionally, platforms like Continue.dev allow code completion via local models, with Startcoder via Ollama as the default.

Expanding LLM support to include more flexible, local, or privacy-focused options would greatly enhance Zed's appeal and utility for a wider range of developers.

@JosephTLyons JosephTLyons added ai Improvement related to Assistant, Copilot, or other AI features inline completion Umbrella label for Copilot, Supermaven, etc. completions and removed triage labels Oct 12, 2024
@ggerganov
Copy link

We have recently extended the llama.cpp server with a specialized /infill endpoint that enables FIM requests with large contexts to run efficiently in local environments. A simple example of using this endpoint with Qwen2.5-Coder in Neovim is demonstrated here: ggerganov/llama.cpp#9787

I believe it could be an interesting option to explore in the scope of this issue. Feel free to ping me if you have any questions.

@20manas
Copy link

20manas commented Oct 23, 2024

I think Zed AI should also provide its own code completion functionality.

@bersace
Copy link

bersace commented Dec 16, 2024

Does Zed allows to use Codestral cloud for FIM ?

@josharian
Copy link

https://arstechnica.com/ai/2025/01/nvidias-first-desktop-pc-can-run-local-ai-models-for-3000/ is relevant here. FIM models aren't all that big; one of these could probably handle an entire office's worth of requests.

I'd really like to be able to use a llama.cpp model for FIM.

@aretrace
Copy link

Configurable FIM support should be a priority. While the improvements to the Assistant panel over time have been great, the lack of customizable AI autocomplete in Zed is becoming a significant drawback compared to the code assistance experiences offered by others, such as IntelliSense.

@agu-z agu-z changed the title Expand AI Code Completion beyond Copilot and Supermaven Custom inline completion providers Jan 30, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ai Improvement related to Assistant, Copilot, or other AI features enhancement [core label] inline completion Umbrella label for Copilot, Supermaven, etc. completions
Projects
None yet
Development

No branches or pull requests

7 participants