Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Obs AI Assistant] Use preconfigured elser inference endpoint #200908

Open
sorenlouv opened this issue Nov 20, 2024 · 2 comments
Open

[Obs AI Assistant] Use preconfigured elser inference endpoint #200908

sorenlouv opened this issue Nov 20, 2024 · 2 comments
Assignees
Labels
Team:Obs AI Assistant Observability AI Assistant

Comments

@sorenlouv
Copy link
Member

Problem
The Observability AI Assistant currently has a setup step for creating an inference endpoint obs_ai_assistant_kb_inference. When created this will download .elser_model_2. This is required to use the knowledge base and adds complexity to the codebase.

Solution
8.17 ships with pre-configured inference endpoints for elser and e5 models (elastic/elasticsearch#116931). We should replace our own endpoint obs_ai_assistant_kb_inference with .elser-2-elasticsearch:

Current setup UI

This screenshot shows the setup UI that we can get rid of if we use the preconfigured inference endpoint

@botelastic botelastic bot added the needs-team Issues missing a team label label Nov 20, 2024
@sorenlouv sorenlouv added the Team:Obs AI Assistant Observability AI Assistant label Nov 20, 2024
@botelastic botelastic bot removed the needs-team Issues missing a team label label Nov 20, 2024
@elasticmachine
Copy link
Contributor

Pinging @elastic/obs-ai-assistant (Team:Obs AI Assistant)

@sorenlouv sorenlouv changed the title [Obs AI Assistant] Use built-in inference endpoint [Obs AI Assistant] Use preconfigured elser inference endpoint Nov 21, 2024
@sorenlouv
Copy link
Member Author

This has been postponed because the preconfigured endpoints don't allow us to adjust min_number_of_allocations. As soon as that is possible or we have a multi-tenant inference service we can pick this up again. In the meantime we'll be using custom inference endpoints.

@sorenlouv sorenlouv removed the v8.17.0 label Nov 29, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Team:Obs AI Assistant Observability AI Assistant
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants