You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Each LLM provider and model has limits on the input size of prompts. When a user asks questions about indexed docs with /ask, the chunk size used to split docs in indexing affects the amount of context passed to the LLM in the prompt. Especially for smaller models that are running locally, an augmented prompt can end up exceeding the prompt size limit.
Proposed Solution
Add token limit attribute to provider class and implement this for different providers/models.
Use the token limit to derive the chunk size, overlap or a different chain that optimizes the indexing for that particular model.
The text was updated successfully, but these errors were encountered:
JasonWeill
changed the title
Add token limit attribute to providers, derive chuck size for indexing docs from this
Add token limit attribute to providers, derive chunk size for indexing docs from this
Jul 27, 2023
Problem
Each LLM provider and model has limits on the input size of prompts. When a user asks questions about indexed docs with
/ask
, the chunk size used to split docs in indexing affects the amount of context passed to the LLM in the prompt. Especially for smaller models that are running locally, an augmented prompt can end up exceeding the prompt size limit.Proposed Solution
The text was updated successfully, but these errors were encountered: