Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Do not timeout when calling HF through acomplete #885

Merged
merged 1 commit into from
Nov 23, 2023

Conversation

okotek
Copy link
Contributor

@okotek okotek commented Nov 23, 2023

I'm trying to use litellm for a HF text-generation-inference server. When using completion everything works as expected. When using acompletion, a timeout exception occurs after 5 seconds.

This PR changes the default timeout of httpx to None.

Copy link

vercel bot commented Nov 23, 2023

The latest updates on your projects. Learn more about Vercel for Git ↗︎

Name Status Preview Comments Updated (UTC)
litellm ✅ Ready (Inspect) Visit Preview 💬 Add feedback Nov 23, 2023 2:02pm

@krrishdholakia
Copy link
Contributor

lgtm! thanks for the fix @okotek

@krrishdholakia krrishdholakia merged commit 6ba4eeb into BerriAI:main Nov 23, 2023
@krrishdholakia
Copy link
Contributor

@okotek curious - what're you using async hf completions for?

context - trying to improve our own implementation, and want to make sure we're good for your scenario.

@okotek
Copy link
Contributor Author

okotek commented Nov 23, 2023

Over here https://github.com/Codium-ai/pr-agent/blob/main/pr_agent/algo/ai_handler.py
It's your implementation :)

I've been running HF TGI on a GPU server and tried to connect to it.

@krrishdholakia
Copy link
Contributor

got it - 😄

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants