Skip to content

This issue was moved to a discussion.

You can continue the conversation there. Go to discussion →

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug]: "Continue generation" request to API is missing a parameter #5654

Closed
1 task done
Originalimoc opened this issue Feb 5, 2025 · 0 comments
Closed
1 task done
Labels
🐛 bug Something isn't working

Comments

@Originalimoc
Copy link

Originalimoc commented Feb 5, 2025

When you click stop generation, there will be an additional continue generation button which won't appear if you let the generation finish itself. That button should trigger a generation request contains the specific parameter in another project's issue closing comment or the HF endpoint will assume it's a new round and model will not behave as what continue generation intended. It's called "Prefilled" generation request. Another use case of this feature is model steering. To prevent this, add add_generation_prompt to false in the request to prevent adding an extra turn:

Relavant:
"https://docs.vllm.ai/en/latest/serving/openai_compatible_server.html?ref=blog.mozilla.ai#:~:text=role.%22)%2C%0A%20%20%20%20)-,add_generation_prompt,-%3A%20bool%20%3D"
&
"huggingface/transformers#33198"

Code of Conduct

  • I agree to follow this project's Code of Conduct
@Originalimoc Originalimoc added the 🐛 bug Something isn't working label Feb 5, 2025
Repository owner locked and limited conversation to collaborators Feb 5, 2025
@danny-avila danny-avila converted this issue into discussion #5659 Feb 5, 2025

This issue was moved to a discussion.

You can continue the conversation there. Go to discussion →

Labels
🐛 bug Something isn't working
Projects
None yet
Development

No branches or pull requests

1 participant