-
Notifications
You must be signed in to change notification settings - Fork 28.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Chat] Add Chat from TRL 🐈 #35714
[Chat] Add Chat from TRL 🐈 #35714
Conversation
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Super nice! Just missing documentation IMO and maybe some examples of actually calling the CLI! 🤗
@ArthurZucker I've added docs, but the preview isn't showing them 🤔 I'm merging and then I'll check the docs on |
* tmp commit * add working chat * add docts * docs 2 * use auto dtype by default
* tmp commit * add working chat * add docts * docs 2 * use auto dtype by default
What does this PR do?
Adds a
chat
command, allowing users to explore text models with a minimal effort. Except for the transformers-specific command boilerplate, this chat is a copy of the chat in TRL.The purpose of this PR is to establish a baseline from something that already works well. In subsequent PRs, a few QoL upgrades will be added to this chat, including:
transformers
GenerationConfig
to be setgeneration_config.json
file to parameterize the chat, and allow users to export current chat settings into fileslocal-gemma
, including compilationUsage
transformers-cli chat --model_name_or_path Qwen/Qwen2.5-0.5B-Instruct
transformers-cli chat --help
help