From 31b02d0cd0a23e77d4554a2a621e7b818e216723 Mon Sep 17 00:00:00 2001 From: Mark Sturdevant Date: Wed, 14 Feb 2024 19:38:17 -0800 Subject: [PATCH] Update README.md to clarify model requirement (#1315) Clarify that language models must be transformers models for text. This is a bit redundant with intro description, but attempts to better address a question that that comes up (issue 1257). Closes: #1257 --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 25e501769d..6425f431ab 100644 --- a/README.md +++ b/README.md @@ -168,7 +168,7 @@ train_stats = ppo_trainer.step([query_tensor[0]], [response_tensor[0]], reward) The PPO implementation largely follows the structure introduced in the paper **"Fine-Tuning Language Models from Human Preferences"** by D. Ziegler et al. \[[paper](https://arxiv.org/pdf/1909.08593.pdf), [code](https://github.com/openai/lm-human-preferences)]. ### Language models -The language models utilize the `transformers` library by 🤗 Hugging Face. +The language models utilize the `transformers` library by 🤗 Hugging Face. Currently, `trl` only supports `transformers` models **for text**. ## Citation