You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
[The documentation](https://huggingface.co/docs/optimum/habana/usage_guides/deepspeed) provides more information about how to use DeepSpeed within Optimum Habana.
If training on a different language, you should be sure to change the `language` argument. The `language` and `task` arguments should be omitted for English speech recognition.
You can look at the [documentation](https://huggingface.co/docs/optimum/habana/usage_guides/deepspeed) for more information about how to use DeepSpeed in Optimum Habana.
Copy file name to clipboardexpand all lines: examples/text-generation/README.md
+22-11
Original file line number
Diff line number
Diff line change
@@ -79,7 +79,8 @@ python run_generation.py \
79
79
--use_kv_cache \
80
80
--max_new_tokens 100 \
81
81
--do_sample \
82
-
--prompt "Here is my prompt"
82
+
--prompt "Here is my prompt" \
83
+
--sdp_on_bf16
83
84
```
84
85
85
86
If you want to provide several prompts as inputs, here is how to do it:
@@ -91,7 +92,8 @@ python run_generation.py \
91
92
--max_new_tokens 100 \
92
93
--do_sample \
93
94
--batch_size 2 \
94
-
--prompt "Hello world" "How are you?"
95
+
--prompt "Hello world" "How are you?" \
96
+
--sdp_on_bf16
95
97
```
96
98
97
99
> The batch size should be larger than or equal to the number of prompts. Otherwise, only the first N prompts are kept with N being equal to the batch size.
0 commit comments