From 22a5e9827b1e56d24803c9d0b4ed38b650bc664a Mon Sep 17 00:00:00 2001 From: Luca-Calabria Date: Wed, 4 Dec 2024 12:25:13 +0100 Subject: [PATCH 1/3] added gemma_2b_it deepspeed baseline --- tests/baselines/gemma_2b_it.json | 12 ++++++++++++ 1 file changed, 12 insertions(+) diff --git a/tests/baselines/gemma_2b_it.json b/tests/baselines/gemma_2b_it.json index f103c2e7ac..eb92692ed2 100644 --- a/tests/baselines/gemma_2b_it.json +++ b/tests/baselines/gemma_2b_it.json @@ -25,6 +25,18 @@ "--dataset_config_name wikitext-2-raw-v1", "--use_hpu_graphs_for_inference" ] + }, + "deepspeed": { + "learning_rate": 8e-4, + "train_batch_size": 4, + "perplexity": 924.062, + "train_runtime": 75.518, + "train_samples_per_second": 81.097, + "extra_arguments": [ + "--dataset_config_name wikitext-2-raw-v1", + "--use_hpu_graphs_for_inference", + "--deepspeed tests/configs/deepspeed_zero_2.json" + ] } } } From f85332365f033b492ced03795e15b0c54638cbe8 Mon Sep 17 00:00:00 2001 From: Luca-Calabria Date: Fri, 6 Dec 2024 11:20:42 +0100 Subject: [PATCH 2/3] removed gemma from G1 CI --- tests/test_examples.py | 10 ++++++---- 1 file changed, 6 insertions(+), 4 deletions(-) diff --git a/tests/test_examples.py b/tests/test_examples.py index b6f07b0512..73cf3c59c4 100644 --- a/tests/test_examples.py +++ b/tests/test_examples.py @@ -288,6 +288,8 @@ def to_test( return False elif eager_mode and model_name not in models_measured_on_eager_mode: return False + elif "gemma" in model_name and not IS_GAUDI2: + return False elif model_name not in models_with_specific_rules and not deepspeed: return True elif model_name == "gpt2-xl" and deepspeed: @@ -324,8 +326,6 @@ def to_test( return True elif "huggyllama" in model_name and IS_GAUDI2 and deepspeed: return True - elif "gemma" in model_name and IS_GAUDI2: - return True return False @@ -509,7 +509,9 @@ def test(self): if os.environ.get("DATA_CACHE", None) is not None and self.EXAMPLE_NAME == "run_clip": extra_command_line_arguments[0] = "--data_dir {}".format(os.environ["DATA_CACHE"]) elif torch_compile and ( - model_name == "bert-large-uncased-whole-word-masking" or model_name == "roberta-large" + model_name == "bert-large-uncased-whole-word-masking" + or model_name == "roberta-large" + or model_name == "albert-xxlarge-v1" ): extra_command_line_arguments.append("--torch_compile_backend hpu_backend") extra_command_line_arguments.append("--torch_compile") @@ -857,7 +859,7 @@ class ProteinFoldingExampleTester2(ExampleTesterBase, metaclass=ExampleTestMeta, class CausalLanguageModelingLORAExampleTester( ExampleTesterBase, metaclass=ExampleTestMeta, example_name="run_lora_clm" ): - TASK_NAME = ["tatsu-lab/alpaca", "databricks/databricks-dolly-15k"] + TASK_NAME = "databricks/databricks-dolly-15k" class MultiCardCausalLanguageModelingLORAExampleTester2( From 1158a5c30bdc443e669aeab4795da920e6b6ca22 Mon Sep 17 00:00:00 2001 From: regisss <15324346+regisss@users.noreply.github.com> Date: Fri, 6 Dec 2024 16:59:41 +0000 Subject: [PATCH 3/3] Fix CI rules --- tests/test_examples.py | 2 ++ 1 file changed, 2 insertions(+) diff --git a/tests/test_examples.py b/tests/test_examples.py index 73cf3c59c4..85d49fe6dd 100644 --- a/tests/test_examples.py +++ b/tests/test_examples.py @@ -326,6 +326,8 @@ def to_test( return True elif "huggyllama" in model_name and IS_GAUDI2 and deepspeed: return True + elif "gemma" in model_name and IS_GAUDI2: + return True return False