From a240462409b08cffc55cab26abdb87fd62e79e68 Mon Sep 17 00:00:00 2001 From: Xin Yang Date: Fri, 15 Nov 2024 14:22:23 -0800 Subject: [PATCH] update --- tests/integration/llm/client.py | 6 +++--- tests/integration/llm/prepare.py | 4 ++-- 2 files changed, 5 insertions(+), 5 deletions(-) diff --git a/tests/integration/llm/client.py b/tests/integration/llm/client.py index c4131986ef..59dd63d098 100644 --- a/tests/integration/llm/client.py +++ b/tests/integration/llm/client.py @@ -389,14 +389,14 @@ def get_model_name(): }, "gemma-7b-unmerged-lora": { "batch_size": [4], - "seq_length": [16, 32, 1024], + "seq_length": [16, 32], "worker": 1, "adapters": ["alpaca"], "tokenizer": "unsloth/gemma-7b" }, "phi2-unmerged-lora": { "batch_size": [4], - "seq_length": [16, 32, 1024], + "seq_length": [16, 32], "worker": 1, "adapter_names": ["sql"], "tokenizer": "microsoft/phi-2" @@ -603,7 +603,7 @@ def get_model_name(): }, "gemma-7b-unmerged-lora": { "batch_size": [4], - "seq_length": [16, 32, 1024], + "seq_length": [16, 32], "worker": 1, "adapters": ["alpaca"], "tokenizer": "unsloth/gemma-7b" diff --git a/tests/integration/llm/prepare.py b/tests/integration/llm/prepare.py index 3fe50ca846..8fc9d5b602 100644 --- a/tests/integration/llm/prepare.py +++ b/tests/integration/llm/prepare.py @@ -608,11 +608,11 @@ "phi2-unmerged-lora": { "option.model_id": "s3://djl-llm/phi-2/", "option.tensor_parallel_degree": "max", - "option.adapters": "adapters", "option.enable_lora": "true", "option.max_loras": 1, "option.max_lora_rank": 64, "option.long_lora_scaling_factors": "4.0", + "option.adapters": "adapters", "adapter_ids": [ "isotr0py/phi-2-test-sql-lora", ], @@ -959,11 +959,11 @@ "phi2-unmerged-lora": { "option.model_id": "s3://djl-llm/phi-2/", "option.tensor_parallel_degree": "max", - "option.adapters": "adapters", "option.enable_lora": "true", "option.max_loras": 1, "option.max_lora_rank": 64, "option.long_lora_scaling_factors": "4.0", + "option.adapters": "adapters", "adapter_ids": [ "isotr0py/phi-2-test-sql-lora", ],