Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add an option to launch cacheflow without ray #51

Merged
merged 3 commits into from
Apr 30, 2023
Merged

Add an option to launch cacheflow without ray #51

merged 3 commits into from
Apr 30, 2023

Conversation

zhuohan123
Copy link
Member

@zhuohan123 zhuohan123 commented Apr 27, 2023

Fix #23

@zhuohan123 zhuohan123 requested a review from WoosukKwon April 27, 2023 10:34
Copy link
Collaborator

@WoosukKwon WoosukKwon left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for the PR. Left some comments. BTW, could you show the latency benchmarks before and after this PR?

.gitignore Outdated Show resolved Hide resolved
cacheflow/master/server.py Outdated Show resolved Hide resolved
cacheflow/master/server.py Outdated Show resolved Hide resolved
@zhuohan123
Copy link
Member Author

zhuohan123 commented Apr 29, 2023

Latency with Ray:

ubuntu@ray-zhuohan-cf-head-6dd317a2-compute:~/nfs/cacheflow/cacheflow/benchmark$ python benchmark_latency.py --model facebook/opt-13b --use-ray
Namespace(batch_size=8, block_size=16, dtype='half', input_len=32, max_num_batched_tokens=2560, max_num_sequences=256, model='facebook/opt-13b', model_path='~/.cacheflow/model_weights', n=1, output_len=128, pipeline_parallel_size=1, seed=0, swap_space=20, tensor_parallel_size=1, use_beam_search=False, use_dummy_weights=False, use_ray=True)
2023-04-29 16:28:04,226 INFO worker.py:1622 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265
# GPU blocks: 987, # CPU blocks: 1638
SamplingParams(n=1, temperature=1.0, top_p=1.0, use_beam_search=False, stop_token_ids=set(), max_num_steps=128, num_logprobs=0, context_window_size=None)
Warm up step
Profile step: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 3/3 [00:10<00:00,  3.55s/it]
Avg latency: 3.54734468460083 seconds

Latency without Ray:

ubuntu@ray-zhuohan-cf-head-6dd317a2-compute:~/nfs/cacheflow/cacheflow/benchmark$ python benchmark_latency.py --model facebook/opt-13b
Namespace(batch_size=8, block_size=16, dtype='half', input_len=32, max_num_batched_tokens=2560, max_num_sequences=256, model='facebook/opt-13b', model_path='~/.cacheflow/model_weights', n=1, output_len=128, pipeline_parallel_size=1, seed=0, swap_space=20, tensor_parallel_size=1, use_beam_search=False, use_dummy_weights=False, use_ray=False)
# GPU blocks: 987, # CPU blocks: 1638
SamplingParams(n=1, temperature=1.0, top_p=1.0, use_beam_search=False, stop_token_ids=set(), max_num_steps=128, num_logprobs=0, context_window_size=None)
Warm up step
Profile step: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 3/3 [00:09<00:00,  3.32s/it]
Avg latency: 3.3214046160380044 seconds

Copy link
Collaborator

@WoosukKwon WoosukKwon left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks!

@zhuohan123 zhuohan123 merged commit 4858f3b into main Apr 30, 2023
@zhuohan123 zhuohan123 deleted the no-ray branch May 24, 2023 04:40
hongxiayang pushed a commit to hongxiayang/vllm that referenced this pull request Feb 13, 2024
dtrifiro pushed a commit to dtrifiro/vllm that referenced this pull request Jun 13, 2024
…karound

Dockerfile.ubi: remove vllm-nccl workaround
yukavio pushed a commit to yukavio/vllm that referenced this pull request Jul 3, 2024
Summary:
In Streaming mode, the vllm server returns responses as soon as a token
is available. However, it doesn't do it in parts, instead, each response
is already an aggregate of all the previous responses. Therefore, it is
sufficient to record just the last response.

Test:
Manual testing

Co-authored-by: Varun <varun@neuralmagic.com>
dllehr-amd pushed a commit to dllehr-amd/vllm that referenced this pull request Jul 22, 2024
…tch-1

Revert "Tune fused_moe_kernel for TP 1,2,4,8 and bf16 and fp16, updated moe kern…"
JHLEE17 pushed a commit to JHLEE17/vllm that referenced this pull request Aug 1, 2024
@alixiaodi alixiaodi mentioned this pull request Aug 2, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Add an option to disable Ray when using a single GPU
2 participants