Skip to content

Commit

Permalink
Merge pull request vllm-project#1 from bigPYJ1151/fix_ans
Browse files Browse the repository at this point in the history
Fix key cache block shape.
  • Loading branch information
bigPYJ1151 authored Oct 27, 2023
2 parents 3a4c79a + eff780b commit b5e7066
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion vllm/worker/cache_engine.py
Original file line number Diff line number Diff line change
Expand Up @@ -58,7 +58,7 @@ def get_key_block_shape(self) -> Tuple[int, int, int, int]:
element_size = torch.tensor([], dtype=self.dtype).element_size()
x = (
16 // element_size
) if not self.cache_config.cpu_only else 1 # TODO: whether need to follow the tiling method on GPU
)
return (
self.num_heads,
self.head_size // x,
Expand Down

0 comments on commit b5e7066

Please sign in to comment.