We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 363d548 commit b3f7431Copy full SHA for b3f7431
vllm/platforms/xpu.py
@@ -80,13 +80,11 @@ def inference_mode(cls):
80
@classmethod
81
def check_and_update_config(cls, vllm_config: VllmConfig) -> None:
82
cache_config = vllm_config.cache_config
83
- # in V1(or with ipex chunked prefill) block_size is 64
84
- if cache_config and \
85
- cache_config.block_size is None and \
86
- envs.VLLM_USE_V1:
87
- cache_config.block_size = 64
88
if cache_config and cache_config.block_size is None:
89
- cache_config.block_size = 16
+ if envs.VLLM_USE_V1:
+ cache_config.block_size = 64
+ else:
+ cache_config.block_size = 16
90
91
# Instances created using VllmConfig() typically have model_config as
92
# None by default. The modification involves adding a check to prevent
0 commit comments