Skip to content

Commit 1650d36

Browse files
committed
support vllm==0.4.2
1 parent e46e480 commit 1650d36

File tree

4 files changed

+5
-5
lines changed

4 files changed

+5
-5
lines changed

api/config.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -203,8 +203,8 @@ class VLLMSetting(BaseModel):
203203
default=get_bool_env("ENFORCE_EAGER"),
204204
description="Always use eager-mode PyTorch. If False, will use eager mode and CUDA graph in hybrid for maximal performance and flexibility."
205205
)
206-
max_context_len_to_capture: Optional[int] = Field(
207-
default=int(get_env("MAX_CONTEXT_LEN_TO_CAPTURE", 8192)),
206+
max_seq_len_to_capture: Optional[int] = Field(
207+
default=int(get_env("MAX_SEQ_LEN_TO_CAPTURE", 8192)),
208208
description="aximum context length covered by CUDA graphs. When a sequence has context length larger than this, we fall back to eager mode."
209209
)
210210
max_loras: Optional[int] = Field(

api/models.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -107,7 +107,7 @@ def create_vllm_engine():
107107
"gpu_memory_utilization",
108108
"max_num_seqs",
109109
"enforce_eager",
110-
"max_context_len_to_capture",
110+
"max_seq_len_to_capture",
111111
"max_loras",
112112
"max_lora_rank",
113113
"lora_extra_vocab_size",

docker/Dockerfile.vllm

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,6 +6,6 @@ COPY requirements.txt /workspace/
66

77
RUN pip config set global.index-url https://pypi.tuna.tsinghua.edu.cn/simple && \
88
pip install bitsandbytes --upgrade && \
9-
pip install vllm==0.4.0 && \
9+
pip install vllm==0.4.2 && \
1010
pip install --no-cache-dir -r /workspace/requirements.txt && \
1111
pip uninstall transformer-engine -y

docs/VLLM_SCRIPT.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@ docker build -f docker/Dockerfile.vllm -t llm-api:vllm .
1616

1717
```shell
1818
pip install torch==2.1.0
19-
pip install vllm==0.4.0
19+
pip install vllm==0.4.2
2020
pip install -r requirements.txt
2121
pip uninstall transformer-engine -y
2222
```

0 commit comments

Comments
 (0)