Skip to content

Commit a1d8b5e

Browse files
committed
doc
Signed-off-by: Cody Yu <hao.yu.cody@gmail.com>
1 parent e06d14f commit a1d8b5e

File tree

1 file changed

+4
-4
lines changed

1 file changed

+4
-4
lines changed

vllm/v1/engine/core.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -182,10 +182,10 @@ def step_with_batch_queue(self) -> Optional[EngineCoreOutputs]:
182182
Note that if nothing to output in this step, None is returned.
183183
184184
The execution flow is as follows:
185-
1. Try to schedule a new batch if there are unscheduled requests
186-
and the job queue is not full. If a new batch is scheduled, directly
187-
return an empty engine core output. In other words, we won't check
188-
and return model outputs before the batch queue is full.
185+
1. Try to schedule a new batch if the batch queue is not full.
186+
If a new batch is scheduled, directly return an empty engine core
187+
output. In other words, fulfilling the batch queue has a higher priority
188+
then getting model outputs.
189189
2. If there is no new scheduled batch, meaning that the batch queue
190190
is full or no other requests can be scheduled, we block until the first
191191
batch in the job queue is finished.

0 commit comments

Comments
 (0)