Skip to content

Commit c935b70

Browse files
authored
[doc] fix feature support (vllm-project#70)
Check and update the feature support table. - both multi-step and speculative decoding require adaptation of corresponding workers - prompt adapter (finetune method) require adaption in worker.py and model_runner.py Signed-off-by: MengqingCao <cmq0113@163.com>
1 parent 36ea38f commit c935b70

File tree

1 file changed

+8
-6
lines changed

1 file changed

+8
-6
lines changed

docs/source/features/suppoted_features.md

Lines changed: 8 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -3,17 +3,19 @@
33
| Feature | Supported | Note |
44
|---------|-----------|------|
55
| Chunked Prefill || Plan in 2025 Q1 |
6-
| Automatic Prefix Caching || Improve performance in 2025 Q1 |
6+
| Automatic Prefix Caching || Improve performance in 2025 Q2 |
77
| LoRA || Plan in 2025 Q1 |
8-
| Prompt adapter | ||
9-
| Speculative decoding | | Improve accuracy in 2025 Q1|
10-
| Pooling || Plan in 2025 Q1 |
11-
| Enc-dec || Plan in 2025 Q1 |
8+
| Prompt adapter | | Plan in 2025 Q1 |
9+
| Speculative decoding | | Plan in 2025 Q1 |
10+
| Pooling || Plan in 2025 Q2 |
11+
| Enc-dec || Plan in 2025 Q2 |
1212
| Multi Modality | ✅ (LLaVA/Qwen2-vl/Qwen2-audio/internVL)| Add more model support in 2025 Q1 |
1313
| LogProbs |||
1414
| Prompt logProbs |||
1515
| Async output |||
16-
| Multi step scheduler | ||
16+
| Multi step scheduler | | Plan in 2025 Q1 |
1717
| Best of |||
1818
| Beam search |||
1919
| Guided Decoding || Plan in 2025 Q1 |
20+
| Tensor Parallel || Only "mp" supported now |
21+
| Pipeline Parallel || Only "mp" supported now |

0 commit comments

Comments
 (0)