-
-
Notifications
You must be signed in to change notification settings - Fork 4.4k
Issues: vllm-project/vllm
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Feature]: Llama 3 and Command-R Chat Templates
feature request
#9904
opened Nov 1, 2024 by
hammer-ai
1 task done
[Feature]: automatically release graphics card memory
feature request
#9903
opened Nov 1, 2024 by
turkeymz
1 task done
[Feature]: Support guided decoding with multistep decoding
feature request
#9893
opened Oct 31, 2024 by
joerunde
1 task done
[Feature]: Online video support for VLMs
feature request
#9842
opened Oct 30, 2024 by
litianjian
1 task done
[Feature]: host wheel via pypi index?
feature request
#9831
opened Oct 30, 2024 by
youkaichao
1 task done
[Feature]: Public binary release supports torch 2.5.1
feature request
#9830
opened Oct 30, 2024 by
luohao123
1 task done
[Feature]: Integrate Writing in the Margins inference pattern ($5,000 Bounty)
feature request
#9807
opened Oct 29, 2024 by
melisa-writer
1 task done
[Feature]: soft limit-mm-per-prompt for MM OAI API
feature request
#9805
opened Oct 29, 2024 by
SinanAkkoyun
1 task done
[Feature]: Is it supported Qwen2.5 tool_choice: auto?
feature request
#9789
opened Oct 29, 2024 by
deku0818
1 task done
[Feature]: Image-Modality Throughput Benchmark
feature request
#9778
opened Oct 28, 2024 by
lk-chen
1 task done
[Feature]: Qwen2.5 model : ValueError: This model does not support the 'embedding' task. Supported tasks: {'generate'}
feature request
#9761
opened Oct 28, 2024 by
tarikbeijing
1 task done
[Feature]: Support a
flush_cache
API to clean the kvcache after load_weights
feature request
#9744
opened Oct 28, 2024 by
zhuzilin
1 task done
[Feature]: OpenAI supply image path
feature request
#9742
opened Oct 27, 2024 by
SinanAkkoyun
1 task done
[Feature]: support for Cambricon MLU
feature request
#9649
opened Oct 24, 2024 by
a120092009
1 task done
[Feature]: support
x-request-id
header
feature request
#9593
opened Oct 22, 2024 by
cjackal
1 task done
[Feature]: Support for 1.58-bit models.
feature request
#9566
opened Oct 21, 2024 by
RealMrCactus
1 task done
[Feature]: Support for Controlled Decoding
feature request
#9541
opened Oct 21, 2024 by
simonucl
1 task done
[Feature]: EAGLE fp8 quantization
feature request
#9534
opened Oct 20, 2024 by
fengyang95
1 task done
[Feature]: google/gemma-2-2b supports 8K context length but vllm does not support it.
feature request
#9517
opened Oct 18, 2024 by
yananchen1989
1 task done
[Feature]: LoRA support for InternVLChatModel
feature request
#9495
opened Oct 18, 2024 by
AkshataABhat
1 task done
[Feature]: Support for Diff-Transformer to limit noise in attention calculation @ runtime
feature request
#9480
opened Oct 18, 2024 by
nightflight-dk
1 task done
[Feature]: Alternating local-global attention layers
feature request
#9464
opened Oct 17, 2024 by
griff4692
1 task done
Previous Next
ProTip!
Add no:assignee to see everything that’s not assigned.