Skip to content

Commit bb0a761

Browse files
Set vllm-hpu-extension to 6ac93fb (#684)
remove expert_max hard code (#47) vLLM-Ext: Full enabling of ALiBi (#34) Add version inference via setuptools-scm (#58) Revert "vLLM-Ext: Full enabling of ALiBi (#34)" (#59) Remove punica_hpu.py from vllm_hpu_extension (#66) Removed previous (not-pipelined) pa implementation (#72) Add flag to enable running softmax in fp32 (#71) Update calibration readme link (#73) allow lm_head quantization in calibration process (#65) Pad to bmin if value is less (#67) Update pyproject.toml (#75) --------- Co-authored-by: Michał Kuligowski <mkuligowski@habana.ai>
1 parent ffcf007 commit bb0a761

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

requirements-hpu.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,4 +8,4 @@ pandas
88
tabulate
99
setuptools>=61
1010
setuptools-scm>=8
11-
vllm-hpu-extension @ git+https://github.com/HabanaAI/vllm-hpu-extension.git@4312768
11+
vllm-hpu-extension @ git+https://github.com/HabanaAI/vllm-hpu-extension.git@d05c0a7

0 commit comments

Comments
 (0)