We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
2 parents 4c6db55 + 8bee2e8 commit ef628a3Copy full SHA for ef628a3
Dockerfile
@@ -1,4 +1,4 @@
1
-FROM nvidia/cuda:12.3.1-devel-ubuntu20.04
+FROM nvidia/cuda:12.4.1-devel-ubuntu20.04
2
3
# Non-interactive apt-get commands
4
ARG DEBIAN_FRONTEND=noninteractive
@@ -41,8 +41,10 @@ COPY . /vec-inf
41
42
# Install project dependencies with build requirements
43
RUN PIP_INDEX_URL="https://download.pytorch.org/whl/cu121" uv pip install --system -e .[dev]
44
-# Install Flash Attention
+# Install FlashAttention
45
RUN python3.10 -m pip install flash-attn --no-build-isolation
46
+# Install FlashInfer
47
+RUN python3.10 -m pip install flashinfer-python -i https://flashinfer.ai/whl/cu124/torch2.6/
48
49
# Final configuration
50
RUN mkdir -p /vec-inf/nccl && \
0 commit comments