Switching to flashinfer 0.6.x pre-release wheels
This commit is contained in:
@@ -69,9 +69,9 @@ RUN --mount=type=cache,id=pip-cache,target=/root/.cache/pip \
|
|||||||
|
|
||||||
# Install FlashInfer packages
|
# Install FlashInfer packages
|
||||||
RUN --mount=type=cache,id=pip-cache,target=/root/.cache/pip \
|
RUN --mount=type=cache,id=pip-cache,target=/root/.cache/pip \
|
||||||
pip install flashinfer-python --no-deps --index-url https://flashinfer.ai/whl && \
|
pip install flashinfer-python --no-deps --index-url https://flashinfer.ai/whl --pre && \
|
||||||
pip install flashinfer-cubin --index-url https://flashinfer.ai/whl && \
|
pip install flashinfer-cubin --index-url https://flashinfer.ai/whl --pre && \
|
||||||
pip install flashinfer-jit-cache --index-url https://flashinfer.ai/whl/cu130 && \
|
pip install flashinfer-jit-cache --index-url https://flashinfer.ai/whl/cu130 --pre && \
|
||||||
pip install apache-tvm-ffi nvidia-cudnn-frontend nvidia-cutlass-dsl nvidia-ml-py tabulate
|
pip install apache-tvm-ffi nvidia-cudnn-frontend nvidia-cutlass-dsl nvidia-ml-py tabulate
|
||||||
|
|
||||||
# =========================================================
|
# =========================================================
|
||||||
|
|||||||
Reference in New Issue
Block a user