Removed temporary patch for NVFP4 quants support as it's been merged into main
This commit is contained in:
@@ -75,8 +75,8 @@ RUN python3 use_existing_torch.py && \
|
|||||||
sed -i "/flashinfer/d" requirements/cuda.txt && \
|
sed -i "/flashinfer/d" requirements/cuda.txt && \
|
||||||
pip install -r requirements/build.txt
|
pip install -r requirements/build.txt
|
||||||
|
|
||||||
# TEMPORARY - apply NVFP4 patch
|
# TEMPORARY - apply NVFP4 patch - MERGED ON 11/25/2025 - TODO: test and remove from Dockerfile
|
||||||
RUN curl -L https://patch-diff.githubusercontent.com/raw/vllm-project/vllm/pull/29242.diff | git apply
|
# RUN curl -L https://patch-diff.githubusercontent.com/raw/vllm-project/vllm/pull/29242.diff | git apply
|
||||||
|
|
||||||
# Final Build
|
# Final Build
|
||||||
# Uses --no-build-isolation to respect the pre-installed Torch/FlashInfer
|
# Uses --no-build-isolation to respect the pre-installed Torch/FlashInfer
|
||||||
|
|||||||
Reference in New Issue
Block a user