From 0605911b7800034e16a35a75e816a261de518b8d Mon Sep 17 00:00:00 2001 From: Imran Ullah <103720343+imrankh46@users.noreply.github.com> Date: Sun, 12 Oct 2025 18:11:14 +0500 Subject: [PATCH] Update Dockerfile always use latest version of vllm for new model support --- Dockerfile | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/Dockerfile b/Dockerfile index bb22ceb..e633c7c 100644 --- a/Dockerfile +++ b/Dockerfile @@ -12,7 +12,7 @@ RUN --mount=type=cache,target=/root/.cache/pip \ python3 -m pip install --upgrade -r /requirements.txt # Install vLLM (switching back to pip installs since issues that required building fork are fixed and space optimization is not as important since caching) and FlashInfer -RUN python3 -m pip install vllm==0.10.0 && \ +RUN python3 -m pip install --upgrade vllm && \ python3 -m pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.3 # Setup for Option 2: Building the Image with the Model included