diff --git a/container/deps/vllm/install_vllm.sh b/container/deps/vllm/install_vllm.sh index 8365deecf6..8dabc5c1a3 100755 --- a/container/deps/vllm/install_vllm.sh +++ b/container/deps/vllm/install_vllm.sh @@ -126,7 +126,7 @@ git checkout $VLLM_REF echo "\n=== Installing vLLM & FlashInfer ===" echo "Installing vLLM $VLLM_REF from PyPI..." -uv pip install vllm[flashinfer]==$VLLM_REF --torch-backend=${TORCH_BACKEND} +uv pip install vllm[flashinfer,runai]==$VLLM_REF --torch-backend=${TORCH_BACKEND} uv pip install flashinfer-cubin==$FLASHINF_REF uv pip install flashinfer-jit-cache==$FLASHINF_REF --extra-index-url https://flashinfer.ai/whl/${TORCH_BACKEND} diff --git a/pyproject.toml b/pyproject.toml index 2100d47951..04e7a78bb2 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -56,7 +56,7 @@ trtllm =[ vllm = [ "uvloop", "nixl[cu12]<=0.7.1", - "vllm[flashinfer]==0.12.0", + "vllm[flashinfer,runai]==0.12.0", ] sglang = [