Skip to content

Commit 86e81da

Browse files
committed
build: add new vLLM CUDA image
Signed-off-by: Dorin Geman <[email protected]>
1 parent b8802e9 commit 86e81da

File tree

2 files changed

+42
-0
lines changed

2 files changed

+42
-0
lines changed

.github/workflows/release.yml

Lines changed: 27 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -19,6 +19,11 @@ on:
1919
required: false
2020
type: string
2121
default: "latest"
22+
vllmVersion:
23+
description: 'vLLM version'
24+
required: false
25+
type: string
26+
default: "latest"
2227

2328
jobs:
2429
test:
@@ -59,6 +64,12 @@ jobs:
5964
echo "docker/model-runner:latest-cuda" >> "$GITHUB_OUTPUT"
6065
fi
6166
echo 'EOF' >> "$GITHUB_OUTPUT"
67+
echo "vllm-cuda<<EOF" >> "$GITHUB_OUTPUT"
68+
echo "docker/model-runner:${{ inputs.releaseTag }}-vllm-cuda" >> "$GITHUB_OUTPUT"
69+
if [ "${{ inputs.pushLatest }}" == "true" ]; then
70+
echo "docker/model-runner:latest-vllm-cuda" >> "$GITHUB_OUTPUT"
71+
fi
72+
echo 'EOF' >> "$GITHUB_OUTPUT"
6273
6374
- name: Log in to DockerHub
6475
uses: docker/login-action@v3
@@ -99,3 +110,19 @@ jobs:
99110
sbom: true
100111
provenance: mode=max
101112
tags: ${{ steps.tags.outputs.cuda }}
113+
114+
- name: Build vLLM CUDA image
115+
uses: docker/build-push-action@v5
116+
with:
117+
file: Dockerfile
118+
target: vllm
119+
platforms: linux/amd64
120+
build-args: |
121+
"LLAMA_SERVER_VERSION=${{ inputs.llamaServerVersion }}"
122+
"LLAMA_SERVER_VARIANT=cuda"
123+
"BASE_IMAGE=nvidia/cuda:12.9.0-runtime-ubuntu24.04"
124+
"VLLM_VERSION=${{ inputs.vllmVersion }}""
125+
push: true
126+
sbom: true
127+
provenance: mode=max
128+
tags: ${{ steps.tags.outputs.vllm-cuda }}

Dockerfile

Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -77,3 +77,18 @@ ENV LD_LIBRARY_PATH=/app/lib
7777
LABEL com.docker.desktop.service="model-runner"
7878

7979
ENTRYPOINT ["/app/model-runner"]
80+
81+
# --- vLLM variant ---
82+
FROM final AS vllm
83+
84+
ARG VLLM_VERSION
85+
86+
USER root
87+
88+
# Install uv
89+
RUN curl -LsSf https://astral.sh/uv/install.sh | sh
90+
91+
USER modelrunner
92+
93+
# Install vLLM using uv
94+
RUN uv pip install vllm==${VLLM_VERSION}

0 commit comments

Comments
 (0)