Skip to content

Commit 706d1bd

Browse files
bug: fixing start docker command, fixing openai port expose, fixing dockerfile (#111)
1 parent b8de595 commit 706d1bd

File tree

2 files changed

+6
-4
lines changed

2 files changed

+6
-4
lines changed

src/triton_cli/docker/Dockerfile

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -34,4 +34,4 @@ RUN mkdir -p /opt/tritonserver/backends/vllm && \
3434
rm -r /tmp/vllm_backend
3535

3636
# vLLM runtime dependencies
37-
RUN pip install "vllm==0.6.3.post1" "setuptools==74.0.0"
37+
RUN pip install "vllm==0.6.3.post1" "setuptools>=74.1.1"

src/triton_cli/server/server_docker.py

Lines changed: 5 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -133,13 +133,13 @@ def start(self, env=None):
133133
# Mount required directories
134134
volumes = {}
135135
# Mount model repository at same path in read-only mode for simplicity
136-
volumes[self._server_config["model-repository"]] = {
137-
"bind": self._server_config["model-repository"],
136+
volumes[str(self._server_config["model-repository"])] = {
137+
"bind": str(self._server_config["model-repository"]),
138138
"mode": "ro",
139139
}
140140
# Mount huggingface model cache to save time across runs
141141
# Use default cache in container for now.
142-
volumes[HF_CACHE] = {
142+
volumes[str(HF_CACHE)] = {
143143
"bind": "/root/.cache/huggingface",
144144
"mode": "rw",
145145
}
@@ -155,11 +155,13 @@ def start(self, env=None):
155155
server_http_port = 8000
156156
server_grpc_port = 8001
157157
server_metrics_port = 8002
158+
openai_http_port = 9000
158159

159160
ports = {
160161
server_http_port: server_http_port,
161162
server_grpc_port: server_grpc_port,
162163
server_metrics_port: server_metrics_port,
164+
openai_http_port: openai_http_port,
163165
}
164166
# Construct run command
165167
command = self._server_utils.get_launch_command(

0 commit comments

Comments
 (0)