File tree Expand file tree Collapse file tree 2 files changed +6
-4
lines changed
Expand file tree Collapse file tree 2 files changed +6
-4
lines changed Original file line number Diff line number Diff line change @@ -34,4 +34,4 @@ RUN mkdir -p /opt/tritonserver/backends/vllm && \
3434 rm -r /tmp/vllm_backend
3535
3636# vLLM runtime dependencies
37- RUN pip install "vllm==0.6.3.post1" "setuptools== 74.0.0 "
37+ RUN pip install "vllm==0.6.3.post1" "setuptools>= 74.1.1 "
Original file line number Diff line number Diff line change @@ -133,13 +133,13 @@ def start(self, env=None):
133133 # Mount required directories
134134 volumes = {}
135135 # Mount model repository at same path in read-only mode for simplicity
136- volumes [self ._server_config ["model-repository" ]] = {
137- "bind" : self ._server_config ["model-repository" ],
136+ volumes [str ( self ._server_config ["model-repository" ]) ] = {
137+ "bind" : str ( self ._server_config ["model-repository" ]) ,
138138 "mode" : "ro" ,
139139 }
140140 # Mount huggingface model cache to save time across runs
141141 # Use default cache in container for now.
142- volumes [HF_CACHE ] = {
142+ volumes [str ( HF_CACHE ) ] = {
143143 "bind" : "/root/.cache/huggingface" ,
144144 "mode" : "rw" ,
145145 }
@@ -155,11 +155,13 @@ def start(self, env=None):
155155 server_http_port = 8000
156156 server_grpc_port = 8001
157157 server_metrics_port = 8002
158+ openai_http_port = 9000
158159
159160 ports = {
160161 server_http_port : server_http_port ,
161162 server_grpc_port : server_grpc_port ,
162163 server_metrics_port : server_metrics_port ,
164+ openai_http_port : openai_http_port ,
163165 }
164166 # Construct run command
165167 command = self ._server_utils .get_launch_command (
You can’t perform that action at this time.
0 commit comments