Skip to content

Commit 9fb5f03

Browse files
biswapandaPeaBrane
andauthored
fix: pin sniffio dependency (#4665)
Signed-off-by: PeaBrane <[email protected]> Co-authored-by: PeaBrane <[email protected]>
1 parent fb4432e commit 9fb5f03

File tree

7 files changed

+17
-12
lines changed

7 files changed

+17
-12
lines changed

.github/workflows/container-validation-dynamo.yml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -65,7 +65,7 @@ jobs:
6565
docker compose down
6666
- name: Run pytest (parallel tests with xdist)
6767
env:
68-
PYTEST_MARKS: "pre_merge and parallel and not (vllm or trtllm or sglang or k8s)"
68+
PYTEST_MARKS: "pre_merge and parallel"
6969
run: |
7070
docker run -w /workspace \
7171
--name ${{ env.CONTAINER_ID }}_pytest_parallel \
@@ -77,7 +77,7 @@ jobs:
7777
docker cp ${{ env.CONTAINER_ID }}_pytest_parallel:/workspace/${{ env.PYTEST_PARALLEL_XML_FILE }} . || echo "No parallel test report found"
7878
- name: Run pytest (sequential tests)
7979
env:
80-
PYTEST_MARKS: "(pre_merge and not parallel and not (vllm or trtllm or sglang or k8s)) or mypy"
80+
PYTEST_MARKS: "(pre_merge and not parallel) or mypy"
8181
run: |
8282
docker run -w /workspace \
8383
--name ${{ env.CONTAINER_ID }}_pytest \

container/Dockerfile.vllm

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -174,6 +174,9 @@ WORKDIR /workspace
174174
ENV DYNAMO_HOME=/opt/dynamo
175175
ENV VIRTUAL_ENV=/opt/dynamo/venv
176176
ENV PATH="${VIRTUAL_ENV}/bin:${PATH}"
177+
# Set CUDA_DEVICE_ORDER to ensure CUDA logical device IDs match NVML physical device IDs
178+
# This fixes NVML InvalidArgument errors when CUDA_VISIBLE_DEVICES is set
179+
ENV CUDA_DEVICE_ORDER=PCI_BUS_ID
177180

178181
ARG ARCH_ALT
179182
ARG PYTHON_VERSION

container/deps/requirements.txt

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -43,6 +43,9 @@ PyYAML==6.0.3
4343
scikit-learn==1.7.2
4444
scipy<1.14.0 # Upper bound for pmdarima compatibility
4545
sentencepiece==0.2.1
46+
# Required by kr8s
47+
# https://github.com/kr8s-org/kr8s/blob/750022c3ebbb7988cddb5a979aca2ee8074a1069/examples/kubectl-ng/uv.lock#L988
48+
sniffio==1.3.1
4649
tensorboard==2.19.0
4750
tensorboardX==2.6.2.2
4851
# Transformers version constraint for container builds

tests/planner/unit/test_prometheus.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -140,7 +140,7 @@ def test_get_average_metric_none_result():
140140
mock_query.return_value = None
141141

142142
result = client._get_average_metric(
143-
metric_name="test_metric",
143+
full_metric_name="test_metric",
144144
interval="60s",
145145
operation_name="test operation",
146146
model_name="test_model",
@@ -157,7 +157,7 @@ def test_get_average_metric_empty_result():
157157
mock_query.return_value = []
158158

159159
result = client._get_average_metric(
160-
metric_name="test_metric",
160+
full_metric_name="test_metric",
161161
interval="60s",
162162
operation_name="test operation",
163163
model_name="test_model",
@@ -175,7 +175,7 @@ def test_get_average_metric_no_matching_containers(mock_prometheus_result):
175175
mock_query.return_value = [mock_prometheus_result[0]]
176176

177177
result = client._get_average_metric(
178-
metric_name="test_metric",
178+
full_metric_name="test_metric",
179179
interval="60s",
180180
operation_name="test operation",
181181
model_name="target_model",
@@ -193,7 +193,7 @@ def test_get_average_metric_one_matching_container(mock_prometheus_result):
193193
mock_query.return_value = mock_prometheus_result[:2]
194194

195195
result = client._get_average_metric(
196-
metric_name="test_metric",
196+
full_metric_name="test_metric",
197197
interval="60s",
198198
operation_name="test operation",
199199
model_name="target_model",
@@ -227,7 +227,7 @@ def test_get_average_metric_with_validation_error():
227227
mock_query.return_value = mock_result
228228

229229
result = client._get_average_metric(
230-
metric_name="test_metric",
230+
full_metric_name="test_metric",
231231
interval="60s",
232232
operation_name="test operation",
233233
model_name="target_model",
@@ -245,7 +245,7 @@ def test_get_average_metric_multiple_matching_containers(mock_prometheus_result)
245245
mock_query.return_value = mock_prometheus_result[1:]
246246

247247
result = client._get_average_metric(
248-
metric_name="test_metric",
248+
full_metric_name="test_metric",
249249
interval="60s",
250250
operation_name="test operation",
251251
model_name="target_model",

tests/serve/test_sglang.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -44,7 +44,7 @@ class SGLangConfig(EngineConfig):
4444
name="aggregated",
4545
directory=sglang_dir,
4646
script_name="agg.sh",
47-
marks=[pytest.mark.gpu_1, pytest.mark.pre_merge],
47+
marks=[pytest.mark.gpu_1],
4848
model="Qwen/Qwen3-0.6B",
4949
env={},
5050
models_port=8000,

tests/serve/test_trtllm.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -40,7 +40,7 @@ class TRTLLMConfig(EngineConfig):
4040
name="aggregated",
4141
directory=trtllm_dir,
4242
script_name="agg_metrics.sh",
43-
marks=[pytest.mark.gpu_1, pytest.mark.trtllm, pytest.mark.pre_merge],
43+
marks=[pytest.mark.gpu_1, pytest.mark.trtllm],
4444
model="Qwen/Qwen3-0.6B",
4545
models_port=8000,
4646
request_payloads=[
@@ -140,7 +140,6 @@ def test_deployment(trtllm_config_test, request, runtime_services, predownload_m
140140
# TODO make this a normal guy
141141
@pytest.mark.e2e
142142
@pytest.mark.gpu_1
143-
@pytest.mark.pre_merge
144143
@pytest.mark.trtllm
145144
def test_chat_only_aggregated_with_test_logits_processor(
146145
request, runtime_services, predownload_models, monkeypatch

tests/serve/test_vllm.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,7 @@ class VLLMConfig(EngineConfig):
4343
name="aggregated",
4444
directory=vllm_dir,
4545
script_name="agg.sh",
46-
marks=[pytest.mark.gpu_1, pytest.mark.pre_merge],
46+
marks=[pytest.mark.gpu_1],
4747
model="Qwen/Qwen3-0.6B",
4848
request_payloads=[
4949
chat_payload_default(),

0 commit comments

Comments
 (0)