Skip to content

Commit 55bc8e7

Browse files
committed
lint
1 parent 7e8c4a4 commit 55bc8e7

5 files changed

Lines changed: 22 additions & 21 deletions

File tree

skyrl/backends/skyrl_train/inference_engines/ray_wrapped_inference_engine.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -136,7 +136,6 @@ def create_ray_wrapped_inference_engines(
136136

137137
if backend == "vllm":
138138
import vllm
139-
140139
from skyrl.backends.skyrl_train.inference_engines.vllm.vllm_engine import (
141140
AsyncVLLMRayActor,
142141
VLLMRayActor,

skyrl/backends/skyrl_train/inference_engines/vllm/vllm_engine.py

Lines changed: 16 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -13,8 +13,23 @@
1313
from uuid import uuid4
1414

1515
import ray
16-
import vllm
1716
from loguru import logger
17+
18+
import vllm
19+
from skyrl.backends.skyrl_train.inference_engines.base import (
20+
InferenceEngineInput,
21+
InferenceEngineInterface,
22+
InferenceEngineOutput,
23+
)
24+
from skyrl.backends.skyrl_train.inference_engines.vllm.utils import pop_openai_kwargs
25+
26+
# Backward compatibility: WorkerWrap has moved to inference_servers.vllm_worker
27+
# This alias preserves the old import path for existing scripts/configs.
28+
# TODO (Kourosh): Remove this alias once all references are updated.
29+
from skyrl.backends.skyrl_train.inference_servers.vllm_worker import (
30+
WorkerWrap, # noqa: F401, E402
31+
)
32+
from skyrl.backends.skyrl_train.weight_sync import WeightLoader, WeightUpdateRequest
1833
from vllm import SamplingParams
1934
from vllm.entrypoints.openai.chat_completion.protocol import (
2035
ChatCompletionRequest,
@@ -36,21 +51,6 @@
3651
from vllm.inputs import TokensPrompt
3752
from vllm.lora.request import LoRARequest
3853

39-
from skyrl.backends.skyrl_train.inference_engines.base import (
40-
InferenceEngineInput,
41-
InferenceEngineInterface,
42-
InferenceEngineOutput,
43-
)
44-
from skyrl.backends.skyrl_train.inference_engines.vllm.utils import pop_openai_kwargs
45-
46-
# Backward compatibility: WorkerWrap has moved to inference_servers.vllm_worker
47-
# This alias preserves the old import path for existing scripts/configs.
48-
# TODO (Kourosh): Remove this alias once all references are updated.
49-
from skyrl.backends.skyrl_train.inference_servers.vllm_worker import (
50-
WorkerWrap, # noqa: F401, E402
51-
)
52-
from skyrl.backends.skyrl_train.weight_sync import WeightLoader, WeightUpdateRequest
53-
5454

5555
@dataclass
5656
class Logprob:

skyrl/backends/skyrl_train/inference_engines/vllm/vllm_server.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,8 +2,9 @@
22
import signal
33

44
import uvloop
5-
import vllm.envs as envs
65
from fastapi import Request
6+
7+
import vllm.envs as envs
78
from vllm import AsyncLLMEngine
89
from vllm.engine.arg_utils import AsyncEngineArgs
910
from vllm.entrypoints.launcher import serve_http

skyrl/benchmarks/benchmark_memory.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -77,9 +77,10 @@
7777
from typing import Callable, Literal
7878

7979
import httpx
80+
from transformers import AutoTokenizer
81+
8082
import tinker
8183
from tinker import types
82-
from transformers import AutoTokenizer
8384

8485
# Default configuration
8586
DEFAULT_BASE_MODEL = "Qwen/Qwen3-4B-Instruct-2507"

tests/tinker/test_api.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -9,13 +9,13 @@
99
from urllib.parse import urlparse
1010

1111
import pytest
12-
import tinker
13-
from tinker import types
1412
from transformers import AutoTokenizer
1513

14+
import tinker
1615
from skyrl.tinker.api import _build_uv_run_cmd_engine
1716
from skyrl.tinker.config import EngineConfig
1817
from tests.tinker.conftest import wait_for_condition
18+
from tinker import types
1919

2020
BASE_MODEL = "trl-internal-testing/tiny-Qwen3ForCausalLM"
2121

0 commit comments

Comments
 (0)