You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: DEVELOPMENT.md
+27-27Lines changed: 27 additions & 27 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -19,42 +19,39 @@ Documentation for developing the inference scheduler.
19
19
20
20
## Kind Development Environment
21
21
22
-
> [!Warning]
23
-
> This currently requires you to have manually built the vllm
24
-
> simulator separately on your local system. In a future iteration this will
25
-
> be handled automatically and will not be required. The tag for the simulator
26
-
> currently needs to be `v0.1.0`.
27
-
28
-
You can deploy the current scheduler with a Gateway API implementation into a
29
-
[Kubernetes in Docker (KIND)] cluster locally with the following:
22
+
The following deployment creates a [Kubernetes in Docker (KIND)] cluster with an inference scheduler using a Gateway API implementation, connected to the vLLM simulator.
23
+
To run the deployment, use the following command:
30
24
31
-
```console
25
+
```bash
32
26
make env-dev-kind
33
27
```
34
28
35
29
This will create a `kind` cluster (or re-use an existing one) using the system's
36
30
local container runtime and deploy the development stack into the `default`
37
31
namespace.
38
32
33
+
> [!NOTE]
34
+
> You can download the image locally using `docker pull ghcr.io/llm-d/llm-d-inference-sim:latest`, and the script will load it from your local Docker registry.
0 commit comments