Skip to content

Commit 4c1f949

Browse files
committed
Make vLLM port number explicit in InferenceServerConfig
Signed-off-by: Jun Duan <jun.duan.phd@outlook.com>
1 parent c070b6d commit 4c1f949

File tree

3 files changed

+23
-2
lines changed

3 files changed

+23
-2
lines changed

api/fma/v1alpha1/inferenceserverconfig_types.go

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,11 @@ type InferenceServerConfigSpec struct {
3333

3434
// ModelServerConfig defines the configuration for a model server
3535
type ModelServerConfig struct {
36-
// Options are the vLLM startup options
36+
// Port is the port on which the vLLM server will listen
37+
// Particularly, management of vLLM instances' sleep state is done through this port
38+
Port int32 `json:"port"`
39+
40+
// Options are the vLLM startup options, excluding Port
3741
// +optional
3842
Options string `json:"options,omitempty"`
3943

config/crd/fma.llm-d.ai_inferenceserverconfigs.yaml

Lines changed: 9 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -66,8 +66,16 @@ spec:
6666
type: string
6767
type: object
6868
options:
69-
description: Options are the vLLM startup options
69+
description: Options are the vLLM startup options, excluding Port
7070
type: string
71+
port:
72+
description: |-
73+
Port is the port on which the vLLM server will listen
74+
Particularly, management of vLLM instances' sleep state is done through this port
75+
format: int32
76+
type: integer
77+
required:
78+
- port
7179
type: object
7280
required:
7381
- launcherConfigName

pkg/generated/applyconfiguration/fma/v1alpha1/modelserverconfig.go

Lines changed: 9 additions & 0 deletions
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

0 commit comments

Comments
 (0)