Skip to content

Commit 3df515e

Browse files
authored
update: Inferencepool API group to use v1 than v1alpha2 (#199)
Signed-off-by: Wen Zhou <wenzhou@redhat.com>
1 parent 40732e3 commit 3df515e

File tree

2 files changed

+3
-3
lines changed

2 files changed

+3
-3
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@ Once a model is deployed, inference requests must be routed to it. To do this, t
3838

3939
Note that when using the GAIE [inferencepool chart](https://github.com/kubernetes-sigs/gateway-api-inference-extension/tree/main/config/charts/inferencepool) together with the modelservice chart the following relationships will exist:
4040

41-
- The modelservice field `modelArtifact.routing.servicePort` should match the GAIE field `inferencePool.targetPortNumber` or be an entry in the list `inferencePool.targets` (depending on the apiVersion of InferencePool).
41+
- The modelservice field `modelArtifact.routing.servicePort` should match the GAIE field `inferencePool.targetPortNumber` (for InferencePool v1alpha2) or be an entry in the list `inferencePool.targets` (for InferencePool v1).
4242
- The modelservice field `modelArtifact.labels` should match the GAIE field, `inferencePool.modelServers.matchLabels`.
4343
Note that the field `llm-d.ai/role` will be addition in addition to the labels specified in the `modelArtifacts.labels` field.
4444

examples/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -95,7 +95,7 @@ spec:
9595
name: INSERT_GATEWAY_NAME
9696
rules:
9797
- backendRefs:
98-
- group: inference.networking.x-k8s.io
98+
- group: inference.networking.k8s.io
9999
kind: InferencePool
100100
name: INSERT_INFERENCEPOOL_NAME
101101
port: 8000
@@ -122,7 +122,7 @@ spec:
122122
name: INSERT_GATEWAY_NAME
123123
rules:
124124
- backendRefs:
125-
- group: inference.networking.x-k8s.io
125+
- group: inference.networking.k8s.io
126126
kind: InferencePool
127127
name: INSERT_INFERENCEPOOL_NAME
128128
port: 8000

0 commit comments

Comments
 (0)