@@ -22,23 +22,23 @@ import (
2222
2323 "github.com/stretchr/testify/assert"
2424
25- "sigs.k8s.io/gateway-api-inference-extension/pkg/epp/backend"
2625 backendmetrics "sigs.k8s.io/gateway-api-inference-extension/pkg/epp/backend/metrics"
26+ "sigs.k8s.io/gateway-api-inference-extension/pkg/epp/datalayer"
2727 "sigs.k8s.io/gateway-api-inference-extension/pkg/epp/scheduling/types"
2828)
2929
3030func TestRunningRequestsSizeScorer (t * testing.T ) {
3131 tests := []struct {
3232 name string
33- pods []types.Pod
33+ endpoints []types.Endpoint
3434 expectedScoresPod map [int ]float64 // Map of pod index to expected score
3535 }{
3636 {
3737 name : "Different running queue sizes" ,
38- pods : []types.Pod {
39- & types.PodMetrics {Pod : & backend. Pod {}, MetricsState : & backendmetrics.MetricsState {RunningRequestsSize : 10 }},
40- & types.PodMetrics {Pod : & backend. Pod {}, MetricsState : & backendmetrics.MetricsState {RunningRequestsSize : 5 }},
41- & types.PodMetrics {Pod : & backend. Pod {}, MetricsState : & backendmetrics.MetricsState {RunningRequestsSize : 0 }},
38+ endpoints : []types.Endpoint {
39+ & types.PodMetrics {EndpointMetadata : & datalayer. EndpointMetadata {}, MetricsState : & backendmetrics.MetricsState {RunningRequestsSize : 10 }},
40+ & types.PodMetrics {EndpointMetadata : & datalayer. EndpointMetadata {}, MetricsState : & backendmetrics.MetricsState {RunningRequestsSize : 5 }},
41+ & types.PodMetrics {EndpointMetadata : & datalayer. EndpointMetadata {}, MetricsState : & backendmetrics.MetricsState {RunningRequestsSize : 0 }},
4242 },
4343 expectedScoresPod : map [int ]float64 {
4444 0 : 0.0 , // Longest queue (10) gets lowest score
@@ -48,9 +48,9 @@ func TestRunningRequestsSizeScorer(t *testing.T) {
4848 },
4949 {
5050 name : "Same running queue sizes" ,
51- pods : []types.Pod {
52- & types.PodMetrics {Pod : & backend. Pod {}, MetricsState : & backendmetrics.MetricsState {RunningRequestsSize : 5 }},
53- & types.PodMetrics {Pod : & backend. Pod {}, MetricsState : & backendmetrics.MetricsState {RunningRequestsSize : 5 }},
51+ endpoints : []types.Endpoint {
52+ & types.PodMetrics {EndpointMetadata : & datalayer. EndpointMetadata {}, MetricsState : & backendmetrics.MetricsState {RunningRequestsSize : 5 }},
53+ & types.PodMetrics {EndpointMetadata : & datalayer. EndpointMetadata {}, MetricsState : & backendmetrics.MetricsState {RunningRequestsSize : 5 }},
5454 },
5555 expectedScoresPod : map [int ]float64 {
5656 0 : 1.0 , // When all pods have the same queue size, they get the same neutral score
@@ -59,9 +59,9 @@ func TestRunningRequestsSizeScorer(t *testing.T) {
5959 },
6060 {
6161 name : "Zero running queue sizes" ,
62- pods : []types.Pod {
63- & types.PodMetrics {Pod : & backend. Pod {}, MetricsState : & backendmetrics.MetricsState {RunningRequestsSize : 0 }},
64- & types.PodMetrics {Pod : & backend. Pod {}, MetricsState : & backendmetrics.MetricsState {RunningRequestsSize : 0 }},
62+ endpoints : []types.Endpoint {
63+ & types.PodMetrics {EndpointMetadata : & datalayer. EndpointMetadata {}, MetricsState : & backendmetrics.MetricsState {RunningRequestsSize : 0 }},
64+ & types.PodMetrics {EndpointMetadata : & datalayer. EndpointMetadata {}, MetricsState : & backendmetrics.MetricsState {RunningRequestsSize : 0 }},
6565 },
6666 expectedScoresPod : map [int ]float64 {
6767 0 : 1.0 ,
@@ -74,9 +74,9 @@ func TestRunningRequestsSizeScorer(t *testing.T) {
7474
7575 for _ , test := range tests {
7676 t .Run (test .name , func (t * testing.T ) {
77- scores := scorer .Score (context .Background (), types .NewCycleState (), & types.LLMRequest {}, test .pods )
77+ scores := scorer .Score (context .Background (), types .NewCycleState (), & types.LLMRequest {}, test .endpoints )
7878
79- for i , pod := range test .pods {
79+ for i , pod := range test .endpoints {
8080 expectedScore := test .expectedScoresPod [i ]
8181 assert .InDelta (t , expectedScore , scores [pod ], 0.0001 , "Pod %d should have score %f" , i , expectedScore )
8282 }
0 commit comments