@@ -237,7 +237,7 @@ func TestDistributor_QueryStream_ShouldReturnErrorIfMaxChunksPerQueryLimitIsReac
237237
238238 // Since the number of series (and thus chunks) is equal to the limit (but doesn't
239239 // exceed it), we expect a query running on all series to succeed.
240- queryRes , err := ds [0 ].QueryStream (queryCtx , queryMetrics , math .MinInt32 , math .MaxInt32 , false , nil , allSeriesMatchers ... )
240+ queryRes , err := ds [0 ].QueryStream (queryCtx , queryMetrics , math .MinInt32 , math .MaxInt32 , allSeriesMatchers ... )
241241 require .NoError (t , err )
242242
243243 require .Len (t , queryRes .StreamingSeries , initialSeries )
@@ -265,7 +265,7 @@ func TestDistributor_QueryStream_ShouldReturnErrorIfMaxChunksPerQueryLimitIsReac
265265
266266 // Since the number of series (and thus chunks) is exceeding to the limit, we expect
267267 // a query running on all series to fail.
268- _ , err = ds [0 ].QueryStream (queryCtx , queryMetrics , math .MinInt32 , math .MaxInt32 , false , nil , allSeriesMatchers ... )
268+ _ , err = ds [0 ].QueryStream (queryCtx , queryMetrics , math .MinInt32 , math .MaxInt32 , allSeriesMatchers ... )
269269 require .Error (t , err )
270270 require .ErrorContains (t , err , testCase .expectedError )
271271
@@ -319,7 +319,7 @@ func TestDistributor_QueryStream_ShouldReturnErrorIfMaxSeriesPerQueryLimitIsReac
319319 // exceed it), we expect a query running on all series to succeed.
320320 queryCtx := limiter .AddQueryLimiterToContext (userCtx , limiter .NewQueryLimiter (maxSeriesLimit , 0 , 0 , 0 , stats .NewQueryMetrics (prometheus .NewPedanticRegistry ())))
321321 queryCtx = limiter .ContextWithNewUnlimitedMemoryConsumptionTracker (queryCtx )
322- queryRes , err := ds [0 ].QueryStream (queryCtx , queryMetrics , math .MinInt32 , math .MaxInt32 , false , nil , allSeriesMatchers ... )
322+ queryRes , err := ds [0 ].QueryStream (queryCtx , queryMetrics , math .MinInt32 , math .MaxInt32 , allSeriesMatchers ... )
323323 require .NoError (t , err )
324324
325325 assert .Len (t , queryRes .StreamingSeries , initialSeries )
@@ -342,7 +342,7 @@ func TestDistributor_QueryStream_ShouldReturnErrorIfMaxSeriesPerQueryLimitIsReac
342342
343343 // Since the number of series is exceeding the limit, we expect
344344 // a query running on all series to fail.
345- _ , err = ds [0 ].QueryStream (queryCtx , queryMetrics , math .MinInt32 , math .MaxInt32 , false , nil , allSeriesMatchers ... )
345+ _ , err = ds [0 ].QueryStream (queryCtx , queryMetrics , math .MinInt32 , math .MaxInt32 , allSeriesMatchers ... )
346346 require .Error (t , err )
347347 assert .ErrorContains (t , err , "the query exceeded the maximum number of series" )
348348
@@ -384,7 +384,7 @@ func TestDistributor_QueryStream_ShouldReturnErrorIfMaxChunkBytesPerQueryLimitIs
384384 assert .Nil (t , err )
385385
386386 queryMetrics := stats .NewQueryMetrics (reg [0 ])
387- chunkSizeResponse , err := ds [0 ].QueryStream (ctx , queryMetrics , math .MinInt32 , math .MaxInt32 , false , nil , allSeriesMatchers ... )
387+ chunkSizeResponse , err := ds [0 ].QueryStream (ctx , queryMetrics , math .MinInt32 , math .MaxInt32 , allSeriesMatchers ... )
388388 require .NoError (t , err )
389389
390390 _ , responseChunkSize , err := countStreamingChunksAndBytes (chunkSizeResponse )
@@ -404,7 +404,7 @@ func TestDistributor_QueryStream_ShouldReturnErrorIfMaxChunkBytesPerQueryLimitIs
404404
405405 // Since the number of chunk bytes is equal to the limit (but doesn't
406406 // exceed it), we expect a query running on all series to succeed.
407- queryRes , err := ds [0 ].QueryStream (ctx , queryMetrics , math .MinInt32 , math .MaxInt32 , false , nil , allSeriesMatchers ... )
407+ queryRes , err := ds [0 ].QueryStream (ctx , queryMetrics , math .MinInt32 , math .MaxInt32 , allSeriesMatchers ... )
408408 require .NoError (t , err )
409409 assert .Len (t , queryRes .StreamingSeries , seriesToAdd )
410410
@@ -418,7 +418,7 @@ func TestDistributor_QueryStream_ShouldReturnErrorIfMaxChunkBytesPerQueryLimitIs
418418 // Since the aggregated chunk size is exceeding the limit, we expect
419419 // a query running on all series to fail but only when the chunks are
420420 // actually consumed from the stream.
421- finalResp , err := ds [0 ].QueryStream (ctx , queryMetrics , math .MinInt32 , math .MaxInt32 , false , nil , allSeriesMatchers ... )
421+ finalResp , err := ds [0 ].QueryStream (ctx , queryMetrics , math .MinInt32 , math .MaxInt32 , allSeriesMatchers ... )
422422 require .NoError (t , err )
423423
424424 _ , _ , err = countStreamingChunksAndBytes (finalResp )
@@ -491,7 +491,7 @@ func TestDistributor_QueryStream_ShouldSuccessfullyRunOnSlowIngesterWithStreamin
491491 t .Run (fmt .Sprintf ("Query #%d" , i ), func (t * testing.T ) {
492492 t .Parallel ()
493493
494- res , err := distributors [0 ].QueryStream (ctx , queryMetrics , math .MinInt32 , math .MaxInt32 , false , nil , matchers )
494+ res , err := distributors [0 ].QueryStream (ctx , queryMetrics , math .MinInt32 , math .MaxInt32 , matchers )
495495 require .NoError (t , err )
496496 require .Equal (t , numSeries , len (res .StreamingSeries ))
497497
0 commit comments