@@ -234,20 +234,13 @@ func TestPrefixCacheTracking_Score(t *testing.T) {
234234 })
235235 require .NoError (t , err )
236236
237- // render the chat template
238- renderReq := & preprocessing.ApplyChatTemplateRequest {
237+ // render the chat template and tokenize
238+ renderReq := & preprocessing.RenderChatRequest {
239239 Key : tokenizerCacheKey ,
240- Conversation : [][]preprocessing. Conversation { conversations } ,
240+ Conversation : conversations ,
241241 ChatTemplate : req .ChatCompletions .ChatTemplate ,
242242 }
243- rendered , err := processor .ApplyChatTemplate (t .Context (), renderReq )
244- require .NoError (t , err )
245-
246- // tokenize rendered prompt
247- testTokenizer , err := tokenization .NewCachedLocalTokenizer (t .Context (), model , localTokenizerConfig )
248- require .NoError (t , err )
249-
250- tokens , _ , err := testTokenizer .Encode (rendered , model , false )
243+ tokens , _ , err := processor .RenderChat (t .Context (), renderReq )
251244 require .NoError (t , err )
252245
253246 tokenProcessor := kvblock .NewChunkedTokenDatabase (kvblock .DefaultTokenProcessorConfig ())
@@ -539,10 +532,9 @@ func TestPrefixCacheTracking_Score(t *testing.T) {
539532
540533 kvcacheConfig , err := kvcache .NewDefaultConfig ()
541534 kvcacheConfig .TokenizersPoolConfig = & tokenization.Config {
542- ModelName : "test-model" ,
543- WorkersCount : 1 ,
544- MinPrefixOverlapRatio : 0.8 ,
545- LocalTokenizerConfig : & localTokenizerConfig ,
535+ ModelName : "test-model" ,
536+ WorkersCount : 1 ,
537+ LocalTokenizerConfig : & localTokenizerConfig ,
546538 }
547539 require .NoError (t , err )
548540
0 commit comments