@@ -247,9 +247,12 @@ mod tests {
247247 ( "OPENAI_CUSTOM_HEADERS" , Some ( "" ) ) ,
248248 ] ) ;
249249
250- let provider = create ( "openai" , ModelConfig :: new_or_fail ( "gpt-4o-mini" ) . with_canonical_limits ( "openai" ) )
251- . await
252- . unwrap ( ) ;
250+ let provider = create (
251+ "openai" ,
252+ ModelConfig :: new_or_fail ( "gpt-4o-mini" ) . with_canonical_limits ( "openai" ) ,
253+ )
254+ . await
255+ . unwrap ( ) ;
253256 let lw = provider. as_lead_worker ( ) . unwrap ( ) ;
254257 let ( lead, worker) = lw. get_model_info ( ) ;
255258 assert_eq ! ( lead, "gpt-4o" ) ;
@@ -272,9 +275,12 @@ mod tests {
272275 ( "OPENAI_CUSTOM_HEADERS" , Some ( "" ) ) ,
273276 ] ) ;
274277
275- let provider = create ( "openai" , ModelConfig :: new_or_fail ( "gpt-4o-mini" ) . with_canonical_limits ( "openai" ) )
276- . await
277- . unwrap ( ) ;
278+ let provider = create (
279+ "openai" ,
280+ ModelConfig :: new_or_fail ( "gpt-4o-mini" ) . with_canonical_limits ( "openai" ) ,
281+ )
282+ . await
283+ . unwrap ( ) ;
278284 assert ! ( provider. as_lead_worker( ) . is_none( ) ) ;
279285 assert_eq ! ( provider. get_model_config( ) . model_name, "gpt-4o-mini" ) ;
280286 }
@@ -292,8 +298,9 @@ mod tests {
292298 ( "GOOSE_CONTEXT_LIMIT" , global_limit) ,
293299 ] ) ;
294300
295- let default_model =
296- ModelConfig :: new_or_fail ( "gpt-3.5-turbo" ) . with_canonical_limits ( "openai" ) . with_context_limit ( Some ( 16_000 ) ) ;
301+ let default_model = ModelConfig :: new_or_fail ( "gpt-3.5-turbo" )
302+ . with_canonical_limits ( "openai" )
303+ . with_context_limit ( Some ( 16_000 ) ) ;
297304
298305 let result = create_worker_model_config ( & default_model, "openai" ) . unwrap ( ) ;
299306 assert_eq ! ( result. context_limit, Some ( expected_limit) ) ;
0 commit comments