@@ -683,15 +683,15 @@ async def test2():
683
683
684
684
@pytest .mark .asyncio
685
685
@pytest .mark .parametrize ("sync_mode" , [True , False ])
686
- async def test_async_router_context_window_fallback (sync_mode , large_text ):
686
+ async def test_async_router_context_window_fallback (sync_mode ):
687
687
"""
688
688
- Give a gpt-4 model group with different context windows (8192k vs. 128k)
689
689
- Send a 10k prompt
690
690
- Assert it works
691
691
"""
692
692
import os
693
693
694
- text = large_text
694
+ from large_text import text
695
695
696
696
litellm .set_verbose = False
697
697
litellm ._turn_on_debug ()
@@ -775,17 +775,15 @@ def test_router_rpm_pre_call_check():
775
775
pytest .fail (f"Got unexpected exception on router! - { str (e )} " )
776
776
777
777
778
- def test_router_context_window_check_pre_call_check_in_group_custom_model_info (
779
- large_text ,
780
- ):
778
+ def test_router_context_window_check_pre_call_check_in_group_custom_model_info ():
781
779
"""
782
780
- Give a gpt-3.5-turbo model group with different context windows (4k vs. 16k)
783
781
- Send a 5k prompt
784
782
- Assert it works
785
783
"""
786
784
import os
787
785
788
- text = large_text
786
+ from large_text import text
789
787
790
788
litellm .set_verbose = False
791
789
@@ -831,15 +829,15 @@ def test_router_context_window_check_pre_call_check_in_group_custom_model_info(
831
829
pytest .fail (f"Got unexpected exception on router! - { str (e )} " )
832
830
833
831
834
- def test_router_context_window_check_pre_call_check (large_text ):
832
+ def test_router_context_window_check_pre_call_check ():
835
833
"""
836
834
- Give a gpt-3.5-turbo model group with different context windows (4k vs. 16k)
837
835
- Send a 5k prompt
838
836
- Assert it works
839
837
"""
840
838
import os
841
839
842
- text = large_text
840
+ from large_text import text
843
841
844
842
litellm .set_verbose = False
845
843
@@ -885,15 +883,15 @@ def test_router_context_window_check_pre_call_check(large_text):
885
883
pytest .fail (f"Got unexpected exception on router! - { str (e )} " )
886
884
887
885
888
- def test_router_context_window_check_pre_call_check_out_group (large_text ):
886
+ def test_router_context_window_check_pre_call_check_out_group ():
889
887
"""
890
888
- Give 2 gpt-3.5-turbo model groups with different context windows (4k vs. 16k)
891
889
- Send a 5k prompt
892
890
- Assert it works
893
891
"""
894
892
import os
895
893
896
- text = large_text
894
+ from large_text import text
897
895
898
896
litellm .set_verbose = False
899
897
@@ -1138,7 +1136,7 @@ async def test_aimg_gen_on_router():
1138
1136
"api_base" : os .getenv ("AZURE_SWEDEN_API_BASE" ),
1139
1137
"api_key" : os .getenv ("AZURE_SWEDEN_API_KEY" ),
1140
1138
},
1141
- },
1139
+ }
1142
1140
]
1143
1141
router = Router (model_list = model_list , num_retries = 3 )
1144
1142
response = await router .aimage_generation (
@@ -2789,5 +2787,4 @@ def test_router_get_model_group_info():
2789
2787
assert model_group_info is not None
2790
2788
assert model_group_info .model_group == "gpt-4"
2791
2789
assert model_group_info .input_cost_per_token > 0
2792
-
2793
- assert model_group_info .output_cost_per_token > 0
2790
+ assert model_group_info .output_cost_per_token > 0
0 commit comments