@@ -85,11 +85,6 @@ class ModelType(UnifiedModelType, Enum):
8585 # Groq platform models
8686 GROQ_LLAMA_3_1_8B = "llama-3.1-8b-instant"
8787 GROQ_LLAMA_3_3_70B = "llama-3.3-70b-versatile"
88- GROQ_LLAMA_3_3_70B_PREVIEW = "llama-3.3-70b-specdec"
89- GROQ_LLAMA_3_8B = "llama3-8b-8192"
90- GROQ_LLAMA_3_70B = "llama3-70b-8192"
91- GROQ_MIXTRAL_8_7B = "mixtral-8x7b-32768"
92- GROQ_GEMMA_2_9B_IT = "gemma2-9b-it"
9388
9489 # Cerebras platform models
9590 CEREBRAS_GPT_OSS_120B = "gpt-oss-120b"
@@ -702,11 +697,6 @@ def is_groq(self) -> bool:
702697 return self in {
703698 ModelType .GROQ_LLAMA_3_1_8B ,
704699 ModelType .GROQ_LLAMA_3_3_70B ,
705- ModelType .GROQ_LLAMA_3_3_70B_PREVIEW ,
706- ModelType .GROQ_LLAMA_3_8B ,
707- ModelType .GROQ_LLAMA_3_70B ,
708- ModelType .GROQ_MIXTRAL_8_7B ,
709- ModelType .GROQ_GEMMA_2_9B_IT ,
710700 }
711701
712702 @property
@@ -1191,10 +1181,6 @@ def token_limit(self) -> int:
11911181 elif self in {
11921182 ModelType .CEREBRAS_LLAMA_3_1_8B ,
11931183 ModelType .GPT_4 ,
1194- ModelType .GROQ_LLAMA_3_8B ,
1195- ModelType .GROQ_LLAMA_3_70B ,
1196- ModelType .GROQ_LLAMA_3_3_70B_PREVIEW ,
1197- ModelType .GROQ_GEMMA_2_9B_IT ,
11981184 ModelType .GLM_3_TURBO ,
11991185 ModelType .GLM_4 ,
12001186 ModelType .QWEN_VL_PLUS ,
@@ -1277,7 +1263,6 @@ def token_limit(self) -> int:
12771263 ModelType .MISTRAL_CODESTRAL ,
12781264 ModelType .MISTRAL_7B ,
12791265 ModelType .MISTRAL_MIXTRAL_8x7B ,
1280- ModelType .GROQ_MIXTRAL_8_7B ,
12811266 ModelType .YI_LARGE ,
12821267 ModelType .YI_LARGE_FC ,
12831268 ModelType .QWEN_MAX ,
0 commit comments