|
562 | 562 | { |
563 | 563 | "id": "deepseek/deepseek-v3.1-terminus", |
564 | 564 | "name": "DeepSeek: DeepSeek V3.1 Terminus", |
565 | | - "cost_per_1m_in": 0.27, |
566 | | - "cost_per_1m_out": 1, |
| 565 | + "cost_per_1m_in": 0.21, |
| 566 | + "cost_per_1m_out": 0.79, |
567 | 567 | "cost_per_1m_in_cached": 0, |
568 | | - "cost_per_1m_out_cached": 0, |
| 568 | + "cost_per_1m_out_cached": 0.13, |
569 | 569 | "context_window": 163840, |
570 | | - "default_max_tokens": 81920, |
| 570 | + "default_max_tokens": 16384, |
571 | 571 | "can_reason": true, |
572 | 572 | "reasoning_levels": [ |
573 | 573 | "low", |
|
601 | 601 | "cost_per_1m_in": 0.27, |
602 | 602 | "cost_per_1m_out": 0.41, |
603 | 603 | "cost_per_1m_in_cached": 0, |
604 | | - "cost_per_1m_out_cached": 0.27, |
| 604 | + "cost_per_1m_out_cached": 0, |
605 | 605 | "context_window": 163840, |
606 | | - "default_max_tokens": 81920, |
| 606 | + "default_max_tokens": 32768, |
607 | 607 | "can_reason": true, |
608 | 608 | "reasoning_levels": [ |
609 | 609 | "low", |
|
939 | 939 | "cost_per_1m_in_cached": 0, |
940 | 940 | "cost_per_1m_out_cached": 0, |
941 | 941 | "context_window": 131072, |
942 | | - "default_max_tokens": 13107, |
| 942 | + "default_max_tokens": 8192, |
943 | 943 | "can_reason": false, |
944 | 944 | "supports_attachments": false |
945 | 945 | }, |
946 | 946 | { |
947 | 947 | "id": "meta-llama/llama-3.3-70b-instruct", |
948 | 948 | "name": "Meta: Llama 3.3 70B Instruct", |
949 | | - "cost_per_1m_in": 0.6, |
950 | | - "cost_per_1m_out": 0.6, |
| 949 | + "cost_per_1m_in": 0.25, |
| 950 | + "cost_per_1m_out": 0.75, |
951 | 951 | "cost_per_1m_in_cached": 0, |
952 | | - "cost_per_1m_out_cached": 0.6, |
| 952 | + "cost_per_1m_out_cached": 0, |
953 | 953 | "context_window": 131072, |
954 | | - "default_max_tokens": 65536, |
| 954 | + "default_max_tokens": 13107, |
955 | 955 | "can_reason": false, |
956 | 956 | "supports_attachments": false |
957 | 957 | }, |
|
967 | 967 | "can_reason": false, |
968 | 968 | "supports_attachments": false |
969 | 969 | }, |
| 970 | + { |
| 971 | + "id": "meta-llama/llama-4-maverick", |
| 972 | + "name": "Meta: Llama 4 Maverick", |
| 973 | + "cost_per_1m_in": 0.35, |
| 974 | + "cost_per_1m_out": 1.15, |
| 975 | + "cost_per_1m_in_cached": 0, |
| 976 | + "cost_per_1m_out_cached": 0, |
| 977 | + "context_window": 524288, |
| 978 | + "default_max_tokens": 4096, |
| 979 | + "can_reason": false, |
| 980 | + "supports_attachments": true |
| 981 | + }, |
970 | 982 | { |
971 | 983 | "id": "meta-llama/llama-4-scout", |
972 | 984 | "name": "Meta: Llama 4 Scout", |
|
982 | 994 | { |
983 | 995 | "id": "minimax/minimax-m2", |
984 | 996 | "name": "MiniMax: MiniMax M2", |
985 | | - "cost_per_1m_in": 0.255, |
986 | | - "cost_per_1m_out": 1, |
| 997 | + "cost_per_1m_in": 0.3, |
| 998 | + "cost_per_1m_out": 1.2, |
987 | 999 | "cost_per_1m_in_cached": 0, |
988 | 1000 | "cost_per_1m_out_cached": 0.03, |
989 | | - "context_window": 196608, |
990 | | - "default_max_tokens": 98304, |
| 1001 | + "context_window": 204800, |
| 1002 | + "default_max_tokens": 65536, |
991 | 1003 | "can_reason": true, |
992 | 1004 | "reasoning_levels": [ |
993 | 1005 | "low", |
|
1168 | 1180 | { |
1169 | 1181 | "id": "mistralai/ministral-3b-2512", |
1170 | 1182 | "name": "Mistral: Ministral 3 3B 2512", |
1171 | | - "cost_per_1m_in": 0.35, |
1172 | | - "cost_per_1m_out": 0.35, |
| 1183 | + "cost_per_1m_in": 0.1, |
| 1184 | + "cost_per_1m_out": 0.1, |
1173 | 1185 | "cost_per_1m_in_cached": 0, |
1174 | | - "cost_per_1m_out_cached": 0, |
| 1186 | + "cost_per_1m_out_cached": 0.01, |
1175 | 1187 | "context_window": 131072, |
1176 | | - "default_max_tokens": 65536, |
| 1188 | + "default_max_tokens": 13107, |
1177 | 1189 | "can_reason": false, |
1178 | 1190 | "supports_attachments": true |
1179 | 1191 | }, |
1180 | 1192 | { |
1181 | 1193 | "id": "mistralai/ministral-8b-2512", |
1182 | 1194 | "name": "Mistral: Ministral 3 8B 2512", |
1183 | | - "cost_per_1m_in": 0.3, |
1184 | | - "cost_per_1m_out": 0.3, |
| 1195 | + "cost_per_1m_in": 0.15, |
| 1196 | + "cost_per_1m_out": 0.15, |
1185 | 1197 | "cost_per_1m_in_cached": 0, |
1186 | | - "cost_per_1m_out_cached": 0, |
| 1198 | + "cost_per_1m_out_cached": 0.015, |
1187 | 1199 | "context_window": 262144, |
1188 | | - "default_max_tokens": 131072, |
| 1200 | + "default_max_tokens": 26214, |
1189 | 1201 | "can_reason": false, |
1190 | 1202 | "supports_attachments": true |
1191 | 1203 | }, |
|
1384 | 1396 | { |
1385 | 1397 | "id": "moonshotai/kimi-k2.5", |
1386 | 1398 | "name": "MoonshotAI: Kimi K2.5", |
1387 | | - "cost_per_1m_in": 0.6, |
1388 | | - "cost_per_1m_out": 3, |
| 1399 | + "cost_per_1m_in": 0.75, |
| 1400 | + "cost_per_1m_out": 3.5, |
1389 | 1401 | "cost_per_1m_in_cached": 0, |
1390 | | - "cost_per_1m_out_cached": 0.1, |
| 1402 | + "cost_per_1m_out_cached": 0.4, |
1391 | 1403 | "context_window": 262144, |
1392 | | - "default_max_tokens": 26214, |
| 1404 | + "default_max_tokens": 131072, |
1393 | 1405 | "can_reason": true, |
1394 | 1406 | "reasoning_levels": [ |
1395 | 1407 | "low", |
|
2140 | 2152 | { |
2141 | 2153 | "id": "openai/gpt-oss-120b", |
2142 | 2154 | "name": "OpenAI: gpt-oss-120b", |
2143 | | - "cost_per_1m_in": 0.09, |
2144 | | - "cost_per_1m_out": 0.36, |
| 2155 | + "cost_per_1m_in": 0.15, |
| 2156 | + "cost_per_1m_out": 0.6, |
2145 | 2157 | "cost_per_1m_in_cached": 0, |
2146 | | - "cost_per_1m_out_cached": 0.045, |
| 2158 | + "cost_per_1m_out_cached": 0, |
2147 | 2159 | "context_window": 131072, |
2148 | | - "default_max_tokens": 32768, |
| 2160 | + "default_max_tokens": 13107, |
2149 | 2161 | "can_reason": true, |
2150 | 2162 | "reasoning_levels": [ |
2151 | 2163 | "low", |
|
2176 | 2188 | { |
2177 | 2189 | "id": "openai/gpt-oss-20b", |
2178 | 2190 | "name": "OpenAI: gpt-oss-20b", |
2179 | | - "cost_per_1m_in": 0.03, |
2180 | | - "cost_per_1m_out": 0.14, |
| 2191 | + "cost_per_1m_in": 0.045, |
| 2192 | + "cost_per_1m_out": 0.18, |
2181 | 2193 | "cost_per_1m_in_cached": 0, |
2182 | 2194 | "cost_per_1m_out_cached": 0, |
2183 | 2195 | "context_window": 131072, |
2184 | | - "default_max_tokens": 13107, |
| 2196 | + "default_max_tokens": 16384, |
2185 | 2197 | "can_reason": true, |
2186 | 2198 | "reasoning_levels": [ |
2187 | 2199 | "low", |
|
2626 | 2638 | { |
2627 | 2639 | "id": "qwen/qwen3-30b-a3b-thinking-2507", |
2628 | 2640 | "name": "Qwen: Qwen3 30B A3B Thinking 2507", |
2629 | | - "cost_per_1m_in": 0.09, |
| 2641 | + "cost_per_1m_in": 0.1, |
2630 | 2642 | "cost_per_1m_out": 0.3, |
2631 | 2643 | "cost_per_1m_in_cached": 0, |
2632 | 2644 | "cost_per_1m_out_cached": 0, |
2633 | 2645 | "context_window": 262144, |
2634 | | - "default_max_tokens": 65536, |
| 2646 | + "default_max_tokens": 26214, |
2635 | 2647 | "can_reason": true, |
2636 | 2648 | "reasoning_levels": [ |
2637 | 2649 | "low", |
|
2644 | 2656 | { |
2645 | 2657 | "id": "qwen/qwen3-32b", |
2646 | 2658 | "name": "Qwen: Qwen3 32B", |
2647 | | - "cost_per_1m_in": 0.104, |
2648 | | - "cost_per_1m_out": 0.416, |
| 2659 | + "cost_per_1m_in": 0.14, |
| 2660 | + "cost_per_1m_out": 0.57, |
2649 | 2661 | "cost_per_1m_in_cached": 0, |
2650 | 2662 | "cost_per_1m_out_cached": 0, |
2651 | 2663 | "context_window": 131072, |
2652 | | - "default_max_tokens": 4096, |
| 2664 | + "default_max_tokens": 65536, |
2653 | 2665 | "can_reason": true, |
2654 | 2666 | "reasoning_levels": [ |
2655 | 2667 | "low", |
|
2680 | 2692 | { |
2681 | 2693 | "id": "qwen/qwen3-coder-30b-a3b-instruct", |
2682 | 2694 | "name": "Qwen: Qwen3 Coder 30B A3B Instruct", |
2683 | | - "cost_per_1m_in": 0.07, |
2684 | | - "cost_per_1m_out": 0.28, |
| 2695 | + "cost_per_1m_in": 0.1, |
| 2696 | + "cost_per_1m_out": 0.3, |
2685 | 2697 | "cost_per_1m_in_cached": 0, |
2686 | 2698 | "cost_per_1m_out_cached": 0, |
2687 | 2699 | "context_window": 262144, |
2688 | | - "default_max_tokens": 131072, |
| 2700 | + "default_max_tokens": 26214, |
2689 | 2701 | "can_reason": false, |
2690 | 2702 | "supports_attachments": false |
2691 | 2703 | }, |
|
2782 | 2794 | { |
2783 | 2795 | "id": "qwen/qwen3-next-80b-a3b-instruct", |
2784 | 2796 | "name": "Qwen: Qwen3 Next 80B A3B Instruct", |
2785 | | - "cost_per_1m_in": 0.1, |
| 2797 | + "cost_per_1m_in": 0.09, |
2786 | 2798 | "cost_per_1m_out": 1.1, |
2787 | 2799 | "cost_per_1m_in_cached": 0, |
2788 | 2800 | "cost_per_1m_out_cached": 0, |
2789 | 2801 | "context_window": 262144, |
2790 | | - "default_max_tokens": 131072, |
| 2802 | + "default_max_tokens": 26214, |
2791 | 2803 | "can_reason": false, |
2792 | 2804 | "supports_attachments": false |
2793 | 2805 | }, |
|
2806 | 2818 | { |
2807 | 2819 | "id": "qwen/qwen3-next-80b-a3b-thinking", |
2808 | 2820 | "name": "Qwen: Qwen3 Next 80B A3B Thinking", |
2809 | | - "cost_per_1m_in": 0.15, |
2810 | | - "cost_per_1m_out": 1.5, |
| 2821 | + "cost_per_1m_in": 0.0975, |
| 2822 | + "cost_per_1m_out": 0.78, |
2811 | 2823 | "cost_per_1m_in_cached": 0, |
2812 | 2824 | "cost_per_1m_out_cached": 0, |
2813 | 2825 | "context_window": 131072, |
|
2980 | 2992 | { |
2981 | 2993 | "id": "qwen/qwen3.5-27b", |
2982 | 2994 | "name": "Qwen: Qwen3.5-27B", |
2983 | | - "cost_per_1m_in": 0.195, |
2984 | | - "cost_per_1m_out": 1.56, |
| 2995 | + "cost_per_1m_in": 0.27, |
| 2996 | + "cost_per_1m_out": 2.16, |
2985 | 2997 | "cost_per_1m_in_cached": 0, |
2986 | | - "cost_per_1m_out_cached": 0, |
| 2998 | + "cost_per_1m_out_cached": 0.27, |
2987 | 2999 | "context_window": 262144, |
2988 | 3000 | "default_max_tokens": 32768, |
2989 | 3001 | "can_reason": true, |
|
3103 | 3115 | "cost_per_1m_in": 0.1, |
3104 | 3116 | "cost_per_1m_out": 0.3, |
3105 | 3117 | "cost_per_1m_in_cached": 0, |
3106 | | - "cost_per_1m_out_cached": 0, |
3107 | | - "context_window": 262144, |
3108 | | - "default_max_tokens": 32768, |
| 3118 | + "cost_per_1m_out_cached": 0.02, |
| 3119 | + "context_window": 256000, |
| 3120 | + "default_max_tokens": 128000, |
3109 | 3121 | "can_reason": true, |
3110 | 3122 | "reasoning_levels": [ |
3111 | 3123 | "low", |
|
3217 | 3229 | "cost_per_1m_in": 0.1, |
3218 | 3230 | "cost_per_1m_out": 0.3, |
3219 | 3231 | "cost_per_1m_in_cached": 0, |
3220 | | - "cost_per_1m_out_cached": 0.1, |
| 3232 | + "cost_per_1m_out_cached": 0.01, |
3221 | 3233 | "context_window": 262144, |
3222 | | - "default_max_tokens": 131072, |
| 3234 | + "default_max_tokens": 32768, |
3223 | 3235 | "can_reason": true, |
3224 | 3236 | "reasoning_levels": [ |
3225 | 3237 | "low", |
|
3298 | 3310 | { |
3299 | 3311 | "id": "z-ai/glm-4.5-air", |
3300 | 3312 | "name": "Z.ai: GLM 4.5 Air", |
3301 | | - "cost_per_1m_in": 0.13, |
3302 | | - "cost_per_1m_out": 0.85, |
| 3313 | + "cost_per_1m_in": 0.14, |
| 3314 | + "cost_per_1m_out": 0.86, |
3303 | 3315 | "cost_per_1m_in_cached": 0, |
3304 | | - "cost_per_1m_out_cached": 0.025, |
| 3316 | + "cost_per_1m_out_cached": 0, |
3305 | 3317 | "context_window": 131072, |
3306 | | - "default_max_tokens": 49152, |
| 3318 | + "default_max_tokens": 65536, |
3307 | 3319 | "can_reason": true, |
3308 | 3320 | "reasoning_levels": [ |
3309 | 3321 | "low", |
|
3406 | 3418 | { |
3407 | 3419 | "id": "z-ai/glm-4.7-flash", |
3408 | 3420 | "name": "Z.ai: GLM 4.7 Flash", |
3409 | | - "cost_per_1m_in": 0.1, |
3410 | | - "cost_per_1m_out": 0.43, |
| 3421 | + "cost_per_1m_in": 0.06, |
| 3422 | + "cost_per_1m_out": 0.4, |
3411 | 3423 | "cost_per_1m_in_cached": 0, |
3412 | | - "cost_per_1m_out_cached": 0, |
| 3424 | + "cost_per_1m_out_cached": 0.01, |
3413 | 3425 | "context_window": 202752, |
3414 | | - "default_max_tokens": 101376, |
| 3426 | + "default_max_tokens": 20275, |
3415 | 3427 | "can_reason": true, |
3416 | 3428 | "reasoning_levels": [ |
3417 | 3429 | "low", |
|
0 commit comments