Change default model to Gemini #425
darkogj
started this conversation in
1. Feature requests
Replies: 3 comments
-
I'm down to do this! If we do this we should make a good blog post with those arguments in it. |
Beta Was this translation helpful? Give feedback.
0 replies
-
Looking at open router Gemini 2.5 has been a bit unstable this week and latency is climbing. I'd like to hold off on making it the default for a week or so to see if it stabilizes. |
Beta Was this translation helpful? Give feedback.
0 replies
-
With Claude 4 released we might want to consider this instead |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Should we consider switching people to Gemini 2.5 as the default model?
Has been number 1 on the daily/weekly list on OpenRouter: https://openrouter.ai/rankings/programming?view=day
Ranking higher than Claude on all major LLM coding leaderboards: https://livebench.ai/#/ https://aider.chat/docs/leaderboards/ https://dubesor.de/benchtable.html
Qualitative evidence it's better than Claude on /r/chatgptcoding
People are using Gemini 2.5 Pro for our free tier most of the time from usage data 2 days ago (not sure how things are now but we're ranking number 3 on "apps using Gemini 2.5 Pro": https://openrouter.ai/google/gemini-2.5-pro-preview
People can get 3x out of it using the$20 free tier: Most $ $$ goes to input tokens, and Claude charges $3 per 1 million tokens for Sonnet. Gemini charges $1.25 (<250k tokens) and $2.5 (>250k tokens). Considering the fact people are starting small when using new extensions in general, it's expected that prob. 90% of the costs will go to this tier.
To summarize: New users get a better experience and get to use our extension for longer if we switch to Gemini as the default model. And $20 with Gemini translates to almost 3x more token usage vs. Claude 3.7 (as I said previously, the vast majority of tokens burned are input tokens).
Beta Was this translation helpful? Give feedback.
All reactions