(Caveat on this comment: I'm the COO of OpenRouter. I'm not here to plug my employer; just ran across this and think this suggestion is helpful)
Feel free to give OpenRouter a try; part of the value prop is that you purchase credits and they are fungible across whatever models & providers you want. We just got Sonnet 4 live. We have a chatroom on the website, that simply uses the API under the covers (and deducts credits). Don't have passkeys yet, but a good handful of auth methods that hopefully work.
Just wanted to provide some (hopefully) helpful feedback from a potential customer that likely would have been, but bounced away due to ambiguity around pricing.
It's too hard to find out what markup y'all charge on top of the APIs. I understand it varies based on the model, but this page (which is what clicking on the "Pricing" link from the website takes you to) https://openrouter.ai/models is way too complicated. My immediate reaction is, "oh shit, this is made for huge enterprises, not for me" followed immediately by "this isn't going to be cheap, I'm not even going to bother." We're building out some AI features in our products so the timing is otherwise pretty good. We're not big fish, but do expect to spending between $3,000 and $5,000 per month once the features hit general availability, so we're not small either. If things go well of course, we'd love to 10x that in the next few years (but time will tell on that one of course).
> We pass through the pricing of the underlying providers; there is no markup on inference pricing (however we do charge a fee when purchasing credits).
Currently, there is a terrible regression UI bug in OpenRouter (at least on Firefor MacOS). Previously, while the LLM was generating the answer I could scroll up to the top of the answer and start reading.
For the past couple of weeks, it keeps force scrolling me down to the bottom as new words come in. I can't start reading till the whole answer is generated. Please fix.
(Caveat on this comment: I'm the COO of OpenRouter. I'm not here to plug my employer; just ran across this and think this suggestion is helpful)
Feel free to give OpenRouter a try; part of the value prop is that you purchase credits and they are fungible across whatever models & providers you want. We just got Sonnet 4 live. We have a chatroom on the website, that simply uses the API under the covers (and deducts credits). Don't have passkeys yet, but a good handful of auth methods that hopefully work.
Just wanted to provide some (hopefully) helpful feedback from a potential customer that likely would have been, but bounced away due to ambiguity around pricing.
It's too hard to find out what markup y'all charge on top of the APIs. I understand it varies based on the model, but this page (which is what clicking on the "Pricing" link from the website takes you to) https://openrouter.ai/models is way too complicated. My immediate reaction is, "oh shit, this is made for huge enterprises, not for me" followed immediately by "this isn't going to be cheap, I'm not even going to bother." We're building out some AI features in our products so the timing is otherwise pretty good. We're not big fish, but do expect to spending between $3,000 and $5,000 per month once the features hit general availability, so we're not small either. If things go well of course, we'd love to 10x that in the next few years (but time will tell on that one of course).
From https://openrouter.ai/docs/faq#how-do-i-get-billed-for-my-us...
> We pass through the pricing of the underlying providers; there is no markup on inference pricing (however we do charge a fee when purchasing credits).
1 reply →
Currently, there is a terrible regression UI bug in OpenRouter (at least on Firefor MacOS). Previously, while the LLM was generating the answer I could scroll up to the top of the answer and start reading.
For the past couple of weeks, it keeps force scrolling me down to the bottom as new words come in. I can't start reading till the whole answer is generated. Please fix.
Looks good, thanx for the suggestion!
I agree about the email login. It is huge barrier and it is why I use other providers for chat over Claude. I still use their API though.