Comment by simianwords 12 days ago ok but the capabilities are also rising. what point are you trying to make? 11 comments simianwords Reply oytis 12 days ago That it's not getting cheaper? jstummbillig 12 days ago But it is, capability adjusted, which is the only way it makes sense. You can definitely produce last years capability at a huge discount. simianwords 12 days ago you are wrong. https://epoch.ai/data-insights/llm-inference-price-trendsthis is accounting for the fact that more tokens are used. techpression 12 days ago The chart shows that they’re right though. Newer models cost more than older models. Sure they’re better but that’s moot if older models are not available or can’t solve the problem they’re tasked with. 7 replies →
oytis 12 days ago That it's not getting cheaper? jstummbillig 12 days ago But it is, capability adjusted, which is the only way it makes sense. You can definitely produce last years capability at a huge discount. simianwords 12 days ago you are wrong. https://epoch.ai/data-insights/llm-inference-price-trendsthis is accounting for the fact that more tokens are used. techpression 12 days ago The chart shows that they’re right though. Newer models cost more than older models. Sure they’re better but that’s moot if older models are not available or can’t solve the problem they’re tasked with. 7 replies →
jstummbillig 12 days ago But it is, capability adjusted, which is the only way it makes sense. You can definitely produce last years capability at a huge discount.
simianwords 12 days ago you are wrong. https://epoch.ai/data-insights/llm-inference-price-trendsthis is accounting for the fact that more tokens are used. techpression 12 days ago The chart shows that they’re right though. Newer models cost more than older models. Sure they’re better but that’s moot if older models are not available or can’t solve the problem they’re tasked with. 7 replies →
techpression 12 days ago The chart shows that they’re right though. Newer models cost more than older models. Sure they’re better but that’s moot if older models are not available or can’t solve the problem they’re tasked with. 7 replies →
That it's not getting cheaper?
But it is, capability adjusted, which is the only way it makes sense. You can definitely produce last years capability at a huge discount.
you are wrong. https://epoch.ai/data-insights/llm-inference-price-trends
this is accounting for the fact that more tokens are used.
The chart shows that they’re right though. Newer models cost more than older models. Sure they’re better but that’s moot if older models are not available or can’t solve the problem they’re tasked with.
7 replies →