Comment by sigmoid10

2 days ago

GPT-4 came out 3 years ago and you can run comparable models for 1% of the cost nowadays. That is not 2x efficiency. That's two orders of magnitude in end-to-end compute efficiency.

you're looking at nearly the entire curve of the tech's development. that's like saying lightbulbs became 99% more energy efficient and therefore will become another 99% more energy efficient. but most techs follow an S curve.

  • >you're looking at nearly the entire curve of the tech's development

    That's a pretty strong statement that would need some data or at least a mathematical argument to back it up. Otherwise it's like saying in the 1980s that PCs with 640kB RAM have reached their pinnacle in terms of what users can expect in real life benefits and there's no reason to keep pushing the tech.

    • *entire curve to-date (I should have clarified). Yes it will get better for a long time, but where we are on the curve is harder to say. Lots of metrics to choose from, like "well it's incorrect 90% less often than a year ago, so that's a 10x improvement!". But the real metric that matters is how useful it is to people, and based on user data it looks like the only area it's getting exponentially more useful YoY is for programming. Lot of coders using it 10x more than before to code 10x faster. Not sure any other profession uses it for more than a juiced-up search engine / proofreader.

      1 reply →

How do we know how much it costs? Or is this just based off the token pricing?

  • That's the bingo of the question... The entire argument is token pricing, which can be subsidized.