← Back to context

Comment by tornikeo

13 hours ago

On paper. There's huge financial incentive to quantize the crap out of a good model to save cash after you've hooked in subscriptions.

And there’s an incentive to publish evidence of this to discourage it, do you have any?

  • Models aren't just big bags of floats you imagine them to be. Those bags are there, but there's a whole layer of runtimes, caches, timers, load balancers, classifiers/sanitizers, etc. around them, all of which have tunable parameters that affect the user-perceptible output.

  • Anybody with more than five years in the tech industry has seen this done in all domains time and again. What evidence you have AI is different, which is the extraordinary claim in this case...