← Back to context

Comment by mirekrusin

1 day ago

MoE is something different - it's a technique to activate just a small subset of parameters during inference.

Whatever is good enough now, can be much better for the same cost (time, computation, actual cost). People will always choose better over worse.

Thanks, I wasn't aware of that. Still - why isn't there a super expensive OpenAI model that uses 1,000 experts and comes up with way better answers? Technically that would be possible to build today. I imagine it just doesn't deliver dramatically better results.

  • That's what GPT-5 Pro and Grok 4 Heavy do. Those are the ones you pay triple digit USD a month for.