Everyone keeps saying that but I’ve found it to be incredibly weak in the real world every single time I’ve reached for it. I think it’s benchmaxxed to an extent.
Why would someone want to spend half a million dollars on GPUs and components (if not more) to run one year old models that genuinely aren't useful? You can't self host trillion parameter models unless you own a datacenter lol (or want to just light money on fire).
They'd only do that if they were some kind of open ai company /s
gpt-oss is pretty great tbh - one of the better all-around local models for knowledge and grounding.
Everyone keeps saying that but I’ve found it to be incredibly weak in the real world every single time I’ve reached for it. I think it’s benchmaxxed to an extent.
lol :)
Why would someone want to spend half a million dollars on GPUs and components (if not more) to run one year old models that genuinely aren't useful? You can't self host trillion parameter models unless you own a datacenter lol (or want to just light money on fire).
Are the mini / omni models really trillion parameter models?
I don't think so, but you're still looking at a giant investment that can't really be justified for their capability.
1 reply →
To do AI research!!!!!!!