Comment by segmondy
1 day ago
probably because maybe 1 or 2 folks on here can run it? It's 1000B model, if 16bit training then you need 2000b of GPU vram to run it. Or about 80 5090s hooked up to the same machine. Or 20 of them to run it in Q2.
1 day ago
probably because maybe 1 or 2 folks on here can run it? It's 1000B model, if 16bit training then you need 2000b of GPU vram to run it. Or about 80 5090s hooked up to the same machine. Or 20 of them to run it in Q2.
Still more accessible then fully closed models.
...but not widespread enough impact/interest to reach/stay on the HN-frontpage.
Infinity minus one is smaller than infinity?
I think there are alot, like alot of alot, of shills pumping up hype around commercial SaaS models. Deepseek was ignored for like forever until it became embarrassing.
I even made a post trying to break the surreal silence.
https://news.ycombinator.com/item?id=38505986