Comment by mike_hearn
5 hours ago
The plan seems to be for lots and lots of smaller satellites.
For inferencing it can work well. One satellite could contain a handful of CPUs and do batch inferencing of even very large models, perhaps in the beginning at low speeds. Currently most AI workloads are interactive but I can't see that staying true for long, as things improve and they can be trusted to work independently for longer it makes more sense to just queue stuff up and not worry about exactly how high your TTFT is.
For training I don't see it today. In future maybe. But then, most AI workloads in future should be inferencing not training anyway.
No comments yet
Contribute on Hacker News ↗