Comment by layer8

1 day ago

Sure, but we could have Hetzners and OVHs who just provide the compute for whatever model we want to run.

Checked the DDR5 price lately?

  • I didn’t claim that it would be cheap. But I’d rather see the real cost of SOTA LLM use exposed. On the other hand, reportedly SOTA LLM inference is profitable nowadays, so it can’t be that expensive.