← Back to context

Comment by energy123

1 month ago

A big wow moment coming up is going to be GPT 5.* in Codex with Cerebras doing inference. The inference speed is going to be a big unlock, because many tasks are intrinsically serial.

It's going to feel literally like playing God, where you type in what you want and it happens ~instantly.

When?

  • I don't know when but I'm going off:

    - "OpenAI is partnering with Cerebras to add 750MW of ultra low-latency AI compute"

    - Sam Altman saying that users want faster inference more than lower cost in his interview.

    - My understanding that many tasks are serial in nature.

    • Speed is really important to me but also I would like higher weekly limits -- which means lower cost I suppose. Building out complex projects can take 6 months to a year on a Pro plan.

      1 reply →