← Back to context

Comment by energy123

1 month ago

I don't know when but I'm going off:

- "OpenAI is partnering with Cerebras to add 750MW of ultra low-latency AI compute"

- Sam Altman saying that users want faster inference more than lower cost in his interview.

- My understanding that many tasks are serial in nature.

Speed is really important to me but also I would like higher weekly limits -- which means lower cost I suppose. Building out complex projects can take 6 months to a year on a Pro plan.

  • Same experience with Pro.

    My trick is to attach the codebase as a txt file to 5-10 different GPT 5.2 Thinking chats, paste in the specs, and then get hard work done there, then just copy paste the final task list into codex to lower codex usage.