Speed is really important to me but also I would like higher weekly limits -- which means lower cost I suppose. Building out complex projects can take 6 months to a year on a Pro plan.
My trick is to attach the codebase as a txt file to 5-10 different GPT 5.2 Thinking chats, paste in the specs, and then get hard work done there, then just copy paste the final task list into codex to lower codex usage.
I don't know when but I'm going off:
- "OpenAI is partnering with Cerebras to add 750MW of ultra low-latency AI compute"
- Sam Altman saying that users want faster inference more than lower cost in his interview.
- My understanding that many tasks are serial in nature.
Speed is really important to me but also I would like higher weekly limits -- which means lower cost I suppose. Building out complex projects can take 6 months to a year on a Pro plan.
Same experience with Pro.
My trick is to attach the codebase as a txt file to 5-10 different GPT 5.2 Thinking chats, paste in the specs, and then get hard work done there, then just copy paste the final task list into codex to lower codex usage.