Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

A big wow moment coming up is going to be GPT 5.* in Codex with Cerebras doing inference. The inference speed is going to be a big unlock, because many tasks are intrinsically serial.

It's going to feel literally like playing God, where you type in what you want and it happens ~instantly.



When?


I don't know when but I'm going off:

- "OpenAI is partnering with Cerebras to add 750MW of ultra low-latency AI compute"

- Sam Altman saying that users want faster inference more than lower cost in his interview.

- My understanding that many tasks are serial in nature.


Speed is really important to me but also I would like higher weekly limits -- which means lower cost I suppose. Building out complex projects can take 6 months to a year on a Pro plan.


Same experience with Pro.

My trick is to attach the codebase as a txt file to 5-10 different GPT 5.2 Thinking chats, paste in the specs, and then get hard work done there, then just copy paste the final task list into codex to lower codex usage.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: