Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Pricing seems good, but the open question is still on tool calling reliability.

Input: $1.25 / 1M tokens Cached: $0.125 / 1M tokens Output: $10 / 1M tokens

With 74.9% on SWE-bench, this inches out Claude Opus 4.1 at 74.5%, but at a much cheaper cost.

For context, Claude Opus 4.1 is $15 / 1M input tokens and $75 / 1M output tokens.

> "GPT-5 will scaffold the app, write files, install dependencies as needed, and show a live preview. This is the go-to solution for developers who want to bootstrap apps or add features quickly." [0]

Since Claude Code launched, OpenAI has been behind. Maybe the RL on tool calling is good enough to be competitive now?

[0]https://github.com/openai/gpt-5-coding-examples



And they included Flex pricing, which is 50% cheaper if you're willing to wait for the reply during periods of high load. But great pricing for agentic use with that cached token pricing, Flex or not.


I switched immediately because of pricing, input token heavy load, but it doesn't even work. For some reason they completely broke the already amateurish API.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: