Hacker News new | past | comments | ask | show | jobs | submit login

The simple answer is: I don't care. I'll statistically figure out what the typical total cost per call is from experience, and that's what matters. Who cares if they lie about it, if the model's cost per call fits my budget?

If it starts costing $1 per call, and that's too high, then I just won't use it commercially. Whether it was $1 because they inflated the token count or because it just actually took a lot of tokens to do its reasoning isn't really material to my economic decision.




The thing is it might increase in cost after you've decided to use it commercially, and have invested a lot of time and resources in it. Now it's very hard to move to something else, but very easy for OpenAI to increase your cost arbitrarily. The statistics you made are not binding for them.


The API returns how many tokens were used in reasoning, so it would be easy to see any average change in reasoning token consumption. And token prices in general have been extremely deflationary over the last 18 months.


This is experimental, frontier stuff, obviously it comes with risks. Building on GPT-4 in March of 2023 was like that as well, but now you can easily switch between a few models of comparable quality made by different companies (yay capitalism and free markets!). You can risk and use just released stuff right now, or, most likely, come back in 6-12 months (probably earlier) and get several different providers with very similar APIs.


Everything that OpenAI does with LLMs has already been done and validated in the open source community well before OpenAI gets around to it. OpenAI is not an innovator. simbianai/taskgen on github is an example of one such project, although there are others too that don't come to mind right now.

As such, I would never call their work "frontier stuff", but they do bring it to the masses with their commercial service.


Same applies to every other API in the world, yes.


No, S3 pricing for example is predictable, and written in a contract. There's no way for AWS to charge you 3x amount of dollars for 1GB tomorrow. They need to announce it in advance, and give you time to exit the contract if you disagree with the new price. It's really not the same. OpenAI can just tell you your prompt from tomorrow used up 20x times reasoning tokens. There's no advance warning or predictability. I really don't understand how you can claim the situations are identical.




Consider applying for YC's Summer 2025 batch! Applications are open till May 13

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: