The pricing and usage of tokens hugely changed recently. To test I gave a simple task to Cursor which would use a 192 line swift code, including comments. This file has 1213 tokens according to https://platform.openai.com/tokenizer .
Here is my prompt: "There is a bug when i click on a raw it plays the audio and it deletes it afterwards, even though without clicking the delete button."
The Cursor dashboard says it used 269,738 tokens, and records to usage accordingly. I am a pro plan user.
Is there anyone else having similar issues?
Most of the tokens are cache read and cache write. It's possible that it's searching the whole codebase with these. My guess is they're embeddings tokens which are substantially cheaper.
Your 1213 should be somewhere in the input column. Probably less because of the caching. This is the magic where your $20 goes for more than $20 worth.
It would be even sadder if the rug pull was because of a bug they didn’t even realise was churning through tokens.