This article is machine translated
Show original
Chatting with friends who aren't developers and have started tinkering with AI, I realized why they spend so many tokens and burn through the limit so quickly:
- They use everything in the same chat, so the chat keeps growing and growing, and you're sending it more and more context.
When planning a project, you have to break it down into smaller functions. Each function will be a development in itself. For example, in OpenCode, you use `/new` to start a new session. OpenCode itself is smart enough to know where to look in the current directory.
I saw some people using Cursor and spending 900k to 3M tokens per request. It's insane! My context windows didn't exceed 150k tokens across the entire chat. Always think of each chat as a feature, not as a project. Not only will you have more tokens for development, but the model will become increasingly unreliable as the number of tokens increases beyond a certain threshold.
From Twitter
Disclaimer: The content above is only the author's opinion which does not represent any position of Followin, and is not intended as, and shall not be understood or construed as, investment advice from Followin.
Like
Add to Favorites
Comments
Share
Relevant content





