← Back to context

Comment by brunooliv

17 hours ago

[flagged]

How is any of what you wrote relevant? People aren't using Claude for the first time and hitting rate limits. They've been using Claude for months, at the very least, and they're hitting rate limits without significant changes to how they prompt.

> People need to understand a few things: vague questions make the models roam endlessly “exploring” dead ends.

> If people were considerably more willing to aggressively prune their context and scope tasks well, they could get a lot more done with it

If this were the problem, people would've encountered this when they started using Claude. The problem is not that they can't get anything done. It's being able to get things done for months, but suddenly hitting rate limits way too easily and response quality being clearly degraded, so they can't get things done that used to be possible.

  • I think in this case, we probably have different experiences that shape how we see some things differently: I see many (very smart) people doing certain things that are not optimal (eg: copy-paste entire files instead of referencing them or tell claude at every message to "read CLAUDE.md and follow its instructions precisely") which can lead to a lot of token waste. If certain system prompts were tweaked internally or some models now read more files than before, keeping these "inneficient prompts" will make limits exhaust faster. Sub-agents or this new agent teams feature didn't exist until a few months ago: that alone eats A LOT of tokens, not intended for this pre-paid API usage, etc.

    The ecosystem is evolving super quickly so, our own experiences and workflows must keep adapting with it to experiment, find limitations and arrive at the "tightest possible scope" that still allows you to get things done, because it is possible.

    Another example: pre-paid monthly subscription aggregates usage towards web and Claude Code, for eg. So if you're checking for holiday itineraries over your lunch break, then decide to sit down and ask a team of agents to refactor a giant codebase with hundreds or thousands of files, context will be exhuasted quickly, etc, etc.

    I see this "context economy" as a new way of managing your "mental models": every token counts, and every token must bear its weight for the task at hand, otherwise, I'm "wasting budget". I am also still learning how to operate in this new way of doing things, and, while there have been genuine issues with Claude Code, not every single issue that people encounter is an upstream problem.

    • This is literally victim blaming. When people haven't been having issues until now, why is it their fault? Anthropic is providing a paid service to paying users. It's not acceptable that they degrade our experience to save some money and it's not acceptable to blame everybody else who didn't cause the issue.

      1 reply →

This is a copypasta right? I'm damn confident I have read the same content before.

> Anthropic CAN change their limits and rates as they see fit, there’s never been hard promises or SLOs on these plans.

No they can't. When I buy an annual subscription and prepay for the year, they can't just go "ok now you get one token a month" a day in. I bought the plan as I bought it. They can't change anything until the next renewal.

  • Agreed. I wouldn't have bought an annual subscription under the current conditions.

  • That probably is somewhere in the EULA or other contract you agreed to. I'm not arguing it's any kind of fair, nor am I a lawyer so IDK if it's enforceable, but I bet it's in there somewhere.

    • Well yes, that's how they get you, but at some point the law will have to change to become fair.

  • > I bought the plan as I bought it. They can't change anything until the next renewal.

    So no new models, no new features?

    • That's up to them. I'd be fine to not get access to new models or features, which is why I'm fine to pay $XX to buy some desktop software and use it forever as-is.

      If they're selling me compute and bundling the features in, they better not go back on the compute I paid for.

  • It's the nature of SaaS software, right? It doesn't need to be an enforced "hard change", but, let's say that they trained Opus 4.6 to be more "verbose" or to explore more files to gain more context for it's own tasks.

    If your limits stay "the same", but you then use Opus 4.6, your quota will be exhausted much faster, it's just how it works.

    Note that some features are simply NOT made for these Pro, Max, Max 5x or whatever pre-paid plans. I'm pretty sure this is by design and not an accident or a bug: If you have 6/7 MCP servers configured or if you want to use this new feature of "Agent Teams", you will exhaust your entire quota before ANY work is even done. This is not a bug. Each agent has its own context window and tools and they all count separately.

    MCP servers, when active, add A LOT of context to your sessions before you even use them, etc, etc.

    It feels to me that people want to have their cake and eat it too, but, that would NOT be a sustainable business model. You can not complain about the tools if you can't understand them in-depth.

    I want to state that I don't think Anthropic are fully aware of the ramifications that ANY small change in ANY of their models might have, because their entire ecosystem is a bit messy atm, but, I'm certain they're aware that if people dont like it, they will cancel the subscription and flock to a competitor very quickly, since there's no real moat anymore. So, it's in their own interest to keep things minimally usable even on the "cheaper plans".

    I have seen people with 5-10 "active MCP servers" that they "wanted to try out" then they forget about it and wonder why their context is always full... Cmon... that's almost bad faith.

    I don't fully defend Anthropic as they've had several issues with degraded model quality after releasing "the latest model", and CLI usability that cost me real money and real tokens, so, there's a lot of room for improvement, but, to claim that quota gets exhausted after 1h it points out to either some forgotten MCP servers, skills or giant files being accidentally read in, or some sort of mis-use which these limits were put in place to prevent exactly.

    There's a very thin line between: quota is exhuasted on a regular, normal session after 1h and I think there's a bug versus I had 3-4 MCP servers active that I am not using at all but forgot to disable and my CLAUDE.md file is 1000 lines...

    • What you say makes sense, but they very actually reduced the token limits. We had, say, 20M tokens/week before, now we have 18M tokens/week (example numbers). They didn't just make a model that eats tokens faster.

      4 replies →

They rolled out 1M context then they start doing this shit? I know Pro doesn't have access to the 1M context but what a joke.