I think it's obvious that they are critically lacking in compute capacity especially since OpenAI has committed billions to locking up all the future compute production.
And I don't necessarily think it's wrong for Anthropic to introduce QoS or throttling on users of their models. It's pretty much a necessity when offering public access to a scarce resource and it's been a common practice for decades.
What is the alternative? We just accept that it doesn't work half the time because the system is overloaded with molt bots?
I agree. If compute is the issue and pricing can't budge then something has to give.
They would have kept my business if they were honest and upfront. Instead they sold me something that worked well, broke it without warning, remained silent about it until enough people caught on, chose to do nothing, then proceeded to release a model that eats ~30% more tokens with no advantage over prior models.
If they chose to unbrick their model and offered what we had a couple months ago at a 50% hike, I would have been onboard. I've seen enough now of how this company treats its customers to continue using or recommending them.
Also, Codex works much better than CC now for anyone who happens to be on the fence.
Codex actually feels severely lacking to me, trying to switch off Claude Code. I'm trying every day. The models honestly feel on par, but the Harness and the CLI are somewhat painful to work with.
The alternative is to price their product transparently. If there is too much demand and supply is limited: Charge more.
Anthropic wants to have their lunch (low apparent prices, increased market share) and eat it too (controlled costs, adequate production to serve the demand).
They're advertising themselves as a $5 All-You-Can-Eat buffet, but then aggressively and arbitrarily restricting admission, sneakily swapping out the high-quality ingredients for garbage-tier slop, and kicking out anyone who even utters the words "to go box" or "doggie bag".
I cancelled my subscription so not really defending them myself but if all of their customers were humans who used it normally I bet they could serve everyone. It's when someone presses a few keys walks away and a bot uses tokens for 72 hours straight that it becomes a problem. Then people buy 3 accounts and do that for weeks at a time.
Could you do that as a human? Sure but you'd likely burn out after a couple of weeks. Also the human would probably use those tokens far more effectively and would not need as many. It's feels the same as someone installing a crypto miner on their servers in my mind. Abhorrent behavior.
I've been very content moving away from OneDrive/GDrive to a personal NAS setup with Synology/Ugreen. You can access a shared drive/photo drive and use Tailscale to mount your volume from anywhere.
I've also configured encrypted cloud backups to a different geographic region and off-site backups to a friend's NAS (following the 3-2-1 backup rule). It does help having 2.5Gb networking as well, but owning your data is more important in the coming age of sloppy/degrading infrastructure and ransomware attacks.
It is, but if you are the kind of person these exploits are likely to target, you should have it on. So far there have been no known exploits that work in Lockdown Mode.
> if you are the kind of person these exploits are likely to target, you should have it on
You can also selectively turn it on in high-risk settings. I do so when I travel abroad or go through a border. (Haven't started doing it yet with TSA domestically. Let's see how the ICE fiasco evolves.)
For entering the US you want to fully wipe your phone first. Lockdown mode is useless since they will just hold you in a basement until you unlock the phone for them to clone.
I'd wager that b2c projects former VP of Product at Instagram & CPO at OpenAI, Kevin Weil, may have championed are getting the boot with the company refocusing on making money under the stewardship of Fidji Simo: https://www.businessinsider.com/fidji-simo-openai-product-re...
>repeated rounds of cuts are destructive to morale, to focus, and to the trust [...] i'd rather take a hard, clear action now [...] than manage a slow reduction of people toward the same outcome
I think this is pretty agreeable, spanning layoffs into a monthly/quarterly "Hunger Games" is very damaging to employee morale.
Selecting your own mute words for your timeline is the best part about Twitter. The algorithm changes have been pretty bad in the last few weeks though.
FYI the “mute by keyword” feature exists both on Bluesky and Mastodon and I use it extensively. I don’t use Threads but a quick search tells me it’s available there too.
Now you know, and you don’t need to use twitter anymore!
It's known that Anthropic's $20 Pro subscription is a gateway plan to their $100 Max subscription, since you'll easily burn your token rate on a single prompt or two. Meanwhile, I've had ample usage testing out Codex on the basic $20 ChatGPT Plus plan without a problem.
As for Anthropic's $100 Max subscription, it's almost always better to start new sessions for tasks since a long conversation will burn your 5-hour usage limit with just a few prompts (assuming they read many files). It's also best to start planning first with Claude, providing line numbers and exact file paths prior, and drilling down the requirements before you start any implementation.
> It's known that Anthropic's $20 Pro subscription is a gateway plan to their $100 Max subscription, since you'll easily burn your token rate on a single prompt or two.
I genuinely have no idea what people mean when I read this kind of thing. Are you abusing the word "prompt" to mean "conversation"? Or are you providing a huge prompt that is meant to spawn 10 subagents and write multiple new full-stack features in one go?
For most users, the $20 Pro subscription, when used with Opus, does not hit the 5-hour limit on "a single prompt or two", i.e. 1-2 user messages.
Today I literally gave Claude a single prompt, asking it to make a plan to implement a relatively simple feature that spanned a couple
different codebases. It churned for a long time, I asked a couple very simple
follow up questions, and then I was out of tokens. I do not consider myself to be any kind of power user at all.
The only time I've ever seen this happen is when you give it a massive codebase, without any meaningful CLAUDE.md to help make sense of it and no explicitly @ mentioning of files/folders to guide, and then ask it for something with huge cross-cutting.
> spanned a couple different codebases
There you go.
If you're looking to prevent this issue I really recommend you set up a number of AGENTS.md files, at least top-level and potentially nested ones for huge, sprawling subfolders. As well as @ mentioning the most relevant 2-3 things, even if it's folder level rather than file.
Not just for Claude, it greatly increases speed and reduces context rot for any model if they have to search less and more quickly understand where things live and how they work together.
I have a tool that scans all code files in a repo and prints the symbols (AST based), it makes orienting around easy, it can be scoped to a file or folder.
I am on $100 max subscription, and I rarely hit the limit, I used to but not anymore, but then again, I stopped building two products at the same time and concentrate to finish up the first/"easiest" one.
> you'll easily burn your token rate on a single prompt or two
My experience has been that I can usually work for a few hours before hitting a rate limit on the $20 subscription. My work time does not frequently overlap with core business hours in PDT, however. I wonder whether there is an aspect of this that is based on real-time dynamic usage.
I’m not surprised they closed the loophole, it always felt a little hacky using an Anthropic monthly sub as an API with a spoofed prompt (“You are Claude Code, Anthropic's official CLI for Claude”) with OpenCode.
Google will probably close off their Antigravity models to 3P tools as well.
reply