This is a tough moment. Claude is simultaneously becoming substantially more expensive, substantially less reliable (single 9 of reliability), and substantially less performant. It's really hard to justify the cost of a subscription over there right now.
There was another thread where some people pointed out, Amazon will give you access to Claude with better uptime for the same price (per million tokens up / down), downside is, it does not have the native ability to browse the web, but maybe that's a hidden blessing, since it's less likely to read some random website that has prompt injection embedded into it.
For coding its fine, I havent experimented too much with Amazon Bedrock myself, but I just might soon to check for any limitations.
From an economics perspective, it makes sense to make it more expensive if you're having trouble keeping up with demand for a service. It'll be tough getting used to because it was so nice and cheap
On the other hand, it was somewhat expected that we would have a correction for the prices. Hopefully after this correction things will be more stable and we won't have to worry too much about future price increases
YMMV. I would still be very happy with Claude if it hard failed on 20% of tasks. You can always come back to it.
I say this as someone working for a tech company who does not have to foot the bill (in the >$1k per month bracket)
I also experienced and accept the 1990s levels of unreliability, which is my “internet generation”. My first access was lifting a handset and placing on a speaker/mic cradle.
Programmers these days are fucking spoiled. If it’s $220 worth of value for $200 - I get it. But I’m getting $100k of value for $10k and so I’ll put up with some shit.
Interestingly, yeah, I can see that this would really cut into your subscription usage with the 5 hour rate limit windows...
I am an API user, and while it being down is super annoying, it isn't really as big of a hit to my overall usage as I can just prepare a bunch of stuff to run in parallel when it does come back up.
Not to mention substantially less open. I've been using an OpenAI subscription in Pi Agent for a couple weeks now and it's great. And from what I can tell, 5.5 is a heck of a model.
Plus, they've dumbed down their models to the point where the value just isn't there like it was. If I have to go in and clean up after it, or constantly wrestle with it through prompts, what's the point? Just spending $200 a month to be frustrated at a machine.
Single nine has good vibes bro. It means when the service is up the results are better. I read about it in a blog. The model hallucinates way less. Even less than grok
It's lazy, does not take ownership and responsibility, wants to defer work, and I have to force it to check reality. It likes to guess and assume it's correct and I am wrong. Agents.md is not helping at all. It's in full enshittification phase, yay!
GitHub is a long running business with a mature software stack running into scaling issues while they move to Azure and becoming Microsoft-ified. Claude is a new company in a new market with an extremely fast growing userbase running relatively novel AI infrastructure with a business model they are still figuring out.
With the TPU deal with Google and their relationship with Amazon they will have access compute coming online.
I worked with 4.6 and found some improvements for better planning and sustained us, but agree some posters 4.7 is slower, overthinking.
What I expect is frontier models to get bigger and more expensive (especially fast mode like on Cerberus). And most of his get much smaller distillations for the more generous subscription tiers.
Anybody else double fist Codex/Claude? They both code, solve problems, and find bugs in unique ways. I find using both is more useful than using either alone. I have them code review each others work, it's great.
So happy to have diversified my model providers this past couple of weeks. GPT-5.5 has had no trouble slotting into Opus workloads. Will be fun to try out more of the models as time goes on to build some resiliency into my engineering workflows :).
What I found was that I *strongly* preferred Claude Code with its defaults. Codex was almost unusable to me -- It would spit out a 4-5 page plan where it kept repeating itself, where Claude would give me a crisp 1-2 pager I could actually review.
*But* I don't work with the defaults -- I work with my own prompt framework based off of superpowers.
Given sufficient prompt scaffolding, I've found the models relatively interchangeable -- _I might_ be getting some of this for free by basing my own system off of superpowers which is used across various harnesses -- In other words achieving this kind of portability may be a lot harder than it looks and I'm benefiting from other people's work.
I don't really mind hopping between claude/codex/glm/kimi except I don't know a good way to resume as session across agent harnesses.
Normally I'd just have it write out what it's doing to a file, if I need to transfer context, but if it goes down mid-session that's a no-go.
I think people have built tools for this, and of course you could reasonably vibe one yourself, but I don't really trust something like that to work reliably or in an ongoing manner.
> Anthropic have blocked usage of your subscription however with third party harnesses.
This is the main reason I use different harnesses, but I also expect (could be wrong) codex is better with codex harness (due to training on it's specific tools) than with other harnesses. I use opencode for everything that's not claude/codex.
Ahh good point -- I've handled this by switching my harness to `pi` but recognize that may not be for everyone and doesn't directly address OP's question.
When Claude is making "0 mistakes", all of his work is 100% done by Claude, therefore "coding is solved!" and we have more time to go on podcasts to tell everyone about it.
However, when there is an incident it is immediately "human error", not Claude.
> Can’t they prompt Mythos to give them better uptime?
Anthropic is currently "vibe coding" the situation right now.
I've been on the $200 plan for 3 months, but this will be my last month. I got great use out of 4.5 for a while, but 4.6 felt like a half step back (conflated with all the random hidden config changes during it), and 4.7 is genuinely terrible.
It's impossible to tell these days whether 4.7 is stuck because it's thinking and Anthropic suppressed all output (seriously, 4.7 will just start making changes without explaining any reasoning - how is that an upgrade?) or because the underlying infrastructure is having issues.
4.5 -> 4.7 feels like going from working with a coach-able, junior engineer that does well with clear guidance to working with a cocky mid-level that will spend too long on pointless tangents and make confidently incorrect changes without any discussion.
I've been a Codex devotee since around last August. I don't know why everyone is so bonkers about Claude Code. It's not the only belle at the ball. Codex is rock solid.
For coding its fine, I havent experimented too much with Amazon Bedrock myself, but I just might soon to check for any limitations.
I say this as someone working for a tech company who does not have to foot the bill (in the >$1k per month bracket)
I also experienced and accept the 1990s levels of unreliability, which is my “internet generation”. My first access was lifting a handset and placing on a speaker/mic cradle.
Programmers these days are fucking spoiled. If it’s $220 worth of value for $200 - I get it. But I’m getting $100k of value for $10k and so I’ll put up with some shit.
Wrong comparison. If a competitor gives you $230 of value for $200, of course you shouldn't pick the $220 one
I am an API user, and while it being down is super annoying, it isn't really as big of a hit to my overall usage as I can just prepare a bunch of stuff to run in parallel when it does come back up.
Is this just the API and I'm too much of luddite to actually use the API?
Say five eights of reliability. Maybe six.
[0]I say December, because that's around the time the models got good enough that non-AI folks started to notice.
[1]https://en.wikipedia.org/wiki/Eternal_September
I don't really blame Anthropic here.
There's a live Claude status board in the corner so you know when it's time to get back to work.
https://github.com/anthropics/claude-code/issues/54497
I worked with 4.6 and found some improvements for better planning and sustained us, but agree some posters 4.7 is slower, overthinking.
What I expect is frontier models to get bigger and more expensive (especially fast mode like on Cerberus). And most of his get much smaller distillations for the more generous subscription tiers.
It's fine for Claude to be unavailable when there is no work at these hours. However, the problem is Claude gave no notice.
At this rate, Claude being unavailable every day is no better than a human on a 9 - 5 working day job.
99.02 % uptime
Ouch.
https://status.claude.com/
*But* I don't work with the defaults -- I work with my own prompt framework based off of superpowers.
Given sufficient prompt scaffolding, I've found the models relatively interchangeable -- _I might_ be getting some of this for free by basing my own system off of superpowers which is used across various harnesses -- In other words achieving this kind of portability may be a lot harder than it looks and I'm benefiting from other people's work.
Normally I'd just have it write out what it's doing to a file, if I need to transfer context, but if it goes down mid-session that's a no-go.
I think people have built tools for this, and of course you could reasonably vibe one yourself, but I don't really trust something like that to work reliably or in an ongoing manner.
Maybe it should just be a skill.
Anthropic have blocked usage of your subscription however with third party harnesses.
This is the main reason I use different harnesses, but I also expect (could be wrong) codex is better with codex harness (due to training on it's specific tools) than with other harnesses. I use opencode for everything that's not claude/codex.
Still, it's pretty crazy that Claude is down to 1 nine.
However, when there is an incident it is immediately "human error", not Claude.
> Can’t they prompt Mythos to give them better uptime?
Anthropic is currently "vibe coding" the situation right now.
Many such cases with humans (given that we continue to compare LLMs to humans these days which you cannot)
It's impossible to tell these days whether 4.7 is stuck because it's thinking and Anthropic suppressed all output (seriously, 4.7 will just start making changes without explaining any reasoning - how is that an upgrade?) or because the underlying infrastructure is having issues.
4.5 -> 4.7 feels like going from working with a coach-able, junior engineer that does well with clear guidance to working with a cocky mid-level that will spend too long on pointless tangents and make confidently incorrect changes without any discussion.