Anthropic Raises Claude Code Usage Limits, Credits New Deal With SpaceX (arstechnica.com) 43
An anonymous reader quotes a report from Ars Technica: At its Code with Claude developer conference on Wednesday, Anthropic announced a deal with SpaceX to utilize the entire compute capacity of the latter's data center in Memphis, Tennessee. On stage at the conference, CEO Dario Amodei said the deal was intended to increase usage limits for Anthropic's Pro and Max plan subscribers. The announcement was accompanied by an increase in those usage limits; Anthropic doubled Claude Code's five-hour window limits for Pro and Max subscribers, removed the peak-hours limit reduction on Claude Code for those same accounts, and raised API limits for its Opus model. The table [here] outlining the Opus changes was shared in the company's blog post on the topic.
Anthropic claims the deal gives the company access to more than 300 megawatts of new compute capacity. For its part, SpaceX focused its announcement on the capability of the Colossus 1 supercomputer that's at the center of the deal. "Colossus 1 features over 220,000 NVIDIA GPUs, including dense deployments of H100, H200, and next-generation GB200 accelerators," SpaceX wrote. Additionally, Anthropic "expressed interest" in working with SpaceX to build up "multiple gigawatts" of orbital compute capacity, tying into a recent (but unproven) focus on exploring orbital data centers as an answer to the problem that "compute required to train and operate the next generation of these systems is outpacing what terrestrial power, land, and cooling can deliver on the timelines that matter." "I spent a lot of time last week with senior members of the Anthropic team to understand what they do to ensure Claude is good for humanity and was impressed," Elon Musk said on Wednesday. "No one set off my evil detector."
Anthropic claims the deal gives the company access to more than 300 megawatts of new compute capacity. For its part, SpaceX focused its announcement on the capability of the Colossus 1 supercomputer that's at the center of the deal. "Colossus 1 features over 220,000 NVIDIA GPUs, including dense deployments of H100, H200, and next-generation GB200 accelerators," SpaceX wrote. Additionally, Anthropic "expressed interest" in working with SpaceX to build up "multiple gigawatts" of orbital compute capacity, tying into a recent (but unproven) focus on exploring orbital data centers as an answer to the problem that "compute required to train and operate the next generation of these systems is outpacing what terrestrial power, land, and cooling can deliver on the timelines that matter." "I spent a lot of time last week with senior members of the Anthropic team to understand what they do to ensure Claude is good for humanity and was impressed," Elon Musk said on Wednesday. "No one set off my evil detector."
no one set off my eveil detector (Score:2, Funny)
RONTHFLMFASO x 10^42 power
Re: (Score:2)
Re: (Score:3, Informative)
Re: (Score:1)
Re: Not me! (Score:2)
meh (Score:2)
Huh, haven't seen a change in my Pro subscription. Though I hit the weekly model limit -- if that's not changing then it's not really a big deal.
Re: (Score:2)
FlashAttention (Score:3)
I did some math the other day on running local AI models and the net result is most homes can't afford to run the current median models.
They don't just need 80GB of VRAM, they need newer architectures - to be supported by CUDA, to be supported by pytorch, etc.
These problems may well be solvable with more clever use of hardware, MoE, acceptable quantization, etc., but today you're in for several grand and something north of 100W idle to use what is effectively a $20/mo plan.
A small enterprise can afford local, so that's good. We paid more than that for one SGI machine back in the day.
The point of the exercise was to plot the position on the curve. We're at something like 2006 YouTube where nobody could afford the drives or bandwidth that YouTube/Google was giving away for free (aka with VC money). Eventually hard drives got cheaper, people got gigabit at home, FlashServer was replaced with h.264/HTML5, phones could stabilize video locally, etc.
So it looks like these AI companies need to stay alive for about seven more years giving away product at a loss, or at least highly oversubscribed, to turn a profit. Hence the low token allowance, the banning of OpenClaw, etc.
On the other hand, I read the blog of a security researcher yesterday who found an exploit with (IIRC) Claude, tried to refine the PoC, but got dinged on "out of tokens" before he could finalize it. So he just deleted the work and moved on.
It sounds like they're trying to not lose money at such a velocity and are trying to find a sweet spot where people don't just declare it too underpowered to use.
A global energy depression may well take out the supermajority of the companies that believe they can burn investment money for seven more years. There is circular financing money, then there is real return on capital money. One is to fool the markets, the other is grounded in current physics.
Re: (Score:2)
> So it looks like these AI companies need to stay alive for about seven more years giving away product at a loss, or at least highly oversubscribed, to turn a profit. Hence the low token allowance, the banning of OpenClaw, etc.
You're assuming that tech will continue to improve as fast as it did over the last twenty years... when we're past the end of Moore's Law.
At a minimum that would likely involve moving away from silicon to something we can run at much higher clock speeds, or finding a much more eff
Re: (Score:2)
That is true. So I guess that's another way to try to beat Moore's Law (or the lack of it).
Though the better the low-end models get, the less need for massive data centres as people can run them locally.
Re: (Score:3)
Nah. If you don't absolutely challenge Opus you can get away with 180 Watt peak and 16 GB VRAM graphics cards like 5060 Ti (probably the best value starter card for AI at the moment). Using MoE models helps and getting up-to-date with new releases is important as nobody will deploy the latest and greatest model if you don't do it yourself. You won't beat Opus, but you're more or less on par with Sonnet, in particular when your client fits your use case well.
Re: (Score:2)
Don't many of these models process tokens at like 5 / sec on something like a 5060ti?
Re: (Score:2)
Re: (Score:2)
I'm getting about 35 tokens/s on a 3060ti 12GB with Qwen 3.6 MoE Q4, KV 8/8, and would get about 50-60 with KV 4/4. That's with 128K context.
Re: (Score:2)
This gives people more time to become addicted to the dopamine rush from using AI and to bond with their chatbots. Anthropic is like the drug dealer who gives out free samples to get more customers.
AI is the performance enhancing drug for tech workers. Someday we'll realize that we need to limit AI time just as we try to limit screen time.
Re: not Claude, but others (Score:1)
Re: (Score:2)
Not loving it (Score:2)
I hate that it uses jet engines to supplement power of the data centers. Makes me think they should scale back the token usage on them.
Re: (Score:1)
Jet Engines are as efficient as the steam turbine generators used at power plants.
Re: (Score:2)
Jet Engines are as efficient as the steam turbine generators used at power plants.
Sure, if the goal is to produce thrust. If it's to produce electricity, then no, because you have to spend thrust to achieve compression. Steam plants produce compression with a pressure vessel, not with thrust.
Re: (Score:2)
Sir, are you aware of how thrust is generated? Compression.
Re: (Score:2)
Sir, are you aware of how thrust is generated? Compression.
Sir, are you aware that you get nothing for free?
Model Limits Double! (Score:1)
Hold on there, space cowboy (Score:2)
Are.
You.
Going.
To.
Cool.
It.
In.
Space.
Though?
I feel like someone needs to keep tweeting this at Musk until he gets it. There's no air in space. You need infrared radiators. Those weigh A LOT and don't work very well.
Re: (Score:2)
This is great. He's going to build this stuff
Like he did FSD, which is neither F nor SD? Suck harder.
Re: (Score:2)
Re: (Score:2)
Re: Hold on there, space cowboy (Score:1)
A totally wasted effort. Good luck finding even one engineer at SpaceX who hasn't already thought of that and some.
Re: (Score:2)
SpaceX owns xAI/Grok and has a bunch of servers in their datacenters that aren't profitable.
As for space cooling - that's a problem, but they only need to dissipate as much energy as they absorb. It's not impossible
ugh. No, Elon. No. Just no (Score:2, Troll)
Buddy, I respect your accomplishments. You deserve your spot as the world's richest man. And, overall, I understand that you're trying to sincerely push humanity forward, since that goal lines up well with your empire-building. But, don't try and act like a caring leader. It's not a good look on yo
Re: (Score:2)
holy shit, have you lost your mind?
Good point. Leon really doesn't deserve to be the world's richest man.
Kessler, meet BOINC... (Score:2)
I'm still rocking a 2009 iMac. I don't need much, check my email, pay my bills, slice some .STLs, play some older games; thanks to Steam sales I have a lot of older games. When it's not being used I let BONIC [berkeley.edu] run in the background. For those who don't know, BONIC is a distributed compute platform which helps Colleges and Universities across the globe leverage unused compute cycles from computers people have volunteered to let the researchers use when their machines are idle.
Over the years my old iMac has
Nothing changes (Score:2)
Anthropics remains a toxic, opaque company that treats its customers with disdain. They change usage limits and terms of use regularly without informing users, banning users for life without explaining the cause.
The competitors might not be much better, but at least they seem to be more transparent. We better hope that Anthropic does not win the AI race without a change of policies.
Re: (Score:2)
Precisely why I jumped to local inference, and spent about $20/mo checking the work and planning on cloud models. You don't need much hardware to effectively replace claude.
Huh (Score:2)
Shouldn't this datacenter be utilized by xAI (which is part of SpaceX now)? Why would they be leasing the capacity to a competitor instead of using it for their own internal AI corp?
Also is this datacenter built with the NVidia processors that were originally purchased by Tesla (for their AI project: Self Driving), before Musk decided to redirect them to his competing AI company xAI that he started because Tesla wouldn't guarantee him the controlling ownership stake he demanded "or else he would build AI e
Re: (Score:2)
Nobody uses xAI except maybe X bots. I doubt X employees use X to do infra/code tasks. Companies are throwing $$ at anthropic for Claude Code (mainly) and I think Anthropic will lease compute power from anyone as long as it keeps their frontier models humming along.
300 megawatts (Score:2)
I hate this new so-called unit of measure. Make it less efficient, and you get a higher number!
broken evil detector (Score:1)