Claude Code and Claude API Limits Just Got a Major Boost: What Anthropic’s SpaceX Compute Deal Means for Developers
On May 6, 2026, Anthropic announced a major update for developers: Claude Code usage limits have doubled, Claude API rate limits have been significantly increased, and all changes are effective immediately.
The immediate driver behind this upgrade is a large-scale compute partnership between Anthropic and SpaceX.

What Changed?
Anthropic rolled out three practical changes that developers will feel right away.
1. Claude Code’s Five-Hour Limit Has Doubled
For users on Pro, Max, Team, and seat-based Enterprise plans, Claude Code’s five-hour rolling usage limit has been doubled.
That means more room for high-intensity coding workflows—such as asking Claude Code to analyze a large codebase, run long agentic workflows, or coordinate multiple concurrent file edits—without hitting limits as quickly as before.

2. Peak-Hour Throttling Has Been Removed
Previously, Pro and Max users could see additional rate reductions during peak usage windows, often during U.S. business hours. Anthropic has now removed that extra restriction.
For developers working across different time zones, this is especially useful. If your most productive hours used to overlap with U.S. peak demand, Claude Code may have felt slower or more likely to hit limits. That bottleneck is now gone.
3. Claude Opus API Rate Limits Have Increased Significantly
For flagship models such as claude-opus-4-6 and claude-opus-4-7, Anthropic has raised API rate limits by a substantial margin, described officially as “considerably.”
For teams using Opus models in production for complex reasoning, long-document processing, or high-quality content generation, this directly reduces scaling pressure.
Why Is This Possible? SpaceX’s Colossus 1 Data Center
These improvements are backed by a new compute agreement between Anthropic and SpaceX:
- Data center: SpaceX’s Colossus 1 data center
- New capacity: More than 300 megawatts (MW)
- GPU scale: More than 220,000 NVIDIA GPUs
- Deployment timeline: Coming online within one month
This is not Anthropic’s first major compute expansion. Combined with its existing infrastructure partnerships, the company is scaling aggressively:
| Partner | Scale |
|---|---|
| Amazon | Up to 5 GW, with nearly 1 GW online by the end of 2026 |
| Google + Broadcom | 5 GW, coming online progressively from 2027 |
| Microsoft + NVIDIA | $30 billion in Azure capacity |
| Fluidstack | $50 billion U.S. AI infrastructure investment |
| SpaceX(new) | 300+ MW, 220,000+ GPUs, connected within one month |
The shift from MW-scale to GW-scale capacity shows that Anthropic is preparing infrastructure for large-scale agentic workflows, not just isolated chatbot use.
What This Means for Developers
Scenario 1: Longer Claude Code Sessions With Fewer Interruptions
A typical heavy Claude Code workflow might involve handing over a refactor across dozens of files, or running a test-edit-verify loop for an extended period. Before the five-hour limit was doubled, hitting a cap meant waiting. Now, it is much easier to complete a full engineering workflow in one run.
Scenario 2: More Reliable Agent Workflows
If you are building automated agents with the Claude API—for example, agents that fetch data on a schedule, generate reports, or process user requests—429 rate-limit errors can hurt production reliability. With higher Opus rate limits, agent systems have more headroom under load.
Scenario 3: Faster Batch Code Review and Documentation Jobs
Batch workloads, such as reviewing multiple pull requests or generating documentation comments at scale, are highly sensitive to API throughput. Higher rate limits raise the ceiling for these batch-processing pipelines.
Scenario 4: No More Slowdowns During Peak Windows
If your workday overlaps with U.S. peak usage hours, removing peak-hour throttling means you no longer need to wait for off-hours to run large jobs.
Calling Claude via ClaudeAPI.com: Quick Setup Guide
If direct access to the Claude API is unreliable from your environment, you can route requests through ClaudeAPI.com. It is fully compatible with the official SDKs—you only need to replace the base_url.
Python(using the official Anthropic SDK)
import anthropic
client = anthropic.Anthropic(
api_key="your-api-key",
base_url="https://gw.claudeapi.com"
)
message = client.messages.create(
model="claude-opus-4-7", # Flagship model with increased rate limits
max_tokens=4096,
messages=[
{"role": "user", "content": "Analyze the performance bottlenecks in this code."}
]
)
print(message.content[0].text)
import anthropic
client = anthropic.Anthropic(
api_key="your-api-key",
base_url="https://gw.claudeapi.com"
)
message = client.messages.create(
model="claude-opus-4-7", # Flagship model with increased rate limits
max_tokens=4096,
messages=[
{"role": "user", "content": "Analyze the performance bottlenecks in this code."}
]
)
print(message.content[0].text)
Node.js
import Anthropic from "@anthropic-ai/sdk";
const client = new Anthropic({
apiKey: "your-api-key",
baseURL: "https://gw.claudeapi.com",
});
const message = await client.messages.create({
model: "claude-opus-4-7",
max_tokens: 4096,
messages: [{ role: "user", content: "Review the edge cases in this function." }],
});
console.log(message.content[0].text);
import Anthropic from "@anthropic-ai/sdk";
const client = new Anthropic({
apiKey: "your-api-key",
baseURL: "https://gw.claudeapi.com",
});
const message = await client.messages.create({
model: "claude-opus-4-7",
max_tokens: 4096,
messages: [{ role: "user", content: "Review the edge cases in this function." }],
});
console.log(message.content[0].text);
Claude Code Configuration
Set the following in your Claude Code configuration file:
{
"env": {
"ANTHROPIC_BASE_URL": "https://gw.claudeapi.com",
"ANTHROPIC_API_KEY": "your-api-key"
}
}
{
"env": {
"ANTHROPIC_BASE_URL": "https://gw.claudeapi.com",
"ANTHROPIC_API_KEY": "your-api-key"
}
}
Model Recommendations
Based on the new limit increases, here are practical model choices for common developer workflows:
| Use Case | Recommended Model | Why |
|---|---|---|
| Everyday coding with Claude Code | claude-sonnet-4-6 |
Fast, cost-effective, and ideal for frequent back-and-forth coding tasks |
| Complex architecture analysis / code refactoring | claude-opus-4-7 |
Higher rate limits and strongest reasoning capability |
| Batch documentation generation | claude-haiku-4-5-20251001 |
High throughput and low cost |
| Production agents | claude-sonnet-4-6 |
Strong price-performance ratio with enough rate-limit headroom |
Full pricing reference: ClaudeAPI.com model pricing page
Final Thoughts
The key signal from this update is clear: Anthropic’s compute expansion is now translating into immediate, practical benefits for everyday developers, not just future promises.
If you are using Claude Code for complex engineering work, or relying on Claude Opus models in production, now is a good time to revisit your workflow limits. Many tasks that previously felt risky because of rate limits may now be worth trying again.
This article is based on Anthropic’s official announcement, with technical parameters sourced from official documentation. To access Claude API through ClaudeAPI.com, visit claudeapi.com to get an API key.



