Anthropic’s $1.8 Billion Akamai Deal Marks a New Kind of Cloud Partnership
Anthropic’s reported seven-year cloud agreement with Akamai Technologies, worth USD 1.8 billion (approx. RM8.28 billion), underscores how frontier AI labs are reshaping cloud infrastructure spending. Akamai, better known for its content delivery network, has been positioning its distributed GPU and edge architecture for low-latency AI inference workloads. That design aligns with Anthropic’s Claude models, which now power longer-running coding sessions, managed agents, and hosted automations rather than brief chat exchanges. These use cases keep inference workloads active, stressing both capacity and responsiveness. While key terms such as the exact hardware mix, regional footprint, and reserved volume remain undisclosed, Akamai’s own description of millisecond-level inference for agentic applications suggests the deal is tailored to production-scale AI services. For Akamai, it is reportedly the largest contract in its history, signaling how AI compute deals are becoming anchor tenants for cloud providers’ future buildouts.

Inside Anthropic’s Proposed $200 Billion Google Cloud and Chips Commitment
Alongside Akamai, Anthropic is reportedly preparing an even larger move: a potential USD 200 billion (approx. RM920 billion) commitment over five years for Google Cloud and chip access, if the figure is confirmed. Public disclosures already show a significant TPU capacity expansion with Google and Broadcom expected to come online starting in 2027, framing this as a long-term, deep infrastructure relationship rather than ordinary cloud usage. The proposed purchasing window would shift Anthropic’s heaviest commitments into the next data center buildout cycle, giving Google clearer visibility into future demand. That visibility helps justify early investments in new facilities, power, networking, and specialized hardware before every rack is installed. For Anthropic, the strategy is about solving supply problems in advance—ensuring there is enough training and inference capacity for frontier models, even as enterprise demand for coding assistance, analytics, and customer support use cases accelerates.
From Training Runs to Always-On Inference: How AI Economics Are Shifting
These large AI compute deals signal a structural shift in AI economics: spending is tilting from sporadic training runs toward continuous, production-scale inference. Anthropic’s Claude models illustrate this shift. Agentic features, tool use, and persistent coding sessions require models to remain active long after the initial answer is generated, putting steady pressure on inference capacity. Akamai’s distributed edge footprint allows Anthropic to spread these workloads closer to end users, reducing latency while keeping data center utilization high. At the same time, the extended runway with Google’s TPU expansion ensures training clusters can be assembled without relying on spot markets or last-minute hardware procurement. Together, these moves show that AI labs now treat capacity planning as a central product decision. Reliability, latency, and scale for AI inference are becoming as strategically important as model quality, redefining how labs allocate capital and negotiate with cloud providers.
Cloud Lockups, Chip Supply Agreements, and the Squeeze on Smaller AI Labs
Anthropic’s multi-supplier strategy—reserving capacity across Google, Akamai, and earlier agreements with other providers—highlights a broader industry trend of locking in cloud and chip supply agreements years ahead. For major labs, long-term reservations reduce operational uncertainty around AI inference scaling, protecting against bottlenecks in land, power, cooling, and hardware installations. However, this consolidation has implications for the rest of the market. As big customers pre-book future capacity, smaller AI startups may face longer waiting times, higher prices, and weaker bargaining power when they seek access to the same GPU clusters or TPUs. Capacity that might once have been available on flexible terms increasingly sits behind multi-year contracts. In effect, the economics of AI infrastructure are converging on a utility model where the largest, earliest buyers secure the most favorable paths to scale, leaving others to navigate a thinner, more expensive spot market for compute.
