Akamai announces an expanded cloud infrastructure offering aimed at supporting AI workloads, featuring an inference cloud designed to run AI models closer to end users. The announcement coincided with first-quarter earnings showing cloud infrastructure revenue rose 40% year over year, alongside disclosure of a $1.8 billion, seven-year commitment from an unnamed frontier AI model provider.
Akamai highlighted its distributed platform, which spans 4,300 points of presence across 700 cities in 130 countries, as infrastructure built to support low-latency compute, storage and security for AI applications. The company described cloud infrastructure as its fastest-growing business segment alongside content delivery and cybersecurity and said it continues expanding capacity to support growing AI demand.
For enterprises, the expanded offering aims to improve inference speed and data locality for AI services while reducing latency for user-facing applications and strengthening deployment security. The deal also reflects rising demand for specialised AI infrastructure and reinforces Akamai’s position in the growing market for distributed AI cloud services.
Cloud Infrastructure Services
Akamai Announces Its $1.8 Billion Deal
Trend Themes
-
Edge Inference Cloud — Akamai’s inference cloud running models nearer to users creates potential for highly responsive AI experiences supported by distributed compute and storage at network edges.
-
Distributed Low-latency Platforms — The company’s 4,300 points of presence illustrate a shift toward geographically dispersed infrastructure that minimizes latency for real-time AI applications and user-facing services.
-
Specialized AI Infrastructure Commitments — Large, long-term financial commitments from frontier model providers indicate growing demand for purpose-built, scalable infrastructure tailored to inference workloads and security requirements.
Industry Implications
-
Cloud Service Providers — Providers with global edge footprints are positioned to differentiate by offering inference-optimized tiers that emphasize locality, throughput, and integrated security.
-
Telecommunications Operators — Carriers owning last-mile networks can leverage edge compute nodes to enable lower-latency AI services for consumers and enterprises across distributed metropolitan areas.
-
Enterprise Software and Applications — Business applications demanding real-time personalization and analytics stand to benefit from closer model hosting that reduces response times and preserves data locality.