• News In Brief
  • AI
  • Education
  • Pro AV
  • Case Study
  • Interview
No Result
View All Result
SUBSCRIBE
Smart Solutions World
  • News In Brief
  • AI
  • Education
  • Pro AV
  • Case Study
  • Interview
No Result
View All Result
No Result
View All Result
Home AI

Akamai Launches AI Grid Intelligent Orchestration for Distributed Inference Across 4,400 Edge Locations

SmartSolutionUser1 by SmartSolutionUser1
March 20, 2026
in AI
0
Akamai Launches AI Grid Intelligent Orchestration for Distributed Inference Across 4,400 Edge Locations
75
SHARES
1.2k
VIEWS
Share on FacebookShare on Twitter

Akamai Technologies reached a major milestone in the evolution of artificial intelligence, unveiling the first global-scale implementation of NVIDIA® AI Grid reference design. By integrating NVIDIA AI infrastructure into Akamai’s infrastructure, and leveraging intelligent workload orchestration across its network, Akamai intends to move the industry beyond isolated AI factories toward a unified, distributed grid for AI inference.

You might also like

Onix Deepens Strategic Collaboration with Google Cloud to Help Accelerate Enterprise-Scale Cloud, Data, and Agentic AI Transformation

4 in 5 Indian Firms Deploy AI Under Pressure, Security Gaps Persist – TrendAI

SiMa.ai Secures Strategic Investment from Micron to Scale High-Performance, Power-Efficient Physical AI

The move marks a significant step in the evolution of Akamai’s Inference Cloud, introduced late last year. As the first to operationalize the AI Grid, Akamai is rolling out thousands of NVIDIA RTX PRO 6000 Blackwell Server Edition GPUs, providing a platform to enable enterprises to run agentic and physical AI with the responsiveness of local compute and the scale of the global web.

Mr. Adam Karon, Chief Operating Officer and General Manager, Cloud Technology Group, Akamai.
Mr. Adam Karon, Chief Operating Officer and General Manager, Cloud Technology Group, Akamai.

“AI factories have been purpose-built for training and frontier model workloads — and centralized infrastructure will continue to deliver the best tokenomics for those use cases,” said Mr. Adam Karon, Chief Operating Officer and General Manager, Cloud Technology Group, Akamai. “But real-time video, physical AI, and highly concurrent personalized experiences demand inference at the point of contact, not a round trip to a centralized cluster. Our AI Grid intelligent orchestration gives AI factories a way to scale inference outward — leveraging the same distributed architecture that revolutionized content delivery to route AI workloads across 4,400 locations, at the right cost, at the right time.”

The Architecture of ‘Tokenomics’

At the heart of the AI Grid is an intelligent orchestrator that acts as a real-time broker for AI requests. Applying Akamai’s expertise in application performance optimization to AI, this workload-aware control plane optimizes “tokenomics” by radically improving cost per token, time-to-first-token, and throughput.

A major differentiator for Akamai is the ability for customers to access fine-tuned or sparsified models through its enormous global edge footprint, which offers a massive cost and performance advantage for the long tail of AI workloads. For example:

  • Cost Efficiency at Scale: Enterprises can dramatically reduce inference costs by matching workloads to the right compute tier automatically. The orchestrator applies techniques like semantic caching and intelligent routing to direct requests to right-sized resources, reserving premium GPU cycles for the workloads that demand them. Underpinning this is Akamai Cloud, built on open-source infrastructure with generous egress allowances to support data-intensive AI operations at scale.
  • Real-Time Responsiveness: Gaming studios can deliver AI-driven NPC interactions that maintain player immersion in milliseconds. Financial institutions can execute personalized fraud detection and marketing recommendations in the moment between login and first screen. Broadcasters can transcode and dub content in real time for global audiences. These outcomes are powered by Akamai’s globally distributed edge network with over 4,400 locations with integrated caching, serverless edge compute, and high-performance connectivity that processes requests at the point of user contact, bypassing the round-trip lag of origin dependent clouds.
  • Production-Grade AI at the Core: Large language models, continuous post-training, and multi-modal inference workloads require sustained, high-density compute that only dedicated infrastructure can deliver. Akamai’s multi-thousand GPU clusters, powered by NVIDIA RTX PRO 6000 Blackwell Server Edition GPUs, provide the concentrated horsepower for the heaviest AI workloads, complementing the distributed edge with centralized scale.

The Continuum of Compute: From Core to Far-Edge

Built on NVIDIA AI Enterprise and leveraging NVIDIA Blackwell architecture and NVIDIA BlueField DPUs for hardware-accelerated networking and security, Akamai is able to manage complex SLAs across edge and core locations:

  • The Edge (4,400+ locations): Delivers rapid response times for physical AI and autonomous agents. It will leverage semantic caching and serverless capabilities like Akamai Functions (WebAssembly-based compute) and EdgeWorkers to deliver model affinity and stable performance at the point of user contact.
  • Akamai Cloud IaaS and Dedicated GPU Clusters: Core public cloud infrastructure enables portability and cost savings for massive-scale workloads, while pods powered by NVIDIA RTX PRO 6000 Blackwell GPUs enable heavy-duty post-training and multi-modal inference.
Mr. Chris Penrose, Global VP - Business Development - Telco at NVIDIA.
Mr. Chris Penrose, Global VP – Business Development – Telco at NVIDIA.

“New AI-native applications demand predictable latency and better cost efficiency at planetary scale,” said Mr. Chris Penrose, Global VP – Business Development – Telco at NVIDIA. “By operationalizing the NVIDIA AI Grid, Akamai is building the connective tissue for generative, agentic, and physical AI, moving intelligence directly to the data to unlock the next wave of real-time applications.”

Powering the Next Wave of Real-Time AI

Akamai is already seeing strong, early adoption for Akamai Inference Cloud across compute-intensive, latency-sensitive industries:

  • Gaming: Studios are deploying sub-50-millisecond inference for AI-driven NPCs and real-time player interactions.
  • Financial Services: Banks rely on the grid for hyper-personalized marketing and rapid recommendations in the critical moments when customers log in.
  • Media and Video: Broadcasters use the distributed network for AI-powered transcoding and real-time dubbing.
  • Retail and Commerce: Retailers are adopting the network for in-store AI applications and associate productivity tools at the point of sale.

Driven by enterprise demand, the platform has also been validated by major technology providers, including a $200 million, four-year service agreement for a multi-thousand GPU cluster in a data center purpose-built for enterprise AI infrastructure at the metro edge.

Scaling AI Factories from Centralized to Distributed

The first wave of AI infrastructure was defined by massive GPU clusters in a handful of centralized locations, optimized for training. But as inference becomes the dominant workload and businesses across every industry focus on building AI agents, that centralized model faces the same scaling constraints that earlier generations of internet infrastructure encountered with media delivery, online gaming, financial transactions, and complex microservices applications.

Akamai is solving each of those challenges through the same fundamental approach: distributed networking, intelligent orchestration, and purpose-built systems that bring content and context together as close as possible to the digital touchpoint. The result has been improved user experiences and stronger ROI for the enterprises that adopted the model. Akamai Inference Cloud applies that same proven architecture to AI factories, enabling the next wave of scaling and growth by distributing dense compute from core to edge.

For enterprises, this means the ability to deploy AI agents that are context-aware and adaptive in their responsiveness. For the industry, it represents a blueprint for how AI factories evolve from isolated installations into a globally distributed utility.

Availability

Akamai Inference Cloud is available today for qualified enterprise customers. Organizations can learn more and request access at https://www.akamai.com/products/akamai-inference-cloud-platform. Akamai representatives will be available for demonstrations and meetings throughout NVIDIA GTC 2026 at the San Jose Convention Center, Booth 621 March 16–19, 2026.

If you have an interesting Article / Report/case study to share, please get in touch with us at editors@roymediative.com roy@roymediative.com, 9811346846/9625243429.

Tags: 400 Edge LocationsAkamaiAkamai Launches AI Grid Intelligent Orchestration for Distributed Inference Across 4Launches AI Grid Intelligent Orchestrationsmart solutions world
Share30Tweet19
SmartSolutionUser1

SmartSolutionUser1

Recommended For You

Onix Deepens Strategic Collaboration with Google Cloud to Help Accelerate Enterprise-Scale Cloud, Data, and Agentic AI Transformation

by SmartSolutionUser1
April 11, 2026
0
Onix Deepens Strategic Collaboration with Google Cloud to Help Accelerate Enterprise-Scale Cloud, Data, and Agentic AI Transformation

Onix, a leading Data and AI services-as-software company, announced a major expanded strategic collaboration with Google Cloud. Leveraging its proprietary Wingspan agentic AI and Data modernization platform, Onix...

Read moreDetails

4 in 5 Indian Firms Deploy AI Under Pressure, Security Gaps Persist – TrendAI

by SmartSolutionUser1
April 10, 2026
0
4 in 5 Indian Firms Deploy AI Under Pressure, Security Gaps Persist – TrendAI

TrendAI™ has published new global research revealing that organizations worldwide are pushing ahead with AI deployment despite known security and compliance risks. A new global study* of 3,700...

Read moreDetails

SiMa.ai Secures Strategic Investment from Micron to Scale High-Performance, Power-Efficient Physical AI

by SmartSolutionUser1
April 10, 2026
0
SiMa.ai Secures Strategic Investment from Micron to Scale High-Performance, Power-Efficient Physical AI

SiMa.ai, a leader in Physical AI, announced a strategic investment from Micron Technology, Inc., strengthening its ability to scale production-ready, high-performance Physical AI solutions for real-world intelligent systems....

Read moreDetails

AI&Beyond Launches ‘AI&Beyond Partner Circle’ to Scale AI Adoption Across Enterprises

by SmartSolutionUser1
April 9, 2026
0
AI&Beyond Launches ‘AI&Beyond Partner Circle’ to Scale AI Adoption Across Enterprises

AI&Beyond, India’s pioneering AI literacy company, today announced the launch of the AI&Beyond Partner Circle, a curated partnership programme designed to accelerate enterprise AI adoption through trusted networks...

Read moreDetails

AI, Cybersecurity, digital and data skills emerge as India’s most critical future capabilities – NIIT India Skills Gap Report

by SmartSolutionUser1
April 8, 2026
0
AI, Cybersecurity, digital and data skills emerge as India’s most critical future capabilities – NIIT India Skills Gap Report

NIIT Ltd., a leading Skills & Talent development corporation, launched the NIIT India Skills Gap Report 2026, a nationwide study conducted in partnership with YouGov. The survey, based...

Read moreDetails
Next Post
Gorilla Technology & Yotta Sign Landmark AI Infrastructure Deal, Establishing Major Position in India’s Sovereign AI Buildout

Gorilla Technology & Yotta Sign Landmark AI Infrastructure Deal, Establishing Major Position in India’s Sovereign AI Buildout

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Browse by Category

Browse by Category

Smart Solutions World

We bring you the best Premium news, magazine, personal blog, etc. Check our landing page for details.

  • News In Brief
  • AI
  • Education
  • Pro AV
  • Case Study
  • Interview

BROWSE BY TAG

Acquisition Agentic AI Agora AI Akamai AMD Aurionpro automation Cloudflare CloudKeeper Coforge CrowdStrike Cybersecurity Databricks Fortinet Gartner GenAI Genesys Google Cloud Honeywell IBM Infosys Kaspersky Kramer LTIMindtree Microsoft New Relic Nvidia OpenAI Palo Alto Networks PPDS Qlik Qualcomm Seqrite ServiceNow SiMa.ai smart solutions world smartsolutionsworld smart solutions world latest news Software Tata Communications Tech Mahindra Technology Tenable Vertiv

© 2024 NCN - Premium news & magazine by NCN.

No Result
View All Result
  • News In Brief
  • AI
  • Education
  • Pro AV
  • Case Study
  • Interview

© 2024 NCN - Premium news & magazine by NCN.

Are you sure want to unlock this post?
Unlock left : 0
Are you sure want to cancel subscription?