Infrastructure
Infrastructure is the foundational layer of technology upon which the entire digital economy — and increasingly, the agentic economy — is built. It encompasses semiconductors, networks, cloud computing, data centers, edge computing, energy systems, and the specialized AI compute fabric that powers everything from large language model training to real-time agent inference. These systems are largely invisible to end users but determine the performance ceiling, cost floor, and geographic reach of every digital experience built on top of them.
The AI Infrastructure Supercycle
The AI era has triggered the largest infrastructure buildout since the early internet — except measured in trillions of dollars rather than billions. In 2026, the hyperscale cloud providers collectively plan over $300 billion in capital expenditure, overwhelmingly directed at AI-optimized data centers. NVIDIA's GPU dominance has made it one of the world's most valuable companies, as demand for AI training and inference compute far outstrips supply. New entrants like Cerebras, Groq, and SambaNova are challenging the GPU orthodoxy with purpose-built AI accelerators, while AMD, Intel, and Qualcomm compete for share across training, inference, and edge workloads.
Semiconductors: The Foundation Layer
Every computation in the digital economy ultimately runs on silicon. The semiconductor supply chain — from ASML's lithography machines to TSMC's fabrication to GPU and AI accelerator design — is the deepest infrastructure layer. The geopolitical significance of this supply chain has driven massive policy interventions like the CHIPS Act, as nations recognize that semiconductor independence is a prerequisite for AI sovereignty. High Bandwidth Memory (HBM) has emerged as a critical bottleneck, with SK Hynix and Micron racing to meet demand from AI accelerator manufacturers.
Cloud and Compute Platforms
The three hyperscale cloud providers — AWS, Azure, and Google Cloud — provide the compute, storage, and networking substrate for most AI workloads. But the AI era has spawned a new category of specialized GPU cloud providers: CoreWeave, Lambda, Together AI, and Fireworks AI offer GPU-optimized infrastructure at competitive prices. Meanwhile, platforms like Databricks and Snowflake provide the data layer that AI systems operate on, and inference-specific platforms like Groq and Replicate optimize for serving trained models at scale.
Networks and Connectivity
Network infrastructure determines how fast data moves between users, devices, data centers, and AI systems. 5G Standalone deployments are accelerating, with over 125 mobile operators expected to launch SA services by end of 2026, while 6G research is transitioning from theoretical to early experimental prototyping. Fiber buildout continues to expand backbone capacity. For AI specifically, high-speed interconnects between GPUs and across data centers — technologies like NVLink, InfiniBand, and optical interconnects — are as critical as the processors themselves, since distributed training and inference are fundamentally network-bound problems.
Edge Computing and On-Device AI
Edge computing is moving from emerging to mainstream, pushing cloud-based processing closer to end users. For AI agents, spatial computing, and real-time applications, the latency of a round-trip to the cloud is often unacceptable. On-device inference — powered by Apple's Neural Engine, Qualcomm's AI Engine, and dedicated NPUs — brings AI to phones, laptops, and IoT devices. The tension between cloud and edge inference, and the hybrid architectures that bridge them, shapes how AI reaches users worldwide.
Energy and Sustainability
AI infrastructure has an energy problem. A single large model training run can consume as much electricity as a small city uses in a month. Data center power consumption is projected to double by 2028, driving unprecedented demand for energy infrastructure. Nuclear power — both traditional and small modular reactors — has emerged as the preferred clean energy source for AI data centers, with Microsoft, Amazon, and Google all signing nuclear power agreements. The energy constraint is increasingly the binding constraint on AI scaling, making power infrastructure as strategic as compute infrastructure.
The Agentic Infrastructure Stack
As the economy shifts toward agentic AI, infrastructure requirements are evolving beyond batch training and simple inference. Agents need low-latency, always-on inference; persistent memory and state management; secure tool access via protocols like MCP and A2A; and the ability to compose actions across multiple services. This is driving demand for a new category of AI-native infrastructure that goes beyond traditional cloud computing. The Agentic Economy Market Map illustrates how infrastructure forms the foundational layer upon which the entire agentic economy is built.
Further Reading
- Market Map of the Agentic Economy — How infrastructure forms the foundation layer
- The State of AI Agents in 2026 — Infrastructure demands of agentic AI
- The Metaverse Value-Chain — Infrastructure's role in the broader value chain
- Compute Capital Markets — The economics of AI compute infrastructure