Infrastructure

Edge Compute

Definition

Edge compute deploys GPU processing capacity at the network edge — closer to end users and data sources rather than in centralised hyperscale facilities. In GPU infrastructure, edge deployments typically serve inference workloads where latency is critical: autonomous vehicles, real-time video analytics, gaming, and local AI inference for enterprise applications. Edge GPU deployments are smaller (1-100 GPUs per site) and geographically distributed, requiring different operational models than centralised training clusters.

Technical Context

Edge GPU infrastructure uses smaller, lower-power accelerators (NVIDIA L4, T4, Jetson) rather than the H100/B200-class GPUs used in training clusters. Deployment environments range from purpose-built micro data centres to ruggedised enclosures in industrial settings. Orchestration across distributed edge sites requires specialised platforms that handle deployment, monitoring, model updates, and failover across potentially hundreds of locations.

Advisory Relevance

Edge compute strategies are relevant for operators and vendors evaluating market positioning. The edge GPU opportunity is distinct from the training cluster market and requires different go-to-market approaches, which we assess in strategy advisory engagements.

This glossary is maintained by Disintermediate as a reference for GPU infrastructure professionals, investors, and operators. Each entry reflects terminology as used in active advisory engagements and market intelligence work.

View all termsDiscuss this topic