Subspace \USA

Subspace positioned itself as a 'real-time internet' infrastructure company, building a dedicated global network optimized for latency-sensitive applications like gaming, video streaming, and real-time communications. The value proposition centered on solving the 'last mile' problem for interactive applications by creating a parallel internet backbone with proprietary routing algorithms that could guarantee sub-50ms latency globally. Founded in 2018 when cloud gaming (Stadia, GeForce Now) and metaverse concepts were gaining momentum, Subspace aimed to be the infrastructure layer enabling the next generation of real-time experiences. They built physical network infrastructure (PoPs across 200+ cities) and software-defined networking to dynamically route traffic through optimal paths, bypassing congested public internet routes. The 'why now' was compelling: 5G rollout, explosion of multiplayer gaming, rise of remote work requiring low-latency video, and increasing demand for sub-100ms experiences. However, they were essentially rebuilding core internet infrastructure—a capital-intensive, decade-long bet requiring massive scale before unit economics worked.

SECTOR Information Technology
PRODUCT TYPE SaaS (B2C)
TOTAL CASH BURNED $26.0M
FOUNDING YEAR 2018
END YEAR 2022

Discover the reason behind the shutdown and the market before & today

Failure Analysis

Failure Analysis

Subspace died from the classic infrastructure startup trap: capital intensity meeting slow enterprise sales cycles during a market downturn. The mechanics of failure were...

Expand
Market Analysis

Market Analysis

The real-time internet infrastructure market in 2024 is dominated by three players: Cloudflare (edge compute + CDN + security platform, $1.2B revenue), AWS (Wavelength...

Expand
Startup Learnings

Startup Learnings

Infrastructure startups require 3x more capital than founders estimate. Subspace likely modeled $50M to scale, but realistically needed $150M+. Modern founders should assume infrastructure...

Expand
Market Potential

Market Potential

The TAM for low-latency infrastructure remains massive and growing. In 2018, the addressable market included cloud gaming ($2B, projected to $20B+ by 2025), live...

Expand
Difficulty

Difficulty

In 2018-2022, building a global network infrastructure company required massive capital expenditure for physical hardware, data center colocation, peering agreements, and custom routing software....

Expand
Scalability

Scalability

Subspace had fundamentally poor scalability economics. Each new geographic market required physical infrastructure deployment (edge nodes, peering agreements, hardware), creating linear cost scaling. Unlike...

Expand

Rebuild & monetization strategy: Resurrect the company

Pivot Concept

+

An AI-native edge inference platform optimized for real-time multimodal AI applications (voice agents, video generation, spatial computing). Instead of building physical infrastructure first, Latency is a software layer that intelligently routes AI inference requests across existing cloud providers (AWS, GCP, Azure, Cloudflare Workers) using ML-optimized routing to guarantee sub-100ms response times globally. The wedge is real-time voice AI agents (ChatGPT voice mode competitors, AI phone agents, voice-first interfaces) where latency is the primary UX differentiator. As revenue scales, deploy specialized AI inference hardware (H100 clusters) in strategic edge locations. The moat is a proprietary latency prediction model trained on billions of inference requests, combined with a developer platform (SDKs, APIs, observability) that makes building real-time AI apps trivial. Monetization is usage-based: $0.01 per 1K inference requests with latency SLA guarantees, targeting the $10B+ real-time AI application market.

Suggested Technologies

+
Cloudflare Workers (initial edge compute layer, 300+ PoPs globally)AWS Lambda@Edge + Wavelength (fallback routing for specialized regions)Modal/Baseten (GPU inference orchestration for model serving)Ray Serve (distributed inference framework for multi-model routing)Temporal (workflow orchestration for complex inference pipelines)ClickHouse (real-time analytics for latency monitoring and ML training)Vercel Edge Functions (developer-facing API gateway)Stripe (usage-based billing infrastructure)LangChain/LlamaIndex (AI agent orchestration layer)WebRTC/LiveKit (real-time audio/video streaming for voice agents)Prometheus + Grafana (observability and latency SLA monitoring)Terraform (multi-cloud infrastructure as code)Claude/GPT-4/Llama 3 (inference models, starting with API-based, moving to self-hosted)

Execution Plan

+

Phase 1

+

Step 1 (Wedge - Month 1-3): Build a developer SDK for real-time voice AI agents that abstracts latency optimization. Launch with a single use case: AI phone agents for customer support (competing with Bland.ai, Retell). Use Cloudflare Workers + OpenAI/Anthropic APIs for inference, with intelligent routing based on user geography. Target 10 early design partners (YC companies building voice AI) with free tier (10K requests/month). Validate that sub-100ms latency improves conversation naturalness by 40%+ in user testing. Deliver a Vercel-like DX: `npm install @latency/voice-sdk`, 10 lines of code to deploy a voice agent with global latency optimization.

Phase 2

+

Step 2 (Validation - Month 4-6): Expand to real-time video generation (Runway/Pika competitors) and multimodal AI agents (voice + vision). Build proprietary latency prediction model using ClickHouse to log every inference request (origin, destination, model, latency) and train an ML model (XGBoost/LightGBM) to predict optimal routing. Launch usage-based pricing: $0.01 per 1K requests with 99.9% uptime SLA and <100ms p95 latency guarantee. Achieve $50K MRR from 50 customers (indie developers, AI startups). Key metric: 60%+ of customers cite latency as primary reason for choosing Latency over direct API calls to OpenAI/Anthropic. Build observability dashboard (Grafana) showing real-time latency heatmaps globally.

Phase 3

+

Step 3 (Growth - Month 7-12): Launch 'Latency Edge' - deploy first 10 self-hosted GPU clusters (H100s) in strategic locations (SF, NYC, London, Singapore, Tokyo) for customers needing <50ms latency or proprietary model hosting. Partner with Modal/Baseten for GPU orchestration. Expand to spatial computing use case: real-time AI for Vision Pro/Quest apps (6DOF tracking, scene understanding, real-time rendering). Achieve $500K MRR from 200 customers. Launch PLG motion: free tier with 100K requests/month, self-serve upgrade to Pro ($99/month + usage), Enterprise (custom SLAs). Key growth loop: developers building real-time AI apps discover Latency through GitHub/Twitter, try free tier, convert to paid when they hit scale. Target 40% month-over-month growth.

Phase 4

+

Step 4 (Moat - Month 13-24): Build the 'Cloudflare for AI Inference' platform. Expand edge network to 50+ locations globally (mix of self-hosted GPU clusters and partnerships with cloud providers). Launch 'Latency Agents' - a managed platform for deploying production-grade AI voice/video agents with built-in latency optimization, observability, and compliance (HIPAA, SOC2). Introduce model marketplace: developers can deploy custom fine-tuned models (Llama, Mistral) on Latency's edge network. Achieve $5M ARR from 1,000+ customers. Moat is threefold: (1) proprietary latency prediction model trained on billions of requests, (2) developer ecosystem (10K+ developers, open-source SDKs, community), (3) physical edge infrastructure that takes competitors 2+ years to replicate. Exit strategy: acquisition by Cloudflare, AWS, or Anthropic/OpenAI as their edge inference layer, or continue scaling to $50M+ ARR as the infrastructure layer for real-time AI applications.

Monetization Strategy

+
Usage-based pricing model with three tiers: (1) FREE TIER: 100K inference requests/month, <200ms p95 latency, community support, perfect for indie developers and prototyping. (2) PRO TIER: $99/month + $0.01 per 1K requests, <100ms p95 latency SLA, priority routing through edge network, email support, observability dashboard. Target: AI startups, agencies building client voice agents, developers launching real-time AI products. (3) ENTERPRISE TIER: Custom pricing starting at $5K/month, <50ms p95 latency SLA, dedicated GPU clusters in custom regions, proprietary model hosting (fine-tuned Llama/Mistral), white-glove support, HIPAA/SOC2 compliance, volume discounts at scale. Target: Fortune 500 companies, gaming studios, spatial computing platforms, telehealth providers. Additional revenue streams: (1) MODEL MARKETPLACE: 20% revenue share on custom models deployed by third-party developers on Latency's edge network. (2) PROFESSIONAL SERVICES: $250/hour consulting for enterprises building complex real-time AI systems (AI phone agents, video generation pipelines, spatial computing apps). (3) INFRASTRUCTURE RESALE: Partner with cloud providers (AWS, GCP) to resell edge compute with markup, creating margin on infrastructure arbitrage. Key unit economics: Gross margin target of 70%+ (software routing is high-margin, GPU inference is 40-50% margin), CAC payback of 6 months through PLG motion, net dollar retention of 130%+ as customers scale usage. At $10M ARR, deploy first self-hosted edge nodes to improve margins and differentiation. At $50M ARR, become the default infrastructure layer for real-time AI applications, with potential acquisition by hyperscaler or IPO path as the 'Twilio for AI Inference.'

Disclaimer: This entry is an AI-assisted summary and analysis derived from publicly available sources only (news, founder statements, funding data, etc.). It represents patterns, opinions, and interpretations for educational purposes—not verified facts, accusations, or professional advice. AI can contain errors or ‘hallucinations’; all content is human-reviewed but provided ‘as is’ with no warranties of accuracy, completeness, or reliability. We disclaim all liability for reliance on or use of this information. If you are a representative of this company and believe any information is inaccurate or wish to request a correction, please click the Disclaimer button to submit a request.