Why Cloud + Robot

The economics of cloud computing have made it practical to run substantial computation far from the physical robot, while network latency advances (5G, WiFi 6E) have expanded what is feasible to offload. Modern robot deployments use a three-tier architecture: onboard compute for safety-critical real-time control, edge servers for low-latency inference and local data buffering, and cloud for model training, fleet management, and teleoperation relay.

The key architectural principle is: everything time-critical runs locally; everything compute-intensive runs in the cloud. The boundary between local and cloud shifts as network latency decreases and edge hardware gets more powerful.

Architecture Layers

Layer 1: Robot Onboard

The robot's onboard computer runs safety-critical, hard real-time tasks. Nothing that affects physical safety should depend on network connectivity.

  • ROS 2 control node: Joint trajectory execution, safety monitoring, e-stop handling. Must run at 500–1,000 Hz with <1 ms jitter. PREEMPT_RT kernel or Xenomai required for deterministic scheduling.
  • Safety controller: Joint limit enforcement, collision monitoring, workspace boundary checks. This is the last line of defense and must never be outsourced to the cloud.
  • Local state estimation: Proprioception, IMU fusion, base odometry. These are prerequisites for any control or perception computation.
  • Hardware: Most collaborative arms include an onboard controller (UR Control Box, Franka FCI). For mobile robots, NVIDIA Jetson AGX Orin (275 TOPS INT8) is the standard edge module for AI-enhanced perception and policy execution.

Layer 2: Edge Server

A local server (co-located in the facility or robot room) handles compute that requires GPU acceleration but must stay below ~50 ms latency.

  • Policy inference: Running ACT, diffusion policy, or OpenVLA fine-tuned models. An NVIDIA RTX 4090 or A100 server handles 5–20 robot inference streams simultaneously at 50 Hz.
  • Local data buffer: Ring-buffer of recent episodes (24–48 hours of operation) for fast retrieval and re-training without cloud round-trips. NVMe RAID recommended for write throughput.
  • Real-time perception: Object detection, pose estimation, depth processing for policy inputs. Should run <20 ms end-to-end.
  • Hardware recommendation: A workstation-class server (NVIDIA A4000 or RTX 4090, 64 GB RAM, 2 TB NVMe) handles a 5–10 robot cluster at approximately $8,000–$15,000 in hardware.

Layer 3: Cloud

Cloud services handle latency-tolerant, compute-intensive workloads and fleet-scale functions.

  • Model training: Training or fine-tuning policy models on accumulated demonstration data. This is the primary use case for cloud GPU clusters (AWS p4d, GCP A100 pods, Lambda Labs).
  • Fleet dashboard: Real-time monitoring of all robots in a deployment — throughput, error rates, joint health, task completion. Latency-tolerant (seconds) aggregation is acceptable.
  • Teleoperation relay: For remote operator teleoperation, the cloud provides STUN/TURN infrastructure for NAT traversal and WebRTC signaling. Video and command streams flow peer-to-peer when possible; through cloud relay when NAT prevents direct connection.
  • Long-term data lake: All episode data eventually flows to cloud object storage (S3 or GCS) for long-term retention, training dataset construction, and compliance.
  • Model serving (latency-tolerant tasks): VLM-based task planning, natural language instruction parsing, and scene understanding — where 300–1,000 ms latency is acceptable.

Latency Budget by Function

FunctionMax LatencyRequired TierNotes
E-stop / safety halt<1 msOnboardNever cloud-dependent
Joint trajectory execution<5 msOnboardHard real-time required
Policy inference (manipulation)10–50 msEdgeContact tasks need <20 ms
Object detection / pose20–80 msEdgeDepends on task speed
Teleoperation video stream30–80 msEdge / P2PAbove 100 ms degrades operator
Task planning (VLM)300–2,000 msCloudLatency-tolerant planning
Fleet monitoring1–10 sCloudAggregation acceptable
Model trainingHoursCloudAsync, not latency-sensitive

Teleoperation Streaming Architecture

Remote robot teleoperation requires bidirectional real-time streaming: video from robot to operator, commands from operator to robot. The recommended stack:

  • Video stream: WebRTC with H.264 or AV1 codec. NVENC hardware encoding on the edge server reduces encoding latency to <10 ms. Adaptive bitrate targeting 10–30 Mbps for HD stereo video.
  • Command stream: WebSocket over TLS for operator controller pose data. JSON or msgpack encoding. 100 Hz update rate. Priority queue so newest commands displace old ones during transient network congestion.
  • NAT traversal: STUN (Coturn or managed STUN service) for most operators on residential internet. TURN relay required for operators behind symmetric NAT or strict corporate firewalls. Budget ~$0.10/GB for TURN relay traffic.
  • Operator latency monitoring: Display end-to-end round-trip latency to the operator in the headset HUD. Alert when latency exceeds 120 ms — this is the threshold where many operators find manipulation uncomfortably imprecise.

Data Pipeline: Robot to Training Cluster

  • Episode capture: At episode completion, the edge server writes an HDF5 file (observations + actions + metadata) to local NVMe buffer.
  • Background upload: A daemon monitors the local buffer and uploads completed episodes to S3/GCS using multipart upload. Bandwidth-throttled to avoid interfering with robot operation. Typical episode size: 100–500 MB.
  • Ingest validation: Cloud-side Lambda/Cloud Function runs schema validation and automated quality checks (see quality metrics article) on each uploaded episode. Failed episodes are quarantined to a review queue.
  • Training trigger: A dataset version is tagged when a new batch of validated episodes is available. This triggers a training job on the GPU cluster using the new data.
  • Model update OTA: After training completes and offline evaluation passes, the new model is pushed to edge servers via an OTA update system. A/B testing framework deploys the new model to 10% of the fleet first, monitors success rates, then rolls out to 100%.

Cloud Provider Comparison

ProviderRobot-Specific ServiceGPU OptionsIoT IntegrationBest For
AWSAWS IoT Greengrass, RoboMakerp4d, p3, g4dnAWS IoT Core, GreengrassEnterprise, broad ecosystem
GCPNone (general compute)A100, H100, TPUv4Cloud IoT Core (deprecated)ML training, BigQuery analytics
AzureAzure IoT Hub, Digital TwinsNC A100, ND H100IoT Hub, IoT EdgeMicrosoft enterprise, mixed reality
Custom (Lambda Labs)NoneA100, H100 on-demandN/AGPU-only cost optimization

The SVRC platform provides a managed implementation of this architecture — edge agents for robot-side data collection, cloud pipeline for episode processing, and a web dashboard for fleet monitoring — without requiring teams to build infrastructure from scratch.