← ALL RFDS
RFD 017 discussing

Inference-Only Verification Package

AUTHOR Romeo Dean CREATED 2024-11-27
verificationhardwareinferencephysical-security

The Idea

AI training and inference have different hardware requirements: training coordinates across entire clusters via high-speed interconnect, while even frontier models can serve inference from single servers. An inference-only verification package exploits this asymmetry by physically removing interconnect from clusters, then verifying that individual servers are doing legitimate inference rather than participating in distributed training.

The verification problem reduces to: (1) ensuring interconnect isn’t reintroduced (physical monitoring, tamper-evident cameras), and (2) ensuring each server is doing whitelisted inference, not contributing to a rogue training run that coordinates over the internet (e.g., expert iteration with gradient updates happening elsewhere).

Why It Matters

If you can verifiably distinguish inference-only clusters from training-capable clusters, you can enforce compute governance at the hardware level. This enables scenarios like: mutual US-China verification that neither is doing undeclared frontier training, or domestic enforcement that licensed inference providers aren’t secretly training. The key insight is that interconnect removal is fast, cheap, and hard to reverse covertly.

Architecture Overview

┌─────────────────────────────────────────────────────────┐
│                      DATACENTER                         │
│                                                         │
│  ┌─────────┐  ┌─────────┐  ┌─────────┐                 │
│  │ Server  │  │ Server  │  │ Server  │  ...            │
│  │(no NVMe)│  │(no NVMe)│  │(no NVMe)│                 │
│  └────┬────┘  └────┬────┘  └────┬────┘                 │
│       │            │            │                       │
│  ┌────┴────┐  ┌────┴────┐  ┌────┴────┐                 │
│  │   SNG   │  │   SNG   │  │   SNG   │  (1:1 ratio)    │
│  └────┬────┘  └────┴────┘  └────┬────┘                 │
│       │            │            │                       │
│       └────────────┼────────────┘                       │
│                    │                                    │
│              ┌─────┴─────┐                              │
│              │    SRC    │  (Secure ReComputer)        │
│              └─────┬─────┘                              │
│                    │                                    │
└────────────────────┼────────────────────────────────────┘

                 Internet

Key components:

  • Servers: GPU nodes with interconnect removed, NVMe removed, airgapped except through SNG
  • SNG (Secure Network Gateway): Logs all I/O, samples for verification, enforces whitelisted models
  • SRC (Secure ReComputer): Receives sampled I/O pairs, does proof-of-work recomputation to verify inference

Server Isolation Requirements

Transform servers into “memoryless individual compute units”:

RequirementImplementation
No interconnectPhysically remove NVLink/InfiniBand
No persistent stateRemove all NVMe/SSD from servers
AirgappedOnly network path is through SNG
Ephemeral modelsSecure boot uploads whitelisted weights on each power cycle

Secure Network Gateway (SNG)

Logical Properties

  1. Tamper = Brick: Device is (OFF ∨ untampered), else non-functional
  2. Upload Mode → Serving Mode: On boot, accepts model weights, checks hash against whitelist. After S bits (where S = model size), transitions to SERVING MODE. Hash mismatch → power kill → restart
  3. Probabilistic Sampling: In SERVING MODE, stores+tags all I/O pairs, deletes with probability 1-ε, sends remainder to SRC

Physical Components

ComponentFunction
IROM (Immutable ROM)Hardcoded secure boot protocol, can’t be overwritten
NVMe (~1TB)Stores whitelist hashes + rolling I/O buffer
Hardware UIDDerived key unknown to manufacturer; tampering breaks UID → breaks encryption → brick
TRNGTrue random number generator for sampling
Input NIC pairReceive from internet → send to server
Output NIC pairReceive from server → send to internet
Volatile Bit CounterTracks upload progress, zeroes on power loss
Input copier+taggerCopies inputs, tags with counter state
Output copier+taggerCopies outputs, matches to input tags
Coordinator + AES engineManages encryption, validates hashes
SRC NIC + samplerSends sampled pairs to SRC

Property Implementation

  • Property 1: UID breaks on tamper → encryption fails → I/O path fails → brick
  • Property 2: Counter fresh on boot, asymmetric encryption on input path, decrypter checks hashes, counter interrupts at S (→ SERVING), hash mismatch interrupts earlier (→ power kill)
  • Property 3: Input path tags all inputs (trusts inputter for delimiter), output path saves + receives tag, TRNG samples, non-deleted pairs → SRC

Secure ReComputer (SRC)

Logical Properties

  1. Tamper = Brick: Same as SNG
  2. Request Types:
    • INPUT = (input_text, model_id) → routes to SNG serving that model
    • ADMIN_REQUEST = (auth, instructions) → [TODO: specify admin protocol]
  3. Proof Requests: Receives (input, output, model_id) from SNGs, does recomputation
  4. Verification: PASS → continue; FAIL → alarms + shutdown

Open Design Questions

  • What happens if no SNG is serving requested model_id?
  • Admin request authentication and authorization model?
  • How to handle non-deterministic inference (sampling temperature, batching)?

Datacenter Physical Security

MeasurePurpose
Vetted guards + entry/exit protocolsPrevent unauthorized physical access
Analog sensorsConfirm power state (cluster on/off)
Radio signal sensorsDetect covert RF communication bypassing network tap
Seals + camerasMake tampering detectable

Deployment Timeline & Costs

Assumptions for 100M H100-equivalents (~40% of global AI compute in early 2029):

MetricValue
Avg H100e per server400
Total servers250K
Team composition4 people (2 US, 2 Chinese)
Servers per team-hour1
Total teams1,000
Wall clock time5 weeks (50-hr weeks)
Manufacturing cost$100M ($400/server)
Labor cost$100M (1M hrs × $100/hr)
Total installation cost$200M
Expected monthly revenue from 100M H100e$100B
Installation cost as % of monthly revenue0.2%

Rollout schedule (early 2029):

PhaseCluster SizeCumulative H100eTime
1>10M H100e clusters50M2.5 weeks
2>1M H100e clusters100M+2.5 weeks
3>100K H100e clusters100M+5 weeks
Total~90% global compute3 months

Clusters 1K–100K H100e (~9% of compute) can be powered off during installation, verified later. Result: 99% of global compute either inference-only or offline after 3-month installation phase.

Attack Vectors & Mitigations

AttackMitigation
Reintroduce interconnectPhysical monitoring, tamper-evident cameras, guards
Distributed training over internetNetwork gateway logs all I/O, random recomputation proves inference
Expert iteration (inference for RL, gradients elsewhere)Whitelist models, periodic power flicker + memory wipe + secure reboot
Rogue model persists in memoryRemove NVMe, periodic forced reboots with fresh model upload
Covert RF communicationRadio signal sensors, triangulation
Tamper with SNG/SRCHardware UID, tamper = brick

Open Questions

  • Nondeterminism: How to verify inference when outputs depend on sampling temperature, batch composition, or timing? Require deterministic mode? Accept probabilistic verification?
  • Upload bandwidth limits: How long does secure boot take for a 1TB model? What’s the attack surface during upload?
  • Proof-of-work coverage: What fraction of I/O pairs must be recomputed for statistical confidence? How does this scale with cluster size?
  • Preparation phase: If US/China haven’t pre-manufactured verification devices, how long is the buffer period? What interim measures (interconnect removal only, human intelligence) provide deterrence?
  • Admin protocol: Who can issue ADMIN_REQUESTs? What instructions are permitted? How to prevent abuse?

References