RFD 016 idea
Zero-Knowledge Proofs for LLM Inference
The Idea
Use zero-knowledge proofs to verify that inference was performed correctly without revealing model weights or inputs. Recent work (zkLLM, zkTorch, zkGPT) has made progress on this, but practical deployment for frontier models remains distant.
Open attack: “Hollow-LLM” shows that computationally trivial weights can pass ZK verification, undermining the guarantee that a real model was used.
Why It Matters
ZKPs offer the strongest possible verification guarantee—mathematical proof rather than physical security. If practical, they would obsolete most hardware-based verification. Understanding their limits helps prioritize the verification research agenda.
Open Questions
- What’s the current overhead for ZK-verified inference?
- Can Hollow-LLM style attacks be prevented?
- Is there a hybrid approach (ZKP + hardware) that’s practical sooner?
References
- zkLLM
- zkTorch
- zkGPT
- ZKML optimizing system
- Hollow-LLM Attack (unpublished, Gong et al.)