
Briefing
This paper addresses the critical problem of centralized Large Language Model (LLM) inference, which poses significant risks including privacy loss, restricted access, single points of failure, and monopolistic control. It proposes VeriLLM, a novel publicly verifiable protocol that achieves security under a one-honest-verifier assumption, significantly reducing the stringent requirements of traditional consensus or the prohibitive computational costs of general-purpose zero-knowledge proofs for LLMs. The core breakthrough lies in a hybrid verification mechanism that leverages Merkle commitments, a Verifiable Random Function (VRF) for unpredictable sampling, and an escalatory dispute resolution protocol, ensuring integrity with near-negligible verification overhead. This new theory fundamentally transforms the landscape of AI infrastructure by enabling truly trustworthy and scalable decentralized LLM services, fostering a more transparent and resilient ecosystem.

Context
Prior to this research, the prevailing paradigm for LLM inference was predominantly centralized, leading to inherent vulnerabilities and concentrated power. While decentralized inference offered a promising alternative, the challenge of ensuring output verifiability in permissionless networks remained largely unsolved. Existing approaches, such as cryptographic proof systems (e.g.
Zero-Knowledge Machine Learning), incurred computational overheads orders of magnitude greater than native inference, rendering them economically infeasible for high-throughput LLM serving. Conversely, consensus-style schemes relied on brittle assumptions like honest majorities or strong synchrony, limiting their practical applicability for public verifiability in dynamic blockchain environments.

Analysis
VeriLLM’s core mechanism for verifiable decentralized LLM inference integrates a commit-then-sample-and-check pipeline with game-theoretic incentives. The system commits all intermediate hidden states to Merkle trees, with root hashes recorded on-chain, creating tamper-evident logs. A Verifiable Random Function (VRF) then unpredictably selects specific positions for off-chain empirical re-computation by designated verifiers. A dispute resolution protocol escalates to zero-knowledge proofs if inconsistencies are challenged, ensuring definitive adjudication.
This approach leverages the efficient “Prefill” phase of LLM inference, reducing verification costs to approximately 1% of the full inference. Unlike previous methods, VeriLLM operates securely under a one-honest-verifier assumption, moving beyond the need for an honest majority. It also employs an isomorphic inference-verification network where all GPU workers can perform both roles indistinguishably, enhancing security by preventing strategic misbehavior and improving resource utilization.

Parameters
- Core Concept ∞ Publicly Verifiable Decentralized LLM Inference
- New System/Protocol ∞ VeriLLM
- Key Authors ∞ Ke Wang, Felix Qu, Zishuo Zhao, Libin Xia, Chris Tong, Lynn Ai, Eric Yang
- Verification Cost ∞ Approximately 1% of underlying inference
- Security Assumption ∞ One-honest-verifier
- Core Mechanism ∞ Commit-then-sample with Merkle trees and VRF

Outlook
This research lays a critical foundation for the future of decentralized AI, potentially unlocking a new generation of trustworthy and scalable on-chain AI agents and private AI services. In the next three to five years, VeriLLM’s principles could enable the widespread adoption of decentralized LLM infrastructure, fostering a more competitive and resilient innovation landscape by democratizing access to foundational AI resources. Future research will explore enhancing collusion resistance with more advanced zero-knowledge proofs, developing automated calibration for diverse hardware, and devising cryptographic mitigations against potential scheduler censorship.