Optimizing ZK-SNARKs by Minimizing Expensive Cryptographic Group Elements
Polymath redesigns zk-SNARKs by shifting proof composition from $mathbb{G}_2$ to $mathbb{G}_1$ elements, significantly reducing practical proof size and on-chain cost.
Horizontally Scalable zkSNARKs via Proof Aggregation Framework
This framework achieves horizontal zkSNARK scalability by distributing large computations for parallel proving, then aggregating results into a single succinct proof.
Zero-Knowledge Proof of Training Secures Federated Learning Consensus
ZKPoT uses zk-SNARKs to verify model contributions privately, eliminating the trade-off between decentralized AI privacy and consensus efficiency.
Fast Zero-Knowledge Proofs for Verifiable Machine Learning via Circuit Optimization
The Constraint-Reduced Polynomial Circuit (CRPC) dramatically lowers ZKP overhead for matrix operations, making private, verifiable AI practical.
Generic Folding Scheme Enables Efficient Non-Uniform Verifiable Computation
Protostar introduces a generic folding scheme for special-sound protocols, drastically reducing recursive overhead for complex, non-uniform verifiable computation.
Zero-Knowledge Proof of Training Secures Decentralized AI Consensus
A new Zero-Knowledge Proof of Training (ZKPoT) consensus mechanism leverages zk-SNARKs to cryptographically verify model performance, eliminating Proof-of-Stake centralization and preserving data privacy in decentralized machine learning.
Zero-Knowledge Proof of Training Secures Decentralized Federated Learning Consensus
ZKPoT uses zk-SNARKs to verify decentralized model accuracy without revealing private data, solving the efficiency-privacy trade-off in federated learning.
Zero-Knowledge Proof of Training Secures Decentralized Federated Learning
ZKPoT consensus uses zk-SNARKs to verify machine learning contributions privately, resolving the privacy-verifiability trade-off for decentralized AI.
Incremental Proofs Maintain Constant-Size Sequential Work for Continuous Verification
This new cryptographic primitive enables constant-size proofs for arbitrarily long sequential computations, fundamentally solving the accumulated overhead problem for VDFs.
