Infra

We build autonomous agents that reason over distributed, high-stakes data (Bio & Finance) without ever centralizing it. We replace trust with cryptography and replace correlation with causal/structural reasoning.

If you are interested in any of this project, please don’t feel free to contact me. :)

Infrastructure

We replace “Trust” (Institutional Reputation) with “Math” (Cryptography) and “Incentives” (Game Theory).

Proof-of-Learning (PoL): The ZK-Training Protocol

The Moonshot: To create the “SSL/TLS” of Machine Learning—a standard protocol where a node proves it trained a model correctly on private data without revealing a single byte of that data.

The Hard Problem: Currently, Zero-Knowledge (ZK) protocols are slow and mostly work for inference (proving a model ran). Proving training (Backpropagation, SGD) involves millions of floating-point multiplications, which is computationally prohibitive for current ZK-SNARKs.

Decentralized-MoE: The Inter-Institutional Mixture of Experts

The Moonshot: A “Global Brain” where the “Neurons” are hospitals. A single logical LLM where the parameters are physically distributed across 1,0000 institutions.

The Hard Problem: Latency. If every token generation requires a network call to 5 different hospitals, the model will be unusably slow. Security. LLM can have biased and can even be weaponized.

Relevant Publications:

  1. Cheng, Zehua, Rui Sun, Jiahao Sun, and Yike Guo. (2025). “Scaling Decentralized Learning with FLock.” In Procedings of 2025 IEEE/WIC International Conference on Web Intelligence and Intelligent Agent Technology (WI-IAT). [WI-IAT 2025 - Best Application Award] [Paper]
  2. Cheng, Zehua, Manying Zhang, Jiahao Sun, and Wei Dai. “On weaponization-resistant large language models with prospect theoretic alignment.” In Proceedings of the 31st International Conference on Computational Linguistics, pp. 10309-10324. 2025. [Paper]

Nash-Equilibrium Data Valuation: The Shapley Protocol

The Moonshot: Solving the “Free Rider Problem.” In current federated learning, a hospital that contributes junk data gets the same model as a hospital that contributes gold. This protocol automatically pays data owners based on the exact marginal utility of their contribution.

The Hard Problem: Calculating the true Shapley Value is $O(N!)$—impossible to compute for thousands of nodes.

The Holographic Gradient: Bandwidth-Efficient Learning

The Moonshot: Training a 1 Trillion parameter model on a network of consumer-grade internet connections (e.g., connecting Starlink satellites or home GPUs).

The Hard Problem: Bandwidth. Sending 1TB of gradients every second is impossible for hospitals or home nodes.