BasedAI: A decentralized P2P network for Zero Knowledge Large Language Models (ZK-LLMs) (2403.01008v1)
Abstract: BasedAI is a distributed network of machines which introduces decentralized infrastructure capable of integrating Fully Homomorphic Encryption (FHE) with any LLM connected to its network. The proposed framework embeds a default mechanism, called "Cerberus Squeezing", into the mining process which enables the transformation of a standard LLMs into encrypted zero-knowledge LLMs, or "ZK-LLMs", leveraging insights from generative adversarial networks for data privacy. This novel quantization mechanism empowers BasedAI miners to process and respond to prompts derived from User interaction with LLMs without the need for decrypting either the queries or their corresponding responses. The introduction of Cerberus Squeezing significantly improves performance degradation caused by quantized functions in current FHE-compliant computing environments by proactively optimizing calls between users, miners, and validators.
- Generative Adversarial Nets. In Advances in Neural Information Processing Systems. https://papers.nips.cc/paper/5423-generative-adversarial-nets.pdf
- Gentry, C. (2009). A Fully Homomorphic Encryption Scheme. PhD thesis, Stanford University. https://crypto.stanford.edu/craig/
- Recursive Composition and Bootstrapping for SNARKs and Proof-Carrying Data. In Proceedings of the Forty-fifth Annual ACM Symposium on Theory of Computing. ACM. https://dl.acm.org/doi/10.1145/2488608.2488706
- Temporal Fusion Transformers for Interpretable Multi-horizon Time Series Forecasting. International Journal of Forecasting. https://arxiv.org/abs/1912.09363
- SNARKs for C: Verifiable Computation with Efficient Proofs. In Proceedings of the 34th Annual Cryptology Conference. https://eprint.iacr.org/2013/507.pdf
- Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding. In International Conference on Learning Representations (ICLR). https://openreview.net/forum?id=rJqFGTslg
- Krishnamoorthi, R. (2018). Quantizing deep convolutional networks for efficient inference: A whitepaper. arXiv preprint arXiv:1806.08342. https://arxiv.org/abs/1806.08342
- Attention Is All You Need. In Advances in Neural Information Processing Systems. https://arxiv.org/abs/1706.03762
- Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR). https://ieeexplore.ieee.org/document/8578799m
Sponsor
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.