Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
162 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Neural Networks with (Low-Precision) Polynomial Approximations: New Insights and Techniques for Accuracy Improvement (2402.11224v2)

Published 17 Feb 2024 in cs.LG and cs.CR

Abstract: Replacing non-polynomial functions (e.g., non-linear activation functions such as ReLU) in a neural network with their polynomial approximations is a standard practice in privacy-preserving machine learning. The resulting neural network, called polynomial approximation of neural network (PANN) in this paper, is compatible with advanced cryptosystems to enable privacy-preserving model inference. Using ``highly precise'' approximation, state-of-the-art PANN offers similar inference accuracy as the underlying backbone model. However, little is known about the effect of approximation, and existing literature often determined the required approximation precision empirically. In this paper, we initiate the investigation of PANN as a standalone object. Specifically, our contribution is two-fold. Firstly, we provide an explanation on the effect of approximate error in PANN. In particular, we discovered that (1) PANN is susceptible to some type of perturbations; and (2) weight regularisation significantly reduces PANN's accuracy. We support our explanation with experiments. Secondly, based on the insights from our investigations, we propose solutions to increase inference accuracy for PANN. Experiments showed that combination of our solutions is very effective: at the same precision, our PANN is 10% to 50% more accurate than state-of-the-arts; and at the same accuracy, our PANN only requires a precision of 2{-9} while state-of-the-art solution requires a precision of 2{-12} using the ResNet-20 model on CIFAR-10 dataset.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (24)
  1. Autofhe: Automated adaption of cnns for efficient evaluation over fhe. Cryptology ePrint Archive, Paper 2023/162, 2023. URL https://eprint.iacr.org/2023/162. https://eprint.iacr.org/2023/162.
  2. Towards the alexnet moment for homomorphic encryption: Hcnn, thefirst homomorphic cnn on encrypted data with gpus, 2020.
  3. ngraph-he: A graph compiler for deep learning on homomorphically encrypted data, 2019.
  4. Faster cryptonets: Leveraging sparsity for real-world encrypted inference. arXiv preprint arXiv:1811.09953, 2018.
  5. Cryptonets: Applying neural networks to encrypted data with high throughput and accuracy. In International conference on machine learning, pp.  201–210. PMLR, 2016.
  6. Explaining and harnessing adversarial examples. arXiv preprint arXiv:1412.6572, 2014.
  7. Deep residual learning for image recognition, 2015.
  8. Cryptodl: Deep neural networks over encrypted data, 2017.
  9. Learning multiple layers of features from tiny images. 2009.
  10. Adversarial machine learning at scale. arXiv preprint arXiv:1611.01236, 2016.
  11. Tiny imagenet visual recognition challenge. 2015. URL https://api.semanticscholar.org/CorpusID:16664790.
  12. Optimization of homomorphic comparison algorithm on rns-ckks scheme. IEEE Access, 10:26163–26176, 2022a.
  13. Low-complexity deep convolutional neural networks on fully homomorphic encryption using multiplexed parallel convolutions. In International Conference on Machine Learning, pp.  12403–12422. PMLR, 2022b.
  14. Privacy-preserving machine learning with fully homomorphic encryption for deep neural network, 2021a.
  15. High-precision bootstrapping of rns-ckks homomorphic encryption using optimal minimax polynomial approximation and inverse sine function. In Advances in Cryptology–EUROCRYPT 2021, Part I 40, pp.  618–647. Springer, 2021b.
  16. Shufflenet v2: Practical guidelines for efficient cnn architecture design. In Proceedings of the European Conference on Computer Vision (ECCV), September 2018.
  17. Towards deep learning models resistant to adversarial attacks. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=rJzIBfZAb.
  18. Delphi: A cryptographic inference service for neural networks. In 29th USENIX Security Symposium, pp.  2505–2522, 2020.
  19. Mobilenetv2: Inverted residuals and linear bottlenecks, 2019.
  20. Deep layer aggregation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp.  2403–2412, 2018.
  21. Three mechanisms of weight decay regularization. In International Conference on Learning Representations, 2018a.
  22. mixup: Beyond empirical risk minimization. In International Conference on Learning Representations, 2018b. URL https://openreview.net/forum?id=r1Ddp1-Rb.
  23. How does mixup help with robustness and generalization? In The Ninth International Conference on Learning Representations, 2021.
  24. Exact solutions of a deep linear network. Advances in Neural Information Processing Systems, 35:24446–24458, 2022.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com