Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
175 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning with Density Matrices and Random Features (2102.04394v5)

Published 8 Feb 2021 in cs.LG, cs.AI, and quant-ph

Abstract: A density matrix describes the statistical state of a quantum system. It is a powerful formalism to represent both the quantum and classical uncertainty of quantum systems and to express different statistical operations such as measurement, system combination and expectations as linear algebra operations. This paper explores how density matrices can be used as a building block for machine learning models exploiting their ability to straightforwardly combine linear algebra and probability. One of the main results of the paper is to show that density matrices coupled with random Fourier features could approximate arbitrary probability distributions over $\mathbb{R}n$. Based on this finding the paper builds different models for density estimation, classification and regression. These models are differentiable, so it is possible to integrate them with other differentiable components, such as deep learning architectures and to learn their parameters using gradient-based optimization. In addition, the paper presents optimization-less training strategies based on estimation and model averaging. The models are evaluated in benchmark tasks and the results are reported and discussed.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (15)
  1. \bibcommenthead
  2. Bishop CM (2006) Pattern recognition and machine learning (information science and statistics)
  3. Chu W, Ghahramani Z (2005) Gaussian Processes for Ordinal Regression. Journal of Machine Learning Research 6:1019–1041. URL http://www.jmlr.org/papers/volume6/chu05a/chu05a.pdf
  4. Efron B (1992) Bootstrap methods: another look at the jackknife. In: Breakthroughs in statistics. Springer, p 569–593
  5. Ji Z, Telgarsky M (2019) Polylogarithmic width suffices for gradient descent to achieve arbitrarily small test error with shallow relu networks. arXiv preprint arXiv:190912292
  6. McNeil BJ, Hanley JA (1984) Statistical approaches to the analysis of receiver operating characteristic (roc) curves. Medical decision making 4(2):137–150
  7. Nadaraya EA (1964) Some new estimates for distribution functions. Theory of Probability & Its Applications 9(3):497–500
  8. Parzen E (1962) On estimation of a probability density function and mode. The annals of mathematical statistics 33(3):1065–1076
  9. Rosenblatt M (1956) Remarks on some nonparametric estimates of a density function. Ann Math Statist 27(3):832–837. 10.1214/aoms/1177728190, URL https://doi.org/10.1214/aoms/1177728190
  10. Schuld M (2018) Supervised learning with quantum computers. Springer
  11. Tang E (2019a) A quantum-inspired classical algorithm for recommendation systems. In: Proceedings of the 51st Annual ACM SIGACT Symposium on Theory of Computing, pp 217–228
  12. Tang E (2019b) Quantum-inspired classical algorithms for principal component analysis and supervised clustering. 1811.00414
  13. Tiwari P, Melucci M (2019) Towards a quantum-inspired binary classifier. IEEE Access 7:42,354–42,372. 10.1109/ACCESS.2019.2904624
  14. Von Neumann J (1927) Wahrscheinlichkeitstheoretischer aufbau der quantenmechanik. Nachrichten von der Gesellschaft der Wissenschaften zu Göttingen, Mathematisch-Physikalische Klasse 1927:245–272
  15. Wolf L (2006) Learning using the born rule. Tech. rep., Massachusetts Institute of Technology Computer Science and Artificial Intelligence Laboratory
Citations (28)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets