Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
129 tokens/sec
GPT-4o
28 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Neural-Kernel Conditional Mean Embeddings (2403.10859v1)

Published 16 Mar 2024 in stat.ML and cs.LG

Abstract: Kernel conditional mean embeddings (CMEs) offer a powerful framework for representing conditional distribution, but they often face scalability and expressiveness challenges. In this work, we propose a new method that effectively combines the strengths of deep learning with CMEs in order to address these challenges. Specifically, our approach leverages the end-to-end neural network (NN) optimization framework using a kernel-based objective. This design circumvents the computationally expensive Gram matrix inversion required by current CME methods. To further enhance performance, we provide efficient strategies to optimize the remaining kernel hyperparameters. In conditional density estimation tasks, our NN-CME hybrid achieves competitive performance and often surpasses existing deep learning-based methods. Lastly, we showcase its remarkable versatility by seamlessly integrating it into reinforcement learning (RL) contexts. Building on Q-learning, our approach naturally leads to a new variant of distributional RL methods, which demonstrates consistent effectiveness across different environments.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (52)
  1. A distributional perspective on reinforcement learning. In International Conference on Machine Learning, pp.  449–458. PMLR, 2017.
  2. Distributional Reinforcement Learning. MIT Press, 2023. http://www.distributional-rl.org.
  3. MMD-FUSE: Learning and combining kernels for two-sample testing without data splitting. 2023. URL https://arxiv.org/abs/2306.08777.
  4. Bishop, C. M. Mixture density networks. Report NCRG/94/004, 1994.
  5. BayesIMP: Uncertainty Quantification for Causal Data Fusion. In Advances in Neural Information Processing Systems (NeurIPS), volume 34, pp.  3466–3477, 2021.
  6. Super-samples from kernel herding. In Proceedings of the 26th Conference on Uncertainty in Artificial Intelligence, pp.  109–116. AUAI Press, 2010.
  7. Distributional reinforcement learning with quantile regression. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 32, 2018.
  8. Asymptotic evaluation of certain markov process expectations for large time. Communications on Pure and Applied Mathematics, 28(1):1–47, 1975.
  9. Uci machine learning repository, 2017. URL http://archive.ics.uci.edu/ml.
  10. Neural spline flows. In Advances in Neural Information Processing Systems, pp.  7509–7520, 2019.
  11. Dimensionality reduction for supervised learning with reproducing kernel hilbert spaces. In Journal of Machine Learning Research, volume 5, pp.  73–99, 2004.
  12. Kernel measures of conditional dependence. In Advances in Neural Information Processing Systems, pp.  489–496, 2008.
  13. Kernel Bayes’ rule: Bayesian inference with positive definite kernels. Journal of Machine Learning Research, 14(82):3753–3783, 2013.
  14. Regularization theory and neural networks architectures. Neural Computation, 7(2):219–269, 1995. doi: 10.1162/neco.1995.7.2.219.
  15. Kernel methods for measuring independence. In Journal of Machine Learning Research, volume 6, pp.  2075–2129, 2005.
  16. A kernel two-sample test. Journal of Machine Learning Research, 13(25):723–773, 2012.
  17. Conditional mean embeddings as regressors. In Proceedings of the 29th International Conference on Machine Learning, pp.  1823–1830. Omnipress, 2012.
  18. CARD: Classification and regression diffusion models. In Advances in Neural Information Processing Systems, 2022.
  19. Gaussian processes for big data. In Proceedings of the Conference on Uncertainty in Artificial Intelligence, pp.  282–290, 2013.
  20. Bayesian learning of conditional kernel mean embeddings for automatic likelihood-free inference. In Proceedings of the Twenty-Second International Conference on Artificial Intelligence and Statistics, volume 89 of Proceedings of Machine Learning Research, pp.  2631–2640. PMLR, 2019.
  21. Recovering Distributions from Gaussian RKHS Embeddings. In Kaski, S. and Corander, J. (eds.), Proceedings of the Seventeenth International Conference on Artificial Intelligence and Statistics, volume 33 of Proceedings of Machine Learning Research, pp.  457–465, Reykjavik, Iceland, 2014. PMLR.
  22. The multiquadric kernel for moment-matching distributional reinforcement learning. Transactions on Machine Learning Research, 2023. ISSN 2835-8856.
  23. Robust kernel density estimation. Journal of Machine Learning Research, 13(82):2529–2565, 2012.
  24. Adam: A method for stochastic optimization. In International Conference on Learning Representations, 2015.
  25. Decoupled weight decay regularization. In International Conference on Learning Representations, 2019.
  26. Benchmarking simulation-based inference. In Proceedings of The 24th International Conference on Artificial Intelligence and Statistics, volume 130 of Proceedings of Machine Learning Research, pp.  343–351. PMLR, 2021.
  27. Overcoming limitations of mixture density networks: A sampling and fitting framework for multimodal future prediction. In IEEE International Conference on Computer Vision and Pattern Recognition, 2019.
  28. Spectral normalization for generative adversarial networks. In International Conference on Learning Representations, 2018.
  29. Human-level control through deep reinforcement learning. Nature, 518(7540):529–533, 2015.
  30. Kernel mean embedding of distributions: A review and beyond. Foundations and Trends® in Machine Learning, 10(1-2):1–141, 2017.
  31. Counterfactual mean embeddings. Journal of Machine Learning Research, 22(162):1–71, 2021.
  32. Distributional reinforcement learning via moment matching. Proceedings of the AAAI Conference on Artificial Intelligence, 35(10):9144–9152, 2021.
  33. Revisiting rainbow: Promoting more insightful and inclusive deep reinforcement learning research. In Proceedings of the 38th International Conference on Machine Learning, Proceedings of Machine Learning Research. PMLR, 2021.
  34. Fast ϵitalic-ϵ\epsilonitalic_ϵ-free inference of simulation models with bayesian conditional density estimation. In Advances in Neural Information Processing Systems, volume 29, pp.  1028–1036, 2016.
  35. Normalizing flows for modeling and inference. Journal of Machine Learning Research, 22(57):1–64, 2021.
  36. Conditional distributional treatment effect with kernel conditional mean embeddings and u-statistic regression. In Proceedings of 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pp.  8401–8412. PMLR, 2021.
  37. Puterman, M. L. Markov decision processes: Discrete stochastic dynamic programming. John Wiley & Sons, 2014.
  38. Variational inference with normalizing flows. In Proceedings of The 32nd International Conference on Machine Learning, volume 37, pp.  1530–1538. PMLR, 2015.
  39. Saitoh, S. Integral transforms, reproducing kernels and their applications. Pitman research notes in mathematics series ; 369. Chapman and Hall/CRC, 1997.
  40. Prioritized experience replay. In International Conference on Learning Representations, 2016.
  41. Kernel instrumental variable regression. In Advances in Neural Information Processing Systems, volume 32, pp.  4595–4607, 2019.
  42. Hilbert space embeddings of conditional distributions with applications to dynamical systems. In International Conference on Machine Learning, pp.  961–968, 2009.
  43. Kernel embeddings of conditional distributions: A unified kernel framework for nonparametric inference in graphical models. IEEE Signal Processing Magazine, 30(4):98–111, 2013.
  44. Hilbert space embeddings and metrics on probability measures. Journal of Machine Learning Research, 11:1517–1561, 2010.
  45. normflows: A pytorch package for normalizing flows. Journal of Open Source Software, 8(86):5361, 2023.
  46. Conditional density estimation via least-squares density ratio estimation. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, volume 9 of Proceedings of Machine Learning Research, pp.  781–788. PMLR, 2010.
  47. Gymnasium, 2023. URL https://zenodo.org/record/8127025.
  48. Deep reinforcement learning with double q-learning. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, pp.  2094–2100. AAAI Press, 2016.
  49. SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nature Methods, 17:261–272, 2020.
  50. Q-learning. Machine learning, 8(3-4):279–292, 1992.
  51. Learning deep features in instrumental variable regression. In International Conference on Learning Representations, 2021.
  52. Importance weighted kernel Bayes’ rule. In Proceedings of the 39th International Conference on Machine Learning, volume 162 of Proceedings of Machine Learning Research, pp.  24524–24538. PMLR, 2022.
Citations (3)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com