Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

The Principle of Uncertain Maximum Entropy (2305.09868v4)

Published 17 May 2023 in cs.IT, cs.CV, cs.LG, and math.IT

Abstract: The principle of maximum entropy is a well-established technique for choosing a distribution that matches available information while minimizing bias. It finds broad use across scientific disciplines and in machine learning. However, the principle as defined by is susceptible to noise and error in observations. This forces real-world practitioners to use relaxed versions of the principle in an ad hoc way, negatively impacting interpretation. To address this situation, we present a new principle we call uncertain maximum entropy that generalizes the classic principle and provides interpretable solutions irrespective of the observational methods in use. We introduce a convex approximation and expectation-maximization based algorithm for finding solutions to our new principle. Finally, we contrast this new technique with two simpler generally applicable solutions theoretically and experimentally show our technique provides superior accuracy.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (17)
  1. E. T. Jaynes, Information theory and statistical mechanics. Physical Review, 106(4), 620 (1957).
  2. M. Marsili, I. Mastromatteo, On the criticality of inferred models, Journal of Statistical Mechanics: Theory and Experiment, 10, P10012 (2011).
  3. S. Singh, T. Joachims, A maximum entropy model for part-of-speech tagging, in Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, (2018), pp. 390–399.
  4. A statistical explanation of MaxEnt for ecologists, Diversity and Distributions, vol. 17, no. 1, 43–57 (2011).
  5. S. Boyd, L. Vandenberghe, Convex Optimization. (Cambridge University press, 2004).
  6. N. Wu, The maximum entropy method. (Springer Science & Business Media, 2012).
  7. G. J. McLachlan, T. Krishnan, The EM algorithm and extensions. (John Wiley & Sons, 2007).
  8. A Hierarchical Bayesian Process for Inverse RL in Partially-Controlled Environments. in Proceedings of the 21st International Conference on Autonomous Agents and Multiagent Systems, (2022), pp. 145–153.
  9. Inferring light fields from shadows. in Proceedings of the IEEE conference on computer vision and pattern recognition, (2018), pp. 6267-06275.
  10. A. Torralba, W. T. Freeman, Accidental pinhole and pinspeck cameras: Revealing the scene outside the picture. in Proceedings of the IEEE conference on computer vision and pattern recognition, (2012), pp. 374–381.
  11. J. Steinhardt, P. Liang, Adaptivity and Optimism : An Improved Exponentiated Gradient Algorithm. in International Conference on Machine Learning, (2014).
  12. J. Kivinen, M. K. Warmuth, Exponentiated gradient versus gradient descent for linear predictors. Information and Computation 132(1), 1-63, (1997).
  13. K. Bogert, M. Kothe, Principle of Uncertain Maximum Entropy Data Sets. https://doi.org/10.17605/OSF.IO/NR59D (2023)
  14. Expectation-maximization for inverse reinforcement learning with hidden data. in Proceedings of the 2016 International Conference on Autonomous Agents & Multiagent Systems, (2016) pp. 1034–1042.
  15. The latent maximum entropy principle. ACM Transactions on Knowledge Discovery from Data, 6(2), (2012).
  16. J. Shore, R. Johnson, Axiomatic derivation of the principle of maximum entropy and the principle of minimum cross-entropy IEEE Transactions on Information Theory, 26(1), 26-37 (1980).
  17. M. Kárnỳ Minimum expected relative entropy principle in 2020 European Control Conference, (2020) pp. 35–40.

Summary

We haven't generated a summary for this paper yet.