Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
144 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Versatile Energy-Based Probabilistic Models for High Energy Physics (2302.00695v5)

Published 1 Feb 2023 in cs.LG, hep-ex, hep-ph, and stat.ML

Abstract: As a classical generative modeling approach, energy-based models have the natural advantage of flexibility in the form of the energy function. Recently, energy-based models have achieved great success in modeling high-dimensional data in computer vision and natural language processing. In line with these advancements, we build a multi-purpose energy-based probabilistic model for High Energy Physics events at the Large Hadron Collider. This framework builds on a powerful generative model and describes higher-order inter-particle interactions. It suits different encoding architectures and builds on implicit generation. As for applicative aspects, it can serve as a powerful parameterized event generator for physics simulation, a generic anomalous signal detector free from spurious correlations, and an augmented event classifier for particle identification.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (71)
  1. LHC Machine. JINST, 3:S08001, 2008.
  2. Georges Aad et al. Observation of a new particle in the search for the Standard Model Higgs boson with the ATLAS detector at the LHC. Phys. Lett. B, 716:1–29, 2012.
  3. A learning algorithm for boltzmann machines. Cognitive Science, 9(1):147–169, 1985.
  4. Madgraph 5: going beyond. Journal of High Energy Physics, 2011(6), Jun 2011.
  5. IRC-Safe Graph Autoencoder for Unsupervised Anomaly Detection. Front. Artif. Intell., 5:943135, 2022.
  6. Neural machine translation by jointly learning to align and translate. CoRR, abs/1409.0473, 2015.
  7. Theory and phenomenology of two-Higgs-doublet models. Phys. Rept., 516:1–102, 2012.
  8. How to GAN LHC Events. SciPost Phys., 7(6):075, 2019.
  9. The anti-kt jet clustering algorithm. Journal of High Energy Physics, 2008(04):063–063, Apr 2008.
  10. J. M. Campbell et al. Event Generators for High-Energy Physics Experiments. In 2022 Snowmass Summer Study, 3 2022.
  11. On contrastive divergence learning. In Robert G. Cowell and Zoubin Ghahramani, editors, Proceedings of the Tenth International Workshop on Artificial Intelligence and Statistics, volume R5 of Proceedings of Machine Learning Research, pages 33–40. PMLR, 06–08 Jan 2005. Reissued by PMLR on 30 March 2021.
  12. Variational Autoencoders for New Physics Mining at the Large Hadron Collider. JHEP, 05:036, 2019.
  13. Stochastic gradient hamiltonian monte carlo. In International Conference on Machine Learning, 2014.
  14. Taoli Cheng. Test sets for jet anomaly detection at the lhc, March 2021.
  15. Variational autoencoders for anomalous jet tagging. Phys. Rev. D, 107(1):016002, 2023.
  16. Invariant representation driven neural classifier for anti-QCD jet tagging. JHEP, 10:152, 2022.
  17. Generative modeling of convolutional neural networks. CoRR, abs/1412.6296, 2014.
  18. Delphes 3: a modular framework for fast simulation of a generic collider experiment. Journal of High Energy Physics, 2014(2), Feb 2014.
  19. Residual energy-based models for text generation. In International Conference on Learning Representations, 2020.
  20. A normalized autoencoder for lhc triggers, 2022.
  21. Compositional visual generation and inference with energy based models.
  22. Unsupervised learning of compositional energy concepts. In NeurIPS, 2021.
  23. Improved contrastive divergence training of energy based models.
  24. Energy-based models for atomic-resolution protein conformations. In International Conference on Learning Representations, 2020.
  25. Implicit generation and generalization in energy-based models.
  26. F. Englert and R. Brout. Broken symmetry and the mass of gauge vector mesons. Phys. Rev. Lett., 13:321–323, Aug 1964.
  27. Precise simulation of electromagnetic calorimeter showers using a Wasserstein Generative Adversarial Network. Comput. Softw. Big Sci., 3(1):4, 2019.
  28. S. Chatrchyan et al. Observation of a new boson at a mass of 125 gev with the cms experiment at the lhc. Physics Letters B, 716(1):30–61, 2012.
  29. Autoencoders for unsupervised anomaly detection in high energy physics. JHEP, 06:161, 2021.
  30. J.A. Fodor and E. LePore. The Compositionality Papers. Clarendon Press, 2002.
  31. Generative adversarial networks. Communications of the ACM, 63(11):139–144, 2020.
  32. No mcmc for me: Amortized sampling for fast and stable training of energy-based models. ArXiv, abs/2010.04230, 2021.
  33. Your classifier is secretly an energy based model and you should treat it like one.
  34. Scaling out-of-distribution detection for real-world settings. In International Conference on Machine Learning, 2022.
  35. Deep anomaly detection with outlier exposure. In International Conference on Learning Representations, 2019.
  36. Peter W. Higgs. Broken symmetries, massless particles and gauge fields. Phys. Lett., 12:132–133, 1964.
  37. Geoffrey E. Hinton. Training products of experts by minimizing contrastive divergence. Neural Comput., 14(8):1771–1800, aug 2002.
  38. J J Hopfield. Neural networks and physical systems with emergent collective computational abilities. Proceedings of the National Academy of Sciences, 79(8):2554–2558, 1982.
  39. Improving Variational Autoencoders for New Physics Detection at the LHC With Normalizing Flows. Front. Big Data, 5:803685, 2022.
  40. Introspective classification with convolutional nets. In I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017.
  41. Anomalous jet identification via sequence modeling. JINST, 16(08):P08012, 2021.
  42. Graph Generative Adversarial Networks for Sparse Data Generation in High Energy Physics. In 34th Conference on Neural Information Processing Systems, 11 2020.
  43. Particle Cloud Generation with Message Passing Generative Adversarial Networks. 6 2021.
  44. Adam: A method for stochastic optimization. In ICLR (Poster), 2015.
  45. Auto-encoding variational bayes. CoRR, abs/1312.6114, 2014.
  46. A tutorial on energy-based learning. 2006.
  47. Qcd jet samples with particle flow constituents, July 2020.
  48. Graphebm: Molecular graph generation with energy-based models. ArXiv, abs/2102.00546, 2021.
  49. Energy-based out-of-distribution detection. In Proceedings of the 34th International Conference on Neural Information Processing Systems, NIPS’20, Red Hook, NY, USA, 2020. Curran Associates Inc.
  50. Do deep generative models know what they don’t know? ArXiv, abs/1810.09136, 2019.
  51. Energy-based reranking: Improving neural machine translation using energy-based models. In ACL, 2021.
  52. Radford Neal. MCMC Using Hamiltonian Dynamics. In Handbook of Markov Chain Monte Carlo, pages 113–162. 2011.
  53. Radford M. Neal. Annealed importance sampling. Statistics and Computing, 11:125–139, 2001.
  54. Learning deep energy models. In Proceedings of the 28th International Conference on International Conference on Machine Learning, ICML’11, page 1105–1112, Madison, WI, USA, 2011. Omnipress.
  55. On the anatomy of mcmc-based maximum likelihood learning of energy-based models. In AAAI, 2020.
  56. Learning non-convergent non-persistent short-run mcmc toward energy-based model. In NeurIPS, 2019.
  57. Boltzmann generators: Sampling equilibrium states of many-body systems with deep learning. Science, 365, 2019.
  58. Accelerating Science with Generative Adversarial Networks: An Application to 3D Particle Showers in Multilayer Calorimeters. Phys. Rev. Lett., 120(4):042003, 2018.
  59. ParticleNet: Jet Tagging via Particle Clouds. Phys. Rev. D, 101(5):056019, 2020.
  60. Likelihood ratios for out-of-distribution detection. ArXiv, abs/1906.02845, 2019.
  61. A brief introduction to pythia 8.1. Computer Physics Communications, 178(11):852–867, Jun 2008.
  62. Improved techniques for training score-based generative models. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 12438–12448. Curran Associates, Inc., 2020.
  63. Tijmen Tieleman. Training restricted boltzmann machines using approximations to the likelihood gradient. In Proceedings of the 25th International Conference on Machine Learning, ICML ’08, page 1064–1071, New York, NY, USA, 2008. Association for Computing Machinery.
  64. Particle-based fast jet simulation at the lhc with variational autoencoders, 2022.
  65. Attention is all you need. ArXiv, abs/1706.03762, 2017.
  66. Bayesian learning via stochastic gradient langevin dynamics. In Proceedings of the 28th International Conference on International Conference on Machine Learning, ICML’11, page 681–688, Madison, WI, USA, 2011. Omnipress.
  67. A theory of generative convnet. In International Conference on Machine Learning, pages 2635–2644. PMLR, 2016.
  68. Generative pointnet: Deep energy-based learning on unordered point sets for 3d generation, reconstruction and classification. 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 14971–14980, 2021.
  69. Autoencoding under normalization constraints. In International Conference on Machine Learning, 2021.
  70. Deep sets. In NIPS, 2017.
  71. Deep structured energy based models for anomaly detection. In International Conference on Machine Learning, 2016.

Summary

We haven't generated a summary for this paper yet.

Github Logo Streamline Icon: https://streamlinehq.com