Beyond Cuts in Small Signal Scenarios -- Enhanced Sneutrino Detectability Using Machine Learning (2108.03125v4)
Abstract: We investigate enhancing the sensitivity of new physics searches at the LHC by machine learning in the case of background dominance and a high degree of overlap between the observables for signal and background. We use two different models, XGBoost and a deep neural network, to exploit correlations between observables and compare this approach to the traditional cut-and-count method. We consider different methods to analyze the models' output, finding that a template fit generally performs better than a simple cut. By means of a Shapley decomposition, we gain additional insight into the relationship between event kinematics and the machine learning model output. We consider a supersymmetric scenario with a metastable sneutrino as a concrete example, but the methodology can be applied to a much wider class of models.
- M. Feickert and B. Nachman, A Living Review of Machine Learning for Particle Physics. arXiv:2102.02770 [hep-ph].
- CERN Yellow Reports: Monographs. CERN, Geneva, 2020. https://cds.cern.ch/record/2749422.
- https://cds.cern.ch/record/2688062.
- T. Chen and C. Guestrin, “XGBoost: A scalable tree boosting system,” in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’16, pp. 785–794. ACM, New York, NY, USA, 2016. http://doi.acm.org/10.1145/2939672.2939785.
- Springer Series in Statistics. Springer New York Inc., New York, NY, USA, 2001.
- L. S. Shapley, A value for n-person games. Contributions to the Theory of Games 2 (1953) 307–317.
- S. M. Lundberg and S.-I. Lee, “A unified approach to interpreting model predictions,” in Advances in Neural Information Processing Systems 30, I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, eds., pp. 4765–4774. Curran Associates, Inc., 2017. http://papers.nips.cc/paper/7062-a-unified-approach-to-interpreting-model-predictions.pdf.
- S. M. Lundberg and S.-I. Lee, “A unified approach to interpreting model predictions,” in Advances in neural information processing systems, pp. 4765–4774. 2017.
- E. Štrumbelj and I. Kononenko, Explaining prediction models and individual predictions with feature contributions. Knowledge and Information Systems 41 (2013) 647–665.
- A. Datta, S. Sen, and Y. Zick, “Algorithmic transparency via quantitative input influence: Theory and experiments with learning systems,” in 2016 IEEE symposium on security and privacy (SP), pp. 598–617, IEEE. 2016.
- arXiv:2007.06011 [stat.ML].
- F. Huettner and M. Sunder, Axiomatic arguments for decomposing goodness of fit according to Shapley and Owen values. Electronic Journal of Statistics 6 (2012) 1239–1250.