A Mallows-like Criterion for Anomaly Detection with Random Forest Implementation (2405.18932v1)
Abstract: The effectiveness of anomaly signal detection can be significantly undermined by the inherent uncertainty of relying on one specified model. Under the framework of model average methods, this paper proposes a novel criterion to select the weights on aggregation of multiple models, wherein the focal loss function accounts for the classification of extremely imbalanced data. This strategy is further integrated into Random Forest algorithm by replacing the conventional voting method. We have evaluated the proposed method on benchmark datasets across various domains, including network intrusion. The findings indicate that our proposed method not only surpasses the model averaging with typical loss functions but also outstrips common anomaly detection algorithms in terms of accuracy and robustness.
- V. Hodge and J. Austin. A survey of outlier detection methodologies. Artificial Intelligence Review, 22(1):85–126, January 2004.
- Anomaly detection: A survey. ACM Computing Surveys (CSUR), 41(3):1–58, August 2009.
- C. C. Aggarwal. An introduction to outlier analysis. Springer International Publishing, 2017.
- A survey of anomaly detection techniques in financial domain. Future Generation Computer Systems, 55:278–288, February 2016.
- A. Patcha and J. M. Park. An overview of anomaly detection techniques: Existing solutions and latest technological trends. Computer Networks, 51(12):3448–3470, February 2007.
- Algorithms for mining distance-based outliers in large datasets. In Proceedings of the International Conference on Very Large Data Bases, pages 392–403. Citeseer, 1998.
- Outlier detection: Techniques and applications. pages 3–11. Springer Nature, 2019.
- C. L. Mallows. Some comments on Cpsubscript𝐶𝑝C_{p}italic_C start_POSTSUBSCRIPT italic_p end_POSTSUBSCRIPT. Technometrics, 42(1):87–94, 2000.
- H. Leeb and B. M. Pötscher. Model selection and inference: Facts and fiction. Econometric Theory, 21(1):21–59, February 2005.
- Bayesian model averaging for linear regression models. Journal of the American Statistical Association, 92(437):179–191, November 1997.
- Regularization and variable selection via the elastic net. Journal of the Royal Statistical Society Series B: Statistical Methodology, 67(2):301–320, April 2005.
- Mallows-type averaging machine learning techniques. to be published, 2023.
- Bruce E. Hansen. Least squares model averaging. Econometrica, 75(4):1175–1189, June 2007.
- C. J. Willmott and K. Matsuura. Advantages of the mean absolute error (mae) over the root mean square error (rmse) in assessing average model performance. Climate Research, 30(1):79–82, December 2005.
- Bayesian model averaging: a tutorial (with comments by m. clyde, david draper and ei george, and a rejoinder by the authors). Statistical Science, 14(4):382–417, November 1999.
- A model-averaging method for high-dimensional regression with missing responses at random. Statistica Sinica, 31(2):1005–1026, January 2021.
- T. Ando and K. C. Li. A model-averaging approach for high-dimensional regression. Journal of the American Statistical Association, 109(505):254–265, March 2014.
- Bayes model averaging with selection of regressors. Journal of the Royal Statistical Society Series B: Statistical Methodology, 64(3):519–536, August 2002.
- Efficient decentralized deep learning by dynamic model averaging. In Machine Learning and Knowledge Discovery in Databases: European Conference, pages 393–409, Dublin, Ireland, 2018. Springer.
- S. Rüping. Incremental learning with support vector machines. In Proceedings 2001 IEEE International Conference on Data Mining, pages 641–642, San Jose, CA, USA, 2001.
- A. Krogh and J. Vedelsby. Neural network ensembles, cross validation, and active learning. In Advances in Neural Information Processing Systems, volume 7, pages 231–238, 1994.
- H. Deng and G. Runger. Gene selection with guided regularized random forest. Pattern Recognition, 46(12):3483–3489, dec 2013.
- Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, pages 2980–2988, 2017.
- Focal loss in 3d object detection. IEEE Robotics and Automation Letters, 4(2):1263–1270, apr 2019.
- K. Doi and A. Iwasaki. The effect of focal loss in semantic segmentation of high resolution aerial image. In IEEE International Geoscience and Remote Sensing Symposium, pages 6919–6922, 2018.
- Practical bayesian optimization of machine learning algorithms. In Advances in Neural Information Processing Systems, volume 25, 2012.
- I. Markovsky and S. Van Huffel. Overview of total least-squares methods. Signal Processing, 87(10):2283–2302, oct 2007.
- S. Dasgupta and D. Hsu. Hierarchical sampling for active learning. In Proceedings of the 25th International Conference on Machine Learning, pages 208–215. ACM, 2008.
- T. Janarthanan and S. Zargari. Feature selection in unsw-nb15 and kddcup’99 datasets. In 2017 IEEE 26th International Symposium on Industrial Electronics (ISIE), pages 1881–1886, 2017.
- Data Mining: Practical Machine Learning Tools and Techniques. ACM Press, New York, NY, USA, 2005.