Information Modified K-Nearest Neighbor (2312.01991v2)
Abstract: The fundamental concept underlying K-Nearest Neighbors (KNN) is the classification of samples based on the majority through their nearest neighbors. Although distance and neighbors' labels are critical in KNN, traditional KNN treats all samples equally. However, some KNN variants weigh neighbors differently based on a specific rule, considering each neighbor's distance and label. Many KNN methodologies introduce complex algorithms that do not significantly outperform the traditional KNN, often leading to less satisfactory outcomes. The gap in reliably extracting information for accurately predicting true weights remains an open research challenge. In our proposed method, information-modified KNN (IMKNN), we bridge the gap by presenting a straightforward algorithm that achieves effective results. To this end, we introduce a classification method to improve the performance of the KNN algorithm. By exploiting mutual information (MI) and incorporating ideas from Shapley's values, we improve the traditional KNN performance in accuracy, precision, and recall, offering a more refined and effective solution. To evaluate the effectiveness of our method, it is compared with eight variants of KNN. We conduct experiments on 12 widely-used datasets, achieving 11.05\%, 12.42\%, and 12.07\% in accuracy, precision, and recall performance, respectively, compared to traditional KNN. Additionally, we compared IMKNN with traditional KNN across four large-scale datasets to highlight the distinct advantages of IMKNN in the impact of monotonicity, noise, density, subclusters, and skewed distributions. Our research indicates that IMKNN consistently surpasses other methods in diverse datasets.
- C. E. Shannon, “A mathematical theory of communication,” The Bell System Technical Journal, vol. 27, no. 3, pp. 379–423, 1948.
- T. Cover and P. Hart, “Nearest neighbor pattern classification,” IEEE Trans. Inform. Theory, vol. IT-13, no. 1, p. 21–27, 1967.
- G. Fei and B. Liu, “Breaking the closed-world assumption in text classification,” in Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. San Diego, California: Association for Computational Linguistics, 2016, pp. 506–514.
- H. Parvin, H. Alizadeh, and B. Minaei-Bidgoli, “Mknn: Modified k-nearest neighbor,” 2008. [Online]. Available: https://api.semanticscholar.org/CorpusID:3089035
- M. Hellman, “The nearest neighbor classification rule with a reject option,” IEEE Trans. Syst. Man Cybern., vol. 3, p. 179–185, 1970.
- S. Uddin, A. Khan, M. E. Hossain, and M. A. Moni, “Comparing different supervised machine learning algorithms for disease prediction,” BMC Medical Informatics and Decision Making, vol. 19, pp. 1–16, 2019.
- S. Zhang, X. Li, M. Zong, X. Zhu, and D. Cheng, “Learning k for knn classification,” ACM Transactions on Intelligent Systems and Technology (TIST), vol. 8, pp. 1–19, 2017.
- H. Parvin, H. Alizadeh, and B. Minaei-Bidgoli, “Mknn: Modified k-nearest neighbor,” 2005.
- K. He, H. Fan, Y. Wu, S. Xie, and R. Girshick, “Momentum contrast for unsupervised visual representation learning,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 9729–9738.
- L. S. Shapley, “Notes on the n-person game – ii: The value of an n-person game,” RAND Corporation, Santa Monica, Calif., Tech. Rep., 1951.
- B. Mahesh, “Machine learning algorithms—a review,” International Journal of Scientific Research, vol. 9, pp. 381–386, 2020.
- M. Gutmann and A. Hyvärinen, “Noise-contrastive estimation: A new estimation principle for unnormalized statistical models,” in Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics. JMLR Workshop and Conference Proceedings, 2010, pp. 297–304.
- S. Aeberhard, D. Coomans, and O. de Vel, “Comparison of classifiers in high dimensional settings,” Dept. of Computer Science and Dept. of Mathematics and Statistics, James Cook University of North Queensland, Tech. Rep. 92-02.
- D. Hendrycks and K. Gimpel, “A baseline for detecting misclassified and out-of-distribution examples in neural networks,” in 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, 2017.
- P. Khosla, P. Teterwak, C. Wang, A. Sarna, Y. Tian, P. Isola, A. Maschinot, C. Liu, and D. Krishnan, “Supervised contrastive learning,” in Advances in Neural Information Processing Systems, vol. 33. Curran Associates, Inc., 2020, pp. 18 661–18 673.
- R. Hadsell, S. Chopra, and Y. LeCun, “Dimensionality reduction by learning an invariant mapping,” in 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), vol. 2. IEEE, 2006, pp. 1735–1742.
- D. Bzdok, M. Krzywinski, and N. Altman, “Machine learning: supervised methods,” Nature Methods, vol. 15, pp. 5–6, 2018.
- N. Bhatia and Vandana, “Survey of nearest neighbor techniques,” International Journal of Computer Science & Information Security, vol. 8, pp. 1–4, 2010.
- A. Lamba and D. Kumar, “Survey on knn and its variants,” International Journal of Advanced Research in Computer and Communication Engineering, vol. 5, pp. 430–435, 2016.
- Z. Pan, Y. Wang, and Y. Pan, “A new locally adaptive k-nearest neighbor algorithm based on discrimination class,” Knowledge-Based Systems, vol. 204, p. 106185, 2020.
- E.-H. S. Han, G. Karypis, and V. Kumar, “In pacific-asia conference on knowledge discovery and data mining.” Springer, 2001, pp. 53–65.
- J. M. Keller, M. R. Gray, and J. A. Givens, “A fuzzy k-nearest neighbor algorithm,” IEEE Transactions on Systems, Man, and Cybernetics, vol. 15, pp. 580–585, 1985.
- J. e. a. Gou, “A generalised mean distance-based k-nearest neighbor classifier,” Expert Systems with Applications, vol. 115, pp. 356–372, 2019.
- J. Dhar, A. Shukla, M. Kumar, and P. Gupta, “A weighted mutual k-nearest neighbor for classification mining,” Journal of Advances in Physics, 2020.
- A. B. Hassanat, M. A. Abbadi, and G. A. Altarawneh, “Solving the problem of the k parameter in the knn classifier using an ensemble learning approach,” Journal of Applied Sciences, 2014.
- C. Blake and C. Merz. (1998) Uci repository of machine learning databases. [Online]. Available: http://www.ics.uci.edu/~mlearn/MLRepository.html
- J. Vanschoren, “Openml,” https://www.openml.org/, 2014.
- R. A. Fisher, “Iris,” UCI Machine Learning Repository, 1988, DOI: https://doi.org/10.24432/C56C76.
- S. Aeberhard and M. Forina, “Wine,” UCI Machine Learning Repository, 1991, DOI: https://doi.org/10.24432/C5PC7J.
- “South africa heart disease dataset,” http://statweb.stanford.edu/~tibs/ElemStatLearn/data.html, source: http://statweb.stanford.edu/~tibs/ElemStatLearn/data.html.
- “Liver Disorders,” UCI Machine Learning Repository, 1990, DOI: https://doi.org/10.24432/C54G67.
- N. Bhat, “Health care: Heart attack possibility,” https://www.kaggle.com/nareshbhat/health-care-data-set-on-heart-attack-possibility, 2020.
- “Statlog (Heart),” UCI Machine Learning Repository, DOI: https://doi.org/10.24432/C57303.
- D. Chicco and G. Jurman, “Machine learning can predict survival of patients with heart failure from serum creatinine and ejection fraction alone,” BMC Medical Informatics and Decision Making, vol. 20, no. 1, p. 16, 2020.
- Stroke prediction dataset. https://www.kaggle.com/datasets/fedesoriano/stroke-prediction-dataset. Accessed on [Date Retrieved].
- fedesoriano, “Wind speed prediction dataset,” https://www.kaggle.com/datasets/fedesoriano/wind-speed-prediction-dataset, April 2022, accessed [Date Retrieved].
- M. Lichman, “Uci machine learning repository,” http://archive.ics.uci.edu/ml, 2013, irvine, CA: University of California, School of Information and Computer Science.