Papers
Topics
Authors
Recent
2000 character limit reached

EasyFS: an Efficient Model-free Feature Selection Framework via Elastic Transformation of Features (2402.05954v2)

Published 4 Feb 2024 in cs.LG

Abstract: Traditional model-free feature selection methods treat each feature independently while disregarding the interrelationships among features, which leads to relatively poor performance compared with the model-aware methods. To address this challenge, we propose an efficient model-free feature selection framework via elastic expansion and compression of the features, namely EasyFS, to achieve better performance than state-of-the-art model-aware methods while sharing the characters of efficiency and flexibility with the existing model-free methods. In particular, EasyFS expands the feature space by using the random non-linear projection network to achieve the non-linear combinations of the original features, so as to model the interrelationships among the features and discover most correlated features. Meanwhile, a novel redundancy measurement based on the change of coding rate is proposed for efficient filtering of redundant features. Comprehensive experiments on 21 different datasets show that EasyFS outperforms state-of-the art methods up to 10.9\% in the regression tasks and 5.7\% in the classification tasks while saving more than 94\% of the time.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (55)
  1. k-best feature selection and ranking via stochastic approximation. Expert Systems with Applications, 213:118864, 2023.
  2. A review of microarray datasets and applied feature selection methods. Information sciences, 282:111–135, 2014.
  3. Unsupervised feature selection for multi-cluster data. In Proceedings of the 16th ACM SIGKDD international conference on Knowledge discovery and data mining, pp.  333–342, 2010.
  4. Efficient ant colony optimization for image feature selection. Signal processing, 93(6):1566–1576, 2013.
  5. Kernel feature selection via conditional covariance minimization. Advances in Neural Information Processing Systems, 30, 2017.
  6. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining, pp.  785–794, 2016.
  7. Few-sample feature selection via feature manifold learning. In International Conference on Machine Learning, pp.  6296–6319. PMLR, 2023.
  8. Gene selection and classification of microarray data using random forest. BMC bioinformatics, 7:1–13, 2006.
  9. Normalized mutual information feature selection. IEEE Transactions on neural networks, 20(2):189–201, 2009.
  10. Frank, A. Uci machine learning repository. http://archive. ics. uci. edu/ml, 2010.
  11. Iteratively local fisher score for feature selection. Applied Intelligence, 51:6167–6181, 2021.
  12. From few to many: Illumination cone models for face recognition under variable lighting and pose. IEEE transactions on pattern analysis and machine intelligence, 23(6):643–660, 2001.
  13. Characterising virtual eigensignatures for general purpose face recognition. In Face recognition: from theory to applications, pp.  446–456. Springer, 1998.
  14. Generalized fisher score for feature selection. arXiv preprint arXiv:1202.3725, 2012.
  15. Hall, M. A. Correlation-based feature selection for machine learning. PhD thesis, The University of Waikato, 1999.
  16. Autoencoder inspired unsupervised feature selection. In 2018 IEEE international conference on acoustics, speech and signal processing (ICASSP), pp.  2941–2945. IEEE, 2018.
  17. Differential evolution for filter feature selection based on information theory and feature ranking. Knowledge-Based Systems, 140:103–119, 2018.
  18. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp.  770–778, 2016.
  19. Laplacian score for feature selection. In Weiss, Y., Schölkopf, B., and Platt, J. (eds.), Advances in Neural Information Processing Systems, volume 18. MIT Press, 2005. URL https://proceedings.neurips.cc/paper_files/paper/2005/file/b5b03f06271f8917685d14cea7c6c50a-Paper.pdf.
  20. Composite feature selection using deep ensembles. Advances in Neural Information Processing Systems, 35:36142–36160, 2022.
  21. Jaeger, H. Adaptive nonlinear system identification with echo state networks. Advances in neural information processing systems, 15, 2002.
  22. Feature dimensionality reduction: a review. Complex & Intelligent Systems, 8(3):2663–2693, 2022.
  23. Face recognition using convolutional neural network and simple logistic classifier. In Soft Computing in Industrial Applications: Proceedings of the 17th Online World Conference on Soft Computing in Industrial Applications, pp.  197–207. Springer, 2014.
  24. Lassonet: Neural networks with feature sparsity. In International conference on artificial intelligence and statistics, pp.  10–18. PMLR, 2021.
  25. Feature selection: A data perspective. ACM computing surveys (CSUR), 50(6):1–45, 2017.
  26. Deep feature selection: theory and application to identify enhancers and promoters. Journal of Computational Biology, 23(5):322–336, 2016.
  27. Conditional infomax learning: An integrated framework for feature extraction and fusion. Lecture Notes in Computer Science,Lecture Notes in Computer Science, Jan 2006.
  28. Phenotype prediction and genome-wide association study using deep convolutional neural network of soybean. Frontiers in genetics, 10:1091, 2019.
  29. The international rice information system. a platform for meta-analysis of rice crop data. Plant Physiology, 139(2):637–642, 2005.
  30. On the Use of Variable Complementarity for Feature Selection in Cancer Classification, pp.  91–102. Jan 2006. doi: 10.1007/11732242˙9. URL http://dx.doi.org/10.1007/11732242_9.
  31. Deep feature selection using a teacher-student network. Neurocomputing, 383:396–408, 2020.
  32. Feature and parameter selection in stochastic linear bandits. In International Conference on Machine Learning, pp.  15927–15958. PMLR, 2022.
  33. Fast and accurate uncertainty estimation in chemical machine learning. Journal of chemical theory and computation, 15(2):906–915, 2019.
  34. Lasso: A feature selection technique in predictive modeling for machine learning. In 2016 IEEE international conference on advances in computer applications (ICACA), pp.  18–20. IEEE, 2016.
  35. Columbia object image library (coil-20). 1996.
  36. Reading digits in natural images with unsupervised feature learning. 2011.
  37. Scikit-learn: Machine learning in python. the Journal of machine Learning research, 12:2825–2830, 2011.
  38. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Transactions on pattern analysis and machine intelligence, 27(8):1226–1238, 2005.
  39. Theoretical and empirical analysis of relieff and rrelieff. Machine learning, 53:23–69, 2003a.
  40. Theoretical and empirical analysis of relieff and rrelieff. Machine learning, 53:23–69, 2003b.
  41. Infinite feature selection: a graph-based feature filtering approach. IEEE Transactions on Pattern Analysis and Machine Intelligence, 43(12):4396–4410, 2020.
  42. Ross, B. C. Mutual information between discrete and continuous data sets. PloS one, 9(2):e87357, 2014.
  43. Application of support vector regression in indonesian stock price prediction with feature selection using particle swarm optimisation. Modelling and Simulation in Engineering, 2019, 2019.
  44. Parameterisation of a stochastic model for human face identification. In Proceedings of 1994 IEEE workshop on applications of computer vision, pp.  138–142. IEEE, 1994.
  45. The cmu pose, illumination, and expression (pie) database. In Proceedings of fifth IEEE international conference on automatic face gesture recognition, pp.  53–58. IEEE, 2002.
  46. Tibshirani, R. Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society Series B: Statistical Methodology, 58(1):267–288, 1996.
  47. Dnngp, a deep neural network-based method for genomic prediction using multi-omics data in plants. Molecular Plant, 16(1):279–293, 2023.
  48. Feature importance ranking for deep learning. Advances in Neural Information Processing Systems, 33:5105–5114, 2020.
  49. Efficient top-k feature selection using coordinate descent method. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pp.  10594–10601, 2023.
  50. A novel ensemble-based wrapper method for feature selection using extreme learning machine and genetic algorithm. Knowledge and Information Systems, 57:389–412, 2018.
  51. Feature selection using stochastic gates. In International Conference on Machine Learning, pp.  10648–10659. PMLR, 2020.
  52. A review on basic data-driven approaches for industrial process monitoring. IEEE Transactions on Industrial electronics, 61(11):6418–6428, 2014.
  53. Learning diverse and discriminative representations via the principle of maximal coding rate reduction. Advances in Neural Information Processing Systems, 33:9422–9434, 2020.
  54. A comprehensive review of dimensionality reduction techniques for feature selection and feature extraction. Journal of Applied Science and Technology Trends, 1(2):56–70, 2020.
  55. Spectral feature selection for supervised and unsupervised learning. In Proceedings of the 24th international conference on Machine learning, pp.  1151–1157, 2007.

Summary

We haven't generated a summary for this paper yet.

Whiteboard

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.