Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Self-Optimizing Feature Transformation (2209.08044v2)

Published 16 Sep 2022 in cs.LG and cs.AI

Abstract: Feature transformation aims to extract a good representation (feature) space by mathematically transforming existing features. It is crucial to address the curse of dimensionality, enhance model generalization, overcome data sparsity, and expand the availability of classic models. Current research focuses on domain knowledge-based feature engineering or learning latent representations; nevertheless, these methods are not entirely automated and cannot produce a traceable and optimal representation space. When rebuilding a feature space for a machine learning task, can these limitations be addressed concurrently? In this extension study, we present a self-optimizing framework for feature transformation. To achieve a better performance, we improved the preliminary work by (1) obtaining an advanced state representation for enabling reinforced agents to comprehend the current feature set better; and (2) resolving Q-value overestimation in reinforced agents for learning unbiased and effective policies. Finally, to make experiments more convincing than the preliminary work, we conclude by adding the outlier detection task with five datasets, evaluating various state representation approaches, and comparing different training strategies. Extensive experiments and case studies show that our work is more effective and superior.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Meng Xiao (114 papers)
  2. Dongjie Wang (53 papers)
  3. Min Wu (201 papers)
  4. Kunpeng Liu (54 papers)
  5. Hui Xiong (244 papers)
  6. Yuanchun Zhou (62 papers)
  7. Yanjie Fu (93 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.