Feature Selection

Supporting Technique

Feature selection is the process of selecting a subset of relevant features for use in model construction.

Feature selection is commonly used in machine learning to improve model performance, reduce overfitting, and decrease training time by focusing on the most relevant features with the highest predictive power.

Feature selection works by evaluating the importance of each feature and removing those that contribute little to the model’s performance. This helps to avoid the curse of dimensionality, which refers to the exponential increase in computational complexity and risk of overfitting due to more and more complex models as the number of features grows.

For example, consider a dataset with numerous predictor variables for predicting house prices. Feature selection can be used to identify the most significant predictors, such as location and size, while excluding less relevant features like the color of the house or the name of the owner.

The objectives of feature selection include removing collinear features and features with low variance. Collinear features are highly correlated and can introduce redundancy, leading to overfitting and instability in the model. Removing or combining collinear features helps to simplify the model and improve its generalization. Features with low variance contribute little to the model’s predictive power and can be removed to reduce noise. One exception to the rule is outlier detection, where it might be specifically insightful if a low-variance feature shows deviations from the most common behavior.

Common methods for feature selection include recursive elimination and recursive greedy addition. Recursive elimination involves removing one feature at a time and evaluating the model’s performance; if performance improves without the feature, it is removed permanently. Recursive greedy addition, such as Forward-SFS, iteratively adds the best new feature to the set of selected features, starting with zero features and maximizing a cross-validated score.

Feature selection is an essential step in the machine learning pipeline to ensure that models are efficient, interpretable, and robust. It is distinct from feature discovery and feature engineering, which involve crafting new features from raw data to improve model performance.

Alias
Related terms
Dimensionality Reduction Curse of Dimensionality Feature Engineering Feature Discovery