Feature Importance and Feature Selection - Framework

Feature Importance and Feature Selection - Framework

No alt text provided for this image

?? Feature Importance and Feature Selection

????Feature selection is itself useful, but it mostly acts as a filter, muting out features that aren’t useful in addition to your existing features.

OR

????Feature selection is the process of selecting a subset of relevant features for use in model construction.

------------------------------

?? Correlation-based Feature Selection (CFS): CFS selects features based on their correlation with the target variable and their correlation with each other. It measures the subset's ability to predict the target variable while also avoiding the inclusion of redundant features.

?? Recursive Feature Elimination (RFE): RFE is a popular feature selection method that recursively eliminates features based on their importance to the model's performance. It uses a model to evaluate feature importance and removes the least important feature iteratively until the desired number of features is reached.

?? Tree-based Feature Importance: Tree-based feature importance is a widely used method for measuring feature importance in decision tree-based algorithms. It calculates the importance of each feature by how much it reduces the impurity or entropy of the model.

?? Mutual Information-based Feature Selection: Mutual information-based feature selection evaluates the mutual information between each feature and the target variable. It selects features with high mutual information, which implies that they contain information that is relevant to predicting the target variable.

?? Lasso Regularization: Lasso regularization is a linear model regularization technique that shrinks the coefficients of less important features to zero. It selects features by setting their coefficients to zero, resulting in a sparse model with only the most important features.

?? Principal Component Analysis (PCA): PCA is a dimensionality reduction technique that transforms high-dimensional data into a lower-dimensional space while preserving most of the variance in the original data. It selects features by identifying the principal components that explain most of the variability in the data.

?? Mutual information: This technique involves measuring the mutual information between each feature and the target variable, and selecting the features with the highest mutual information.


Will share more later, follow for more details.


You can also follow this framework:

?? Feature Importance Analysis: Analyse the importance of each feature in your dataset using one or more of the following methods:

??Correlation Analysis

??Univariate Feature Selection

?? Model-Based Feature Selection

?? Recursive Feature Elimination

?? Feature Selection: Once you have identified the most important features, you can use one or more of the following techniques to select the final set of features:

??Filter Methods

??Wrapper Methods

??Embedded Methods

要查看或添加评论,请登录

社区洞察

其他会员也浏览了