Diving into data

A blog on machine learning, data mining and visualization

Main menu

Skip to content
  • About

Selecting good features – Part IV: stability selection, RFE and everyting side by side

December 20, 2014

In my previous posts, I looked at univariate methods,linear models and regularization and random forests for feature selection. In this post, I’ll look at two other methods: stability selection and recursive feature elimination (RFE), which can both considered wrapper methods. … Continue reading →

Posted in Feature selection | Reply

Selecting good features – Part III: random forests

December 1, 2014

In my previous posts, I looked at univariate feature selection and linear models and regularization for feature selection. In this post, I’ll discuss random forests, another popular approach for feature ranking. Random forest feature importance Random forests are among the … Continue reading →

Posted in Feature selection | Reply

Selecting good features – Part II: linear models and regularization

November 12, 2014

In my previous post I discussed univariate feature selection where each feature is evaluated independently with respect to the response variable. Another popular approach is to utilize machine learning models for feature ranking. Many machine learning models have either some … Continue reading →

Posted in Feature selection, Machine learning | Reply

Selecting good features – Part I: univariate selection

November 2, 2014

Having a good understanding of feature selection/ranking can be a great asset for a data scientist or machine learning practitioner. A good grasp of these methods leads to better performing models, better understanding of the underlying structure and characteristics of … Continue reading →

Posted in Feature selection, Machine learning | Reply

Interpreting random forests

October 19, 2014

Why model interpretation? Imagine a situation where a credit card company has built a fraud detection model using a random forest. The model can classify every transaction as either valid or fraudulent, based on a large number of features. What … Continue reading →

Posted in Machine learning, Random forest | Replies: 3

Recent Posts

  • Selecting good features – Part IV: stability selection, RFE and everyting side by side
  • Selecting good features – Part III: random forests
  • Selecting good features – Part II: linear models and regularization
  • Selecting good features – Part I: univariate selection
  • Interpreting random forests

Archives

  • December 2014
  • November 2014
  • October 2014

Meta

  • Log in
  • Entries RSS
  • Comments RSS
  • WordPress.org