Unsupervised learning

It's time to see some examples of unsupervised learning, given that we spend a majority of this book on supervised learning models.

When to use unsupervised learning

There are many times when unsupervised learning can be appropriate. Some very common examples include the following:

  • When there is no clear response variable. There is nothing that we are explicitly trying to predict or correlate to other variables.
  • To extract structure from data where no apparent structure/patterns exist (can be a supervised learning problem).
  • When an unsupervised concept called feature extraction is used. Feature extraction is the process of creating new features from existing ones. These new features can be even stronger than the original features.

The first tends to be the most common reason that data scientists choose to use unsupervised learning. This case arises frequently when we are working with data and we are not explicitly trying to predict any of the columns and we merely wish to find patterns of similar (and dissimilar) groups of points. The second option comes into play even if we are explicitly attempting to use a supervised model to predict a response variable. Sometimes simple EDA might not produce any clear patterns in the data in the few dimensions that humans can imagine where as a machine might pick up on data points behaving similarly to each other in greater dimensions.

The third common reason to use unsupervised learning is to extract new features from features that already exist. This process (lovingly called feature extraction) might produce features that can be used in a future supervised model or that can be used for presentation purposes (marketing or otherwise).

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset