Summary

In this chapter, we had a closer look at modeling sequences of observations with hidden (or latent) states with the two commonly used algorithms:

  • The generative hidden Markov model to maximize p(X,Y)
  • The discriminative conditional random field to maximize log p(Y|X)

The HMM is a special form of Bayes network. It requires the observations to be independent. Although restrictive, the conditional independence prerequisites make the HMM fairly easy to understand and validate, which is not the case for a CRF.

You learned how to implement three dynamic programming techniques: Viterbi, Baum-Welch, and alpha/beta algorithms in Scala. These algorithms are used to solve diverse type of optimization problems. They should be an essential component of your algorithmic tool box.

The conditional random field relies on the logistic regression to estimate the optimal weights of the model. Such a technique is also used in the multiple layer perceptron, which was introduced in Chapter 9, Artificial Neural Network. The next chapter introduces two important alternatives to the logistic regression for discriminating between observations: the Kernel function for nonlinear models and the maximization of the margin between classes of observations.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset