Estimating with support vector regression

The third regression algorithm that we want to explore is support vector regression (SVR). As the name implies, SVR is part of the support vector family and sibling of the support vector machine (SVM) for classification (or we can just call it SVC) we learned in Chapter 5, Classifying Newsgroup Topic with Support Vector Machine.

To recap, SVC seeks an optimal hyperplane that best segregates observations from different classes. Suppose a hyperplane is determined by a slope vector w and intercept b, the optimal hyperplane is picked so that the distance (which can be expressed as ) from its nearest points in each of segregated spaces to the hyperplane itself is maximized. Such optimal w and b can be learned and solved by the following optimization problem:

  • Minimizing ‖w
  • Subject to wx(i)+b≥1 if y(i)=1 and wx(i)+b≤1 if y(i)=-1, given a training set of (x(1),y(1)), (x(2),y(2)),… (x(i),y(i))…, (x(m),y(m)).

In SVR, our goal is to find a hyperplane (defined by a slope vector w and intercept b) so that two hyperplanes, wx+b=-ε and wx+b=ε, that are a distance away from itself covers most training data. In other words, most of data points are bounded in the ε bands of the optimal hyperplane. And at the same time, the optimal hyperplane is as flat as possible, which means ‖w‖ is as small as possible, as shown in the following diagram:

This translates into deriving the optimal w and b by solving the following optimization problem:

  • Minimizing ‖w
  • Subject to |y(i)-(wx(i)+b)| ≤ ε , given a training set of (x(1),y(1)), (x(2),y(2)),… (x(i),y(i))…, (x(m),y(m))
..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset