Chapter 13 Support Vector Machine

svm1

svm1

Goal: we want to find the hyperplane (i.e. decision boundary) linearly separating (or not) our classes.

Support Vector Machines (SVMs) are a particular classification strategy. SMVs work by transforming the training dataset into a higher dimension, which is then inspected for the optimal separation boundary, or boundaries, between classes. In SVMs, these boundaries are referred to as hyperplanes, which are identified by locating support vectors, or the instances that most essentially define classes, and their margins, which are the lines parallel to the hyperplane defined by the shortest distance between a hyperplane and its support vectors. Consequently, SVMs are able to classify both linear and nonlinear data.

The grand idea with SVMs is that, with a high enough number of dimensions, a hyperplane separating a particular class from all others can always be found, thereby delineating dataset member classes. When repeated a sufficient number of times, enough hyperplanes can be generated to separate all classes in n-dimensional space. Importantly, SVMs look not just for any separating hyperplane but the maximum-margin hyperplane, being that which resides equidistance from respective class support vectors.

The distance between the hyperplane and the nearest data point from either set is known as the margin. The goal is to choose a hyperplane with the greatest possible margin between the hyperplane and any point within the training set, giving a greater chance of new data being classified correctly. svm2 Intuitively, the further from the hyperplane our data points lie, the more confident we are that they have been correctly classified. We therefore want our data points to be as far away from the hyperplane as possible, while still being on the correct side of it.

svm3

svm3

A dataset will often look more like the jumbled balls below which represent a linearly non separable dataset.

In order to classify a dataset like the one above it’s necessary to move away from a 2d view of the data to a 3d view. Explaining this is easiest with another simplified example. Imagine that our two sets of colored balls above are sitting on a sheet and this sheet is lifted suddenly, launching the balls into the air. While the balls are up in the air, you use the sheet to separate them. This ‘lifting’ of the balls represents the mapping of data into a higher dimension. This is known as kernelling. The commonly used kernel functions are: a) Linear, b) Polynomial, c) Sigmoid and d) Radial Basis.

svm4

svm4

Because we are now in three dimensions, our hyperplane can no longer be a line. It must now be a plane as shown in the example above. The idea is that the data will continue to be mapped into higher and higher dimensions until a hyperplane can be formed to segregate it.

13.1 Support Vecotr Regression

A major benefit of using SVR is that it is a non-parametric technique. Unlike SLR, whose results depend on Gauss-Markov assumptions, the output model from SVR does not depend on distributions of the underlying dependent and independent variables. Instead the SVR technique depends on kernel functions. Another advantage of SVR is that it permits for construction of a non-linear model without changing the explanatory variables, helping in better interpretation of the resultant model. The basic idea behind SVR is not to care about the prediction as long as the error (ϵi) is less than certain value. This is known as the principle of maximal margin. This idea of maximal margin allows viewing SVR as a convex optimization problem. The regression can also be penalized using a cost parameter, which becomes handy to avoid over-fit. SVR is a useful technique provides the user with high flexibility in terms of distribution of underlying variables, relationship between independent and dependent variables and the control on the penalty term.

library(tidyverse)

13.1.1 Create data

For once, let’s generate some artificial data.

x <- 1:75
set.seed(2302)
y <- cumsum(rnorm(length(x)))

df <- tibble(x = x, y = y)
glimpse(df)
## Observations: 75
## Variables: 2
## $ x <int> 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18,…
## $ y <dbl> -1.0347913, -0.6663628, -1.6089930, -3.2992371, -2.8075105, -3…

Let’s plot our data to see what we get

ggplot(df, aes(x = x, y = y)) + 
  geom_point(col = "blue") + 
  geom_line(linetype = "dashed") 

If we were going to do a regression, we’ll get

model_lm <- lm(y ~ x, data = df)
prediction_lm <- tibble(y = df$y, y_hat = predict(model_lm, df))
head(prediction_lm, 10)
## # A tibble: 10 x 2
##         y y_hat
##     <dbl> <dbl>
##  1 -1.03  -4.14
##  2 -0.666 -4.14
##  3 -1.61  -4.15
##  4 -3.30  -4.16
##  5 -2.81  -4.16
##  6 -3.17  -4.17
##  7 -3.46  -4.18
##  8 -3.03  -4.19
##  9 -4.36  -4.19
## 10 -3.41  -4.20
ggplot(df, aes(x = x, y = y)) + 
  geom_point(col = "blue") + 
  geom_line(linetype = "dashed") + 
  geom_smooth(method = "lm")

and we’ll get a RMSE for our linear model of

prediction_lm %>% yardstick::rmse(y, y_hat)
## # A tibble: 1 x 3
##   .metric .estimator .estimate
##   <chr>   <chr>          <dbl>
## 1 rmse    standard        2.48

Let’s move onto SVM modelling.

library(e1071)
## Warning: package 'e1071' was built under R version 3.5.2
## 
## Attaching package: 'e1071'
## The following object is masked from 'package:simputation':
## 
##     impute
model_svm <- svm(y ~ x, data = df)

#To plot the prediction, we need a bit of tweaking
prediction_svm <- tibble(y = df$y, y_hat = predict(model_svm, df))

df2 <- bind_cols(df, y_hat = prediction_svm$y_hat)

ggplot(df2, aes(x = x, y = y)) + 
  geom_point(col = "blue") + 
  geom_point(aes(x = x, y = y_hat), col = "red") + 
  geom_line(linetype = "dashed")

and we’ll get a RMSE for our SVM model of

df2 %>% yardstick::rmse(y, y_hat)
## # A tibble: 1 x 3
##   .metric .estimator .estimate
##   <chr>   <chr>          <dbl>
## 1 rmse    standard        1.11

13.1.2 Tuning a SVM model

Identifying best parameters doing a grid search. We’ll vary epsilon between 0 and 1 by increment of 0.1 (10 values total). We’ll vary the cost from 0.5 to 8 by increment of 0.5 (15 values total). So we’ll test 150 models in total.

tune_model_svm <- tune(svm, y ~ x, data = df, 
                       ranges = list(epsilon = seq(0, 1, 0.1), cost = 2^(seq(0.5, 8, 0.5))))

plot(tune_model_svm)

print(tune_model_svm)
## 
## Parameter tuning of 'svm':
## 
## - sampling method: 10-fold cross validation 
## 
## - best parameters:
##  epsilon cost
##      0.2  256
## 
## - best performance: 0.6844569

Let’s even go further in our fine tuning (drilling down.. but overfitting?)

tune_model_svm2 <- tune(svm, y~x, data = df, 
                        ranges = list(epsilon = seq(tune_model_svm$best.model$epsilon-.15, 
                                                    tune_model_svm$best.model$epsilon+.15, 0.01), 
                                      cost = seq(2^(log2(tune_model_svm$best.model$cost)-1), 
                                                 2^(log2(tune_model_svm$best.model$cost)+1), length = 6)))

print(tune_model_svm2)
## 
## Parameter tuning of 'svm':
## 
## - sampling method: 10-fold cross validation 
## 
## - best parameters:
##  epsilon cost
##     0.05  512
## 
## - best performance: 0.6872402

We now use this parameters for prefiction. !!! Overfitting is possible here.

model_svm_tuned <- predict(tune_model_svm2$best.model, data = df)

#To plot the prediction, we need a bit of tweaking
prediction_svm_tuned <- tibble(y = df$y, y_hat = predict(tune_model_svm2$best.model, df))

df3 <- bind_cols(df2, y_hat_tuned = prediction_svm_tuned$y_hat)

ggplot(df3, aes(x = x, y = y)) + 
  geom_point(col = "blue") + 
  geom_point(aes(x = x, y = y_hat), col = "red") + 
  geom_point(aes(x = x, y = y_hat_tuned), col = "green") + 
  geom_line(linetype = "dashed")

13.1.3 Discussion on parameters

The gamma parameter in the RBF kernel determines the reach of a single training instance. If the value of Gamma is low, then every training instance will have a far reach. Conversely, high values of gamma mean that training instances will have a close reach. So, with a high value of gamma, the SVM decision boundary will simply be dependent on just the points that are closest to the decision boundary, effectively ignoring points that are farther away. In comparison, a low value of gamma will result in a decision boundary that will consider points that are further from it. As a result, high values of gamma typically produce highly flexed decision boundaries, and low values of gamma often results in a decision boundary that is more linear.

http://scikit-learn.org/stable/modules/svm.html

https://www.youtube.com/watch?v=m2a2K4lprQw

http://haohanw.blogspot.com/2014/03/ml-how-sigma-matters-in-svm-rbf-kernel.html

https://link.springer.com/protocol/10.1007/978-1-60327-241-4_13

13.2 References

Intro from kdnugget The R code with the creating data part is from a RPUB document and the fine tuning and using it for further prediction