# 7.3: Machine learning

- Page ID
- 3584

Methods explained in this section are not only visualizations. Together, they frequently called “classification with learning”, “supervised classification”, “machine learning”, or just “classification”. All of them are based on the idea of *learning*:

... He scrambled through and rose to his feet. ... He saw nothing but colours—colours that refused to form themselves into things. Moreover, he knew nothing yet well enough to see it: you cannot see things till you know roughly what they are\(^{[1]}\). His first impression was of a bright, pale world—a watercolour world out

of a child’s paint-box; a moment later he recognized the flat belt of light blue as a sheet of water, or of something like water, which came nearly to his feet. They were on the shore of a lake or river...

C.S.Lewis. Out of the Silent Planet.

First, small part of data where identity is already known (*training dataset*) used to develop (fit) the model of classification (Figure \(\PageIndex{2}\)). On the next step, this model is used to classify objects with unknown identity (*testing dataset*). In most of these methods, it is possible to estimate the quality of the classification and also assess the significance of the every character.

Let us create training and testing datasets from iris data:

**Code \(\PageIndex{1}\) (Python):**

iris.train <- iris[seq(1, nrow(iris), 5), ] iris.unknown <- iris[-seq(1, nrow(iris), 5), ]

(iris.unknown is of course the fake unknown so to use it properly, we must specify iris.unknown[, -5]. On the other hand, species information will help to create misclassification table (confusion matrix, see below).)

## Learning with regression

### Linear discriminant analysis

One of the simplest methods of classification is the linear discriminant analysis (LDA). The basic idea is to create the set of linear functions which “decide” how to classify the particular object.

**Code \(\PageIndex{2}\) (Python):**

iris.train <- iris[seq(1, nrow(iris), 5), ] iris.unknown <- iris[-seq(1, nrow(iris), 5), ] library(MASS) iris.lda <- lda(Species ~ . , data=iris.train) iris.predicted <- predict(iris.lda, iris.unknown[, 1:4]) Misclass(iris.predicted$class, iris.unknown[, 5]) # asmisc.r

Training resulted in the hypothesis which allowed almost all plants (with an exception of seven *Iris virginica*) to be placed into the proper group. Please note that LDA does not require scaling of variables.

It is possible to check LDA results with inferential methods. Multidimensional analysis of variation (MANOVA) allows to understand the relation between data and model (classification from LDA):

**Code \(\PageIndex{3}\) (Python):**

iris.unknown <- iris[-seq(1, nrow(iris), 5), ] ldam <- manova(as.matrix(iris.unknown[, 1:4]) ~ iris.predicted$class) summary(ldam, test="Wilks")

Important here are both p-value based on Fisher statistics, and also the value of Wilks’ statistics which is the *likelihood ratio* (in our case, the probability that groups are *not different*).

It is possible to check the relative importance of every character in LDA with ANOVA-like techniques:

**Code \(\PageIndex{4}\) (Python):**

iris.unknown <- iris[-seq(1, nrow(iris), 5), ] summary(aov(as.matrix(iris.unknown[, 1:4]) ~ iris.predicted$class))

(This idea is applicable to other classification methods too.)

... and also visualize LDA results (Figure \(\PageIndex{3}\)):

**Code \(\PageIndex{5}\) (Python):**

library(MASS) iris.lda2 <- lda(iris[, 1:4], iris[, 5]) iris.ldap2 <- predict(iris.lda2, dimen=2)$x plot(iris.ldap2, type="n", xlab="LD1", ylab="LD2") text(iris.ldap2, labels=abbreviate(iris[, 5], 1, method="both.sides")) Ellipses(iris.ldap2, as.numeric(iris[, 5]), centers=TRUE) # asmisc.r

(Please note 95% confidence ellipses with centers.)

To place all points on the plot, we simply used all data as training. Note the good discrimination (higher than in PCA, MDS or clustering), even between close *Iris versicolor* and *I. virginica*. This is because LDA frequently overestimates the differences between groups. This feature, and also the parametricity and linearity of LDA made it less used over the last years.

With LDA, it is easy to illustrate one more important concept of machine learning: *quality of training*. Consider the following example:

**Code \(\PageIndex{6}\) (Python):**

set.seed(230) iris.sample2 <- sample(1:nrow(iris), 30) iris.train2 <- iris[iris.sample2, ] iris.unknown2 <- iris[-iris.sample2, ] library(MASS) iris.lda2 <- lda(Species ~ . , data=iris.train2) iris.predicted2 <- predict(iris.lda2, iris.unknown2[, 1:4]) Misclass(iris.predicted2$class, iris.unknown2[, 5])

Misclassification error here almost two times bigger! Why?

**Code \(\PageIndex{7}\) (Python):**

iris.train <- iris[seq(1, nrow(iris), 5), ] set.seed(230) iris.sample2 <- sample(1:nrow(iris), 30) iris.train2 <- iris[iris.sample2, ] table(iris.train2$Species)

Well, using sample() (and particular set.seed() value) resulted in biased training sample, this is why our second model was trained so poorly. Our first way to sample (every 5th iris) was better, and if there is a need to use sample(), consider to sample each species *separately*.

Now, return to the default random number generator settings:

**Code \(\PageIndex{8}\) (Python):**

set.seed(NULL)

Please note that it is widely known that while LDA was developed on biological material, this kind of data rarely meets two key assumptions of this method: (1) multivariate normality and (2) multivariate homoscedasticity. Amazingly, even Fisher’s *Iris* data with which LDA was invented, does not meet these assumptions! Therefore, we do not recommend to use LDA and keep it here mostly for teaching purposes.

### Recursive partitioning

To replace linear discriminant analysis, multiple methods with similar background ideas were invented. Recursive partitioning, or *decision trees* (regression trees, classification trees), allow, among other, to make and visualize the sort of discrimination key where every step results in splitting objects in two groups (Figure \(\PageIndex{4}\)):

**Code \(\PageIndex{9}\) (Python):**

set.seed(NULL) library(tree) iris.tree <- tree(Species ~ ., data=iris) plot(iris.tree) text(iris.tree)

We loaded first the tree package containing tree() function (rpart is another package which makes classification trees). Then we again used the whole dataset as training data. The plot shows that all plants with petal length less than 2.45 cm belong to *Iris setosa*, and from the rest those plants which have petal width less than 1.75 cm and petal length more than 4.95 cm, are *I. versicolor*; all other irises belong to *I. virginica*.

In fact, these trees are result of something similar to “hierarchical discriminant analysis”, and it is possible to use them for the supervised classification:

**Code \(\PageIndex{10}\) (Python):**

iris.train <- iris[seq(1, nrow(iris), 5), ] iris.unknown <- iris[-seq(1, nrow(iris), 5), ] iris.tree2 <- tree(Species ~ ., data=iris.train) iris.tp2 <- predict(iris.tree2, iris.unknown[,-5], type="class") library(MASS) Misclass(iris.tp2, iris.unknown[, 5])

Try to find out which characters distinguish species of horsetails described in eq.txt data file. File eq_c.txt contains the description of characters.

Package party offers sophisticated recursive partitioning methods together with advanced tree plots (Figure \(\PageIndex{5}\)):

**Code \(\PageIndex{11}\) (Python):**

library(party) SA <- abbreviate(iris$Species, 1, method="both.sides") iris.ct <- ctree(factor(SA) ~ ., data=iris[, 1:4]) plot(iris.ct) library(MASS) Misclass(SA, predict(iris.ct))

(For species names, we used one-letter abbreviations.)

## Ensemble learnig

### Random Forest

The other method, internally similar to regression trees, rapidly gains popularity. This is the *Random Forest*. Its name came from the ability to use numerous decision trees and build the complex classification model. Random Forest belongs to *bagging ensemble methods*; it uses bootstrap (see in Appendix) to multiply the number of trees in the model (hence “forest”). Below is an example of Random Forest classifier made from the iris data:

**Code \(\PageIndex{12}\) (Python):**

iris.train <- iris[seq(1, nrow(iris), 5), ] iris.unknown <- iris[-seq(1, nrow(iris), 5), ] library(randomForest) set.seed(17) iris.rf <- randomForest(Species ~ ., data=iris.train) iris.rfp <- predict(iris.rf, iris.unknown[,-5]) library(MASS) Misclass(iris.rfp, iris.unknown[, 5])

Here results are similar to LDA but Random Forest allows for more. For example, it can clarify the importance of each character (with function `importance()`

), and reveal classification distances (proximities) between all objects of training subset (these distances could be in turn used for clustering). Random Forest could also visualize the multidimensional dataset (Figure \(\PageIndex{6}\)):

**Code \(\PageIndex{13}\) (Python):**

library(randomForest) set.seed(17) # because plot is random iris.urf <- randomForest(iris[,-5]) iris.rfm <- MDSplot(iris.urf, iris[, 5], xlab="", ylab="", pch=abbreviate(iris[, 5], 1, method="both.sides")) Pal <- brewer.pal(nlevels(iris[, 5]), "Set1") Hulls(iris.rfm$points, as.numeric(iris[, 5]), centers=TRUE, usecolor=Pal) # asmisc.r

(We applied several tricks to show convex hulls and their centroids.)

Package ranger implements even faster variant of Random Forest algorithm, it also can employ parallel calculations.

### Gradient boosting

There are many weak classification methods which typically make high misclassification errors. However, many of them are also ultra-fast. So, is it possible to combine many weak learners to make the strong one? Yes! This is what boosting methods do. Gradient boosting employs multi-step optimization and is now among most frequently using learning techniques. In R, there are several gradient boosting packages, for example, xgboost and gbm:

**Code \(\PageIndex{14}\) (Python):**

iris.train <- iris[seq(1, nrow(iris), 5), ] iris.unknown <- iris[-seq(1, nrow(iris), 5), ] library(gbm) set.seed(4) iris.gbm <- gbm(Species ~ ., data=rbind(iris.train, iris.train)) # to make training bigger Distribution not specified, assuming multinomial ... plot(iris.gbm) iris.gbm.p1 <- predict.gbm(iris.gbm, iris.unknown, n.trees=iris.gbm$n.trees) iris.gbm.p2 <- apply(iris.gbm.p1, 1, # membership trick function(.x) colnames(iris.gbm.p1)[which.max(.x)]) library(MASS) Misclass(iris.gbm.p2, iris.unknown[, 5])

(Plot is purely technical; in the above form, it will show the marginal effect (effect on membership) of the 1st variable. Please **make** it yourself. “Membership trick” selects the “best species” from three alternatives as gbm() reports classification result in fuzzy form.)

## Learning with proximity

*k-Nearest Neighbors algorithm* (or kNN) is the “lazy classifier” because it does not work until unknown data is supplied:

**Code \(\PageIndex{15}\) (Python):**

iris.train <- iris[seq(1, nrow(iris), 5), ] iris.unknown <- iris[-seq(1, nrow(iris), 5), ] library(class) iris.knn.pred <- knn(train=iris.train[,-5], test=iris.unknown[,-5], cl=iris.train[, 5], k=5) library(MASS) Misclass(iris.knn.pred, iris.unknown[, 5])

kNN is based on *distance calculation* and “voting”. It calculates distances from every unknown object to the every object of the training set. Next, it considers several (5 in the case above) nearest neighbors with known identity and finds which id is prevalent. This prevalent id assigned to the unknown member. Function knn() uses Euclidean distances but in principle, any distance would work for kNN.

To illustrate idea of nearest neighbors, we use *Voronoi decomposition*, the technique which is close to both kNN and distance calculation:

**Code \(\PageIndex{16}\) (Python):**

iris.train <- iris[seq(1, nrow(iris), 5), ] iris.unknown <- iris[-seq(1, nrow(iris), 5), ] iris.p <- prcomp(iris[, 1:4], scale=TRUE)$x[, 1:2] iris.p1 <- iris.p[seq(1, nrow(iris.p), 5), ] iris.p2 <- iris.p[-seq(1, nrow(iris.p), 5), ] library(tripack) iris.v <- voronoi.mosaic(iris.p1[, 1], iris.p1[, 2], duplicate="remove") plot(iris.v, do.points=FALSE, main="", sub="") points(iris.p1[, 1:2], col=iris.train$Species, pch=16, cex=2) points(iris.p2[, 1:2], col=iris.unknown$Species)

The plot (Figure \(\PageIndex{7}\)) contains multiple cells which represent neighborhoods of training sample (big dots). This is not exactly what kNN does, but idea is just the same. In fact, Voronoi plot is a good tool to visualize any distance-based approach.

*Depth classification* based on how close an arbitrary point of the space is located to an implicitly defined center of a multidimensional data cloud:

**Code \(\PageIndex{17}\) (Python):**

iris.train <- iris[seq(1, nrow(iris), 5), ] iris.unknown <- iris[-seq(1, nrow(iris), 5), ] library(ddalpha) iris.dd <- ddalpha.train(Species ~ ., data=iris.train) iris.p <- predict(iris.dd, iris.unknown[, -5]) library(MASS) Misclass(unlist(iris.p), iris.unknown[, 5]) # asmisc.r iris.pp <- predict(iris.dd, iris.unknown[, -5], outsider.method="Ignore") sapply(iris.pp, as.character) # shows points outside train clouds

## Learning with rules

*Naïve Bayes classifier* is one of the simplest machine learning algorithms which tries to classify objects based on the probabilities of previously seen attributes. Quite unexpectedly, it is typically a good classifier:

**Code \(\PageIndex{18}\) (Python):**

iris.train <- iris[seq(1, nrow(iris), 5), ] iris.unknown <- iris[-seq(1, nrow(iris), 5), ] library(e1071) iris.nb <- naiveBayes(Species ~ ., data=iris.train) iris.nbp <- predict(iris.nb, iris.unknown[,-5]) library(MASS) Misclass(iris.nbp, iris.unknown[, 5]) # asmisc.r

Note that Naïve Bayes classifier could use not only numerical like above, but also nominal predictors (which is similar to correspondence analysis.)

*Apriori* method is similar to regression trees but instead of classifying objects, it researches *association rules* between classes of objects. This method could be used not only to find these rules but also to make classification. Note that measurement iris data is less suitable for association rules then nominal data, and it needs *discretization* first:

**Code \(\PageIndex{19}\) (Python):**

library(arulesCBA) irisd <- as.data.frame(lapply(iris[1:4], discretize, categories=9)) irisd$Species <- iris$Species irisd.train <- irisd[seq(1, nrow(irisd), 5), ] irisd.unknown <- irisd[-seq(1, nrow(irisd), 5), ] irisd.cba <- CBA(Species ~ ., irisd.train, supp=0.05, conf=0.8) inspect(irisd.cba$rules) irisd.cbap <- predict(irisd.cba, irisd.unknown) library(MASS) Misclass(irisd.cbap, irisd.unknown$Species)

(Rules are self-explanatory. What do you think, does this method performs better for the nominal data? Please **find** it out.)

## Learning from the black box

Famous SVM, *Support Vector Machines* is a *kernel* technique which calculates parameters of the hyper-planes dividing multiple groups in the multidimensional space of characters:

**Code \(\PageIndex{20}\) (Python):**

iris.train <- iris[seq(1, nrow(iris), 5), ] iris.unknown <- iris[-seq(1, nrow(iris), 5), ] library(e1071) iris.svm <- svm(Species ~ ., data=iris.train) iris.svmp <- predict(iris.svm, iris.unknown[,-5]) Misclass(iris.svmp, iris.unknown[, 5]) # asmisc.r

Classification, or prediction *grid* often helps to illustrate the SVM method. Data points are arranged with PCA to reduce dimensionality, and then classifier predicts the identity for the every point in the artificially made grid (Figure \(\PageIndex{8}\)). This is possible to perform manually but Gradd() function simplifies plotting:

**Code \(\PageIndex{21}\) (Python):**

iris.p <- prcomp(iris[, 1:4], scale=TRUE)$x[, 1:2] iris.svm.pca <- svm(Species ~ ., data=cbind(iris[5], iris.p)) plot(iris.p, type="n", xlab="", ylab="") Gradd(iris.svm.pca, iris.p) # gmoon.r text(iris.p, col=as.numeric(iris[, 5]), labels=abbreviate(iris[, 5], 1, method="both.sides"))

And finally, *neural networks*! This name is used for the statistical technique based on some features of neural cells, *neurons*. First, we need to prepare data and convert categorical variable Species into three logical *dummy variables*:

**Code \(\PageIndex{22}\) (Python):**

iris.train <- iris[seq(1, nrow(iris), 5), ] iris.dummy <- Tobin(iris.train$Species, convert.names=FALSE) # asmisc.r iris.train2 <- cbind(iris.train[, -5], iris.dummy) str(iris.train2)

Now, we call neuralnet package and proceed to the main calculation. The package “wants” to supply all terms in the model explicitly:

**Code \(\PageIndex{23}\) (Python):**

iris.train <- iris[seq(1, nrow(iris), 5), ] iris.dummy <- Tobin(iris.train$Species, convert.names=FALSE) # asmisc.r iris.train2 <- cbind(iris.train[, -5], iris.dummy) library(neuralnet) set.seed(17) iris.n <- neuralnet(setosa + versicolor + virginica ~ Sepal.Length + Sepal.Width + Petal.Length + Petal.Width, data=iris.train2, hidden=3, lifesign="full")

(Note use of set.seed(), this is to make your results similar to presented here.)

Now predict (with compute() function) and check misclassification:

**Code \(\PageIndex{24}\) (Python):**

iris.train <- iris[seq(1, nrow(iris), 5), ] iris.unknown <- iris[-seq(1, nrow(iris), 5), ] iris.dummy <- Tobin(iris.train$Species, convert.names=FALSE) # asmisc.r iris.np <- compute(iris.n, iris.unknown[,-5]) iris.np2 <- apply(iris.np$net.result, 1, function(.x) colnames(iris.dummy)[which.max(.x)]) Misclass(iris.np2, iris.unknown[, 5])

Results of neural network prediction are fuzzy, similar to the results of fuzzy clustering or regression trees, this is why which.max() was applied for every row. As you see, this is one of the lowest misclassification errors.

It is possible to plot the actual network:

**Code \(\PageIndex{25}\) (Python):**

iris.train <- iris[seq(1, nrow(iris), 5), ] iris.dummy <- Tobin(iris.train$Species, convert.names=FALSE) # asmisc.r iris.train2 <- cbind(iris.train[, -5], iris.dummy) library(neuralnet) set.seed(17) iris.n <- neuralnet(setosa + versicolor + virginica ~ Sepal.Length + Sepal.Width + Petal.Length + Petal.Width, data=iris.train2, hidden=3, lifesign="full") plot(iris.n, rep="best", intercept=FALSE)

The plot (Figure 7.4.1) is a bit esoteric for the newbie, but hopefully will introduce into the method because there is an apparent multi-layered structure which is used for neural networks decisions.

## References

1. Emphasis mine.