Dear reader,

If you are a newbie in the world of machine learning, then this tutorial is exactly what you need in order to introduce yourself to this exciting new part of the data science world.

This post includes a full machine learning project that will guide you step by step to create a “template,” which you can use later on other datasets.

In this step-by-step tutorial you will:

**1.** Use one of the most popular machine learning packages in R.

**2.** Explore a dataset by using statistical summaries and data visualization.

**3.** Build 5 machine-learning models, pick the best, and build confidence that the accuracy is reliable.

The process of a machine learning project may not be exactly the same, but there are certain standard and necessary steps:

**1.** Define Problem.

**2.** Prepare Data.

**3.** Evaluate Algorithms.

**4.** Improve Results.

**5.** Present Results.

**1. PACKAGE INSTALLATION & DATA SET**

The first thing you have to do is install and load the “caret” package with:

`install.packages("caret")`

library(caret)

Moreover, we need a dataset to work with. The dataset we chose in our case is “iris,” which contains 150 observations of iris flowers. There are four columns of measurements of the flowers in centimeters. The fifth column is the species of the flower observed. All observed flowers belong to one of three species. To attach it to the environment, use:

`data(iris)`

**1.1 Create a Validation Dataset**

First of all, we need to validate that our data set is good. Later, we will use statistical methods to estimate the accuracy of the models that we create on unseen data. To be sure about the accuracy of the best model on unseen data, we will evaluate it on actual unseen data. To do this, we will “deposit” some data that the algorithms will not find and use this data later to get a second and independent idea of how accurate the best model really is.

We will split the loaded dataset into two, 80% of which we will use to train our models and 20% of which we will hold back as a validation dataset. Look at the example below:

#create a list of 80% of rows in the original dataset to use them for training

`validation_index <- createDataPartition(dataset$Species, p=0.80, list=FALSE)`

# select 20% of the data for validation

`validation <- dataset[-validation_index,]`

# use the remaining 80% of data to training and testing the models

`dataset <- dataset[validation_index,]`

You now have training data in the dataset variable and a validation set that will be used later in the validation variable.

**2. DATASET SUMMARY**

In this step, we are going to explore our data set. More specifically, we need to know certain features of our dataset, like:

**1.** Dimensions of the dataset.

**2.** Types of the attributes.

**3.** Details of the data.

**4.** Levels of the class attribute.

**5.** Analysis of the instances in each class.

**6.** Statistical summary of all attributes.

**2.1 Dimensions of Dataset**

We can see of how many instances (rows) and how many attributes (columns) the data contains with the dim function. Look at the example below:

`dim(dataset)`

**2.2 Types of Attributes**

Knowing the types is important as it can help you summarize the data you have and possible transformations you might need to use to prepare the data before modeilng. They could be doubles, integers, strings, factors and other types. You can find it with:

`sapply(dataset, class)`

**2.3 Details of the Data**

You can take a look at the first 5 rows of the data with:

`head(dataset)`

**2.4 Levels of the Class**

The class variable is a factor that has multiple class labels or levels. Let’s look at the levels:

`levels(dataset$Species)`

There are two types of classification problems: the multinomial like this one and the binary if there were two levels.

**2.5 Class Distribution**

Let’s now take a look at the number of instances that belong to each class. We can view this as an absolute count and as a percentage with:

`percentage <- prop.table(table(dataset$Species)) * 100`

cbind(freq=table(dataset$Species), percentage=percentage)

**2.6 Statistical Summary**

This includes the mean, the min and max values, as well as some percentiles. Look at the example below:

`summary(dataset)`

**3. DATASET VISUALIZATION**

We now have a basic idea about the data. We need to extend that with some visualizations, and for that reason we are going to use two types of plots:

**1.** Univariate plots to understand each attribute.

**2.** Multivariate plots to understand the relationships between attributes.

**3.1 Univariate Plots**

We can visualize just the input attributes and just the output attributes. Let’s set that up and call the input attributes x and the output attributes y.

`x <- dataset[,1:4]`

y <- dataset[,5]

Since the input variables are numeric, we can create box and whisker plots of each one with:

`par(mfrow=c(1,4))`

for(i in 1:4) {

boxplot(x[,i], main=names(iris)[i])

}

We can also create a barplot of the Species class variable to graphically display the class distribution.

`plot(y)`

**3.2 Multivariate Plots**

First, we create scatterplots of all pairs of attributes and color the points by class. Then, we can draw ellipses around them to make them more easily separated.

You have to install and call the “ellipse” package to do this.

`install.packages("ellipse")`

library(ellipse)

featurePlot(x=x, y=y, plot="ellipse")

We can also create box and whisker plots of each input variable, but this time they are broken down into separate plots for each class.

`featurePlot(x=x, y=y, plot="box")`

Next, we can get an idea of the distribution of each attribute. We will use some probability density plots to give smooth lines for each distribution.

`scales <- list(x=list(relation="free"), y=list(relation="free"))`

featurePlot(x=x, y=y, plot="density", scales=scales)

**4. ALGORITHMS EVALUATION**

Now it is time to create some models of the data and estimate their accuracy on unseen data.

**1.** Use the test harness to use 10-fold cross validation.

**2.** Build 5 different models to predict species from flower measurements.

**3.** Select the best model.

**4.1 Test Harness**

This will split our dataset into 10 parts, train in 9, test on 1, and release for all combinations of train-test splits.

`control <- trainControl(method="cv", number=10)`

metric <- "Accuracy"

We are using the metric of “Accuracy” to evaluate models. This is: (number of correctly predicted instances / divided by the total number of instances in the dataset)*100 to give a percentage.

**4.2 Build Models**

We don’t know which algorithms would be good on this problem or what configurations to use. We get an idea from the plots that we created earlier.

Algorithms evaluation:

**1.** Linear Discriminant Analysis (LDA)

**2.** Classification and Regression Trees (CART).

**3.** k-Nearest Neighbors (kNN).

**4.** Support Vector Machines (SVM) with a linear kernel.

**5.** Random Forest (RF)

This is a good mixture of simple linear (LDA), nonlinear (CART, kNN) and complex nonlinear methods (SVM, RF). We reset the random number seed before reach run to ensure that the evaluation of each algorithm is performed using exactly the same data splits. It ensures the results are directly comparable.

**NOTE**: To proceed, first install and load the following packages: “rpart”, “kernlab”, “e1071” and “randomForest”.

Let’s build our five models:

# a) linear algorithms

`set.seed(7)`

fit.lda <- train(Species~., data=dataset, method="lda", metric=metric, trControl=control)

# b) nonlinear algorithms

# CART

`set.seed(7)`

fit.cart <- train(Species~., data=dataset, method="rpart", metric=metric, trControl=control)

# kNN

`set.seed(7)`

fit.knn <- train(Species~., data=dataset, method="knn", metric=metric, trControl=control)

# c) advanced algorithms

# SVM

`set.seed(7)`

fit.svm <- train(Species~., data=dataset, method="svmRadial", metric=metric, trControl=control)

# Random Forest

`set.seed(7)`

fit.rf <- train(Species~., data=dataset, method="rf", metric=metric, trControl=control)

**4.3 Select the Best Model**

We now have 5 models and accuracy estimations for each so we have to compare them.

It is a good idea to create a list of the created models and use the summary function.

`results <- resamples(list(lda=fit.lda, cart=fit.cart, knn=fit.knn, svm=fit.svm, rf=fit.rf))`

summary(results)

Moreover, we can create a plot of the model evaluation results and compare the spread and the mean accuracy of each model. There is a population of accuracy measures for each algorithm because each algorithm was evaluated 10 times.

`dotplot(results)`

You can summarize the results for just the LDA model that seems to be the most accurate.

`print(fit.lda)`

**5. Make Predictions**

The LDA was the most accurate model. Now we want to get an idea of the accuracy of the model on our validation set.

We can run the LDA model directly on the validation set and summarize the results in a confusion matrix.

`predictions <- predict(fit.lda, validation)`

confusionMatrix(predictions, validation$Species)

**What's next:**

- Become a Top R Programmer Fast with our Individual Coaching Program
- Explore all our (>4000) R exercises
- Find an R course using our R Course Finder directory
- Subscribe to receive weekly updates and bonus sets by email
- Share with your friends and colleagues using the buttons below

Sherri Walker says

I am fascinated by data science but am not a practitioner. For a new R student like me this was useful. It helped me understand the general ML process and how to set it up in R, which I assume is your objective.

The exercise implies that the LDA was the most accurate model, yet in the results it appeared as though the knn was stronger. Does your choice of LDA relate to the k=7 in the knn?

Efthymios Kasvikis says

My choise is based on this:

results <- resamples(list(lda=fit.lda, cart=fit.cart, knn=fit.knn, svm=fit.svm, rf=fit.rf))

summary(results)

and this:

# compare accuracy of models

dotplot(results)

What results did you find when you ran the solution?

zcatav says

Sensitivity of class virginica for lda is 0.7 and for knn is 1.0. KNN seems best for final results.

> predictions confusionMatrix(predictions, validation$Species)

Confusion Matrix and Statistics

Reference

Prediction setosa versicolor virginica

setosa 10 0 0

versicolor 0 10 3

virginica 0 0 7

Overall Statistics

Accuracy : 0.9

95% CI : (0.7347, 0.9789)

No Information Rate : 0.3333

P-Value [Acc > NIR] : 1.665e-10

Kappa : 0.85

Mcnemar’s Test P-Value : NA

Statistics by Class:

Class: setosa Class: versicolor Class: virginica

Sensitivity 1.0000 1.0000 0.7000

Specificity 1.0000 0.8500 1.0000

Pos Pred Value 1.0000 0.7692 1.0000

Neg Pred Value 1.0000 1.0000 0.8696

Prevalence 0.3333 0.3333 0.3333

Detection Rate 0.3333 0.3333 0.2333

Detection Prevalence 0.3333 0.4333 0.2333

Balanced Accuracy 1.0000 0.9250 0.8500

> predictions confusionMatrix(predictions, validation$Species)

Confusion Matrix and Statistics

Reference

Prediction setosa versicolor virginica

setosa 10 0 0

versicolor 0 10 0

virginica 0 0 10

Overall Statistics

Accuracy : 1

95% CI : (0.8843, 1)

No Information Rate : 0.3333

P-Value [Acc > NIR] : 4.857e-15

Kappa : 1

Mcnemar’s Test P-Value : NA

Statistics by Class:

Class: setosa Class: versicolor Class: virginica

Sensitivity 1.0000 1.0000 1.0000

Specificity 1.0000 1.0000 1.0000

Pos Pred Value 1.0000 1.0000 1.0000

Neg Pred Value 1.0000 1.0000 1.0000

Prevalence 0.3333 0.3333 0.3333

Detection Rate 0.3333 0.3333 0.3333

Detection Prevalence 0.3333 0.3333 0.3333

Balanced Accuracy 1.0000 1.0000 1.0000

Kevin says

Actually featurePlot() is in the caret package not the ellipse package

Rodrigo says

Hello,

Everything went fine until the last lines of code in this exercise, when:

predictions <- predict(fit.lda, validation)

Resulted in:

Error in eval(predvars, data, env) : object 'Sepal.Length' not found

I repeated the whole exercise to see if anything went missing, but it looks like I can't find a reason. Did anyone have the same problem?

I'm an R beginner, so it's difficult for me to understand what might have happened.

R.