This is this second post of the “Create your Machine Learning library from scratch with R !” series. Today, we will see how you can implement K nearest neighbors (KNN) using only the linear algebra available in R. Previously, we managed to implement PCA and next time we will deal with SVM and decision trees.
The Knearest neighbors (KNN) is a simple yet efficient classification and regression algorithm. KNN assumes that an observation will be similar to its K closest neighbors. For instance, if most of the neighbors of a given point belongs to a given class, it seems reasonable to assume that the point will belong to the same given class.
The mathematics of KNN
Now, let’s quickly derive the mathematics used for KNN regression (they are similar for classification).
Let be the observations of our training dataset. The points are in . We denote the variable we seek to estimate. We know its value for the train dataset.
Let be a new point in . We do not know and will estimate it using our train dataset.
Let be a positive and nonzero integer (the number of neighbors used for estimation). We want to select the points from the dataset which are the closest to . To do so, we compute the euclidean distance . From all the distance, we can compute , the smallest radius of the circle centered on which includes exactly points from the training sample.
An estimation of is now easy to construct. This is the mean of the of the closest points to :
KNN regression in R
First, we build a “my_knn_regressor” object which stores all the training points, the value of the target variable and the number of neighbors to use.
my_knn_regressor < function(x,y,k=5) { if (!is.matrix(x)) { x < as.matrix(x) } if (!is.matrix(y)) { y < as.matrix(y) } my_knn < list() my_knn[['points']] < x my_knn[['value']] < y my_knn[['k']] < k attr(my_knn, "class") < "my_knn_regressor" return(my_knn) }
The tricky part of KNN is to compute efficiently the distance. We will use the function we created in our previous post on vectorization. The function and mathematical derivations are specified in this post.
gramMatrix < function(X,Y) { tcrossprod(X, Y) } compute_pairwise_distance=function(X,Y) { xn < rowSums(X ** 2) yn < rowSums(Y ** 2) outer(xn, yn,'+')  2 * tcrossprod(X, Y) }
Now we can build our predictor:
predict.my_knn_regressor < function(my_knn,x) { if (!is.matrix(x)) { x=as.matrix(x) } ##Compute pairwise distance dist_pair < compute_pairwise_distance(x,my_knn[['points']]) ##as.matrix(apply(dist_pair,2,order) <= my_knn[['k']]) orders the points by distance and select the kclosest points ##The M[i,j]=1 if x_j is on the k closest point to x_i t(as.matrix(apply(dist_pair,2,order) <= my_knn[['k']])) %*% my_knn[['value']] / my_knn[['k']] }
The last line may seem complicated:

apply(dist_pair,2,order)
orders the points by distance 
apply(dist_pair,2,order)<=my_knn[['k']]
selects the kclosest points to each point in our new dataset 
M=t(as.matrix(apply(dist_pair,2,order) <= my_knn[['k']]))
cast the matrix into a one hot matrix. if is one of the k closest points to . is zero otherwise. 
M %*% my_knn[['value']] / my_knn
sums the value of the k closest points and normalises it by k
KNN Binary Classification in R
The previous code can be reused as it is for binary classification. Your outcome should be encoded as a onehot variable. If the estimated output is greater (resp. less) than 0.5, you can assume that your point belongs to the class encoded as one (resp. zero). We will use the classical Iris dataset and classify the setosa versus the virginica specy.
iris_class < iris[iris[["Species"]]!="versicolor",] iris_class[["Species"]] < as.numeric(iris_class[["Species"]]!="setosa") knn_class < my_knn_regressor(iris_class[,1:2],iris_class[,5]) predict(knn_class,iris_class[,1:2])
Since, we only used 2 variables, we can easily plot the decision boundaries on a 2D plot.
#Build grid x_coord < seq(min(iris_class[,1])  0.2,max(iris_class[,1]) + 0.2,length.out = 200) y_coord < seq(min(iris_class[,2]) 0.2,max(iris_class[,2]) + 0.2 , length.out = 200) coord < expand.grid(x=x_coord, y=y_coord) #predict probabilities coord[['prob']] < predict(knn_class,coord[,1:2]) library(ggplot2) ggplot() + ##Ad tiles according to probabilities geom_tile(data=coord,mapping=aes(x, y, fill=prob)) + scale_fill_gradient(low = "lightblue", high = "red") + ##add points geom_point(data=iris_class,mapping=aes(Sepal.Length,Sepal.Width, shape=Species),size=3 ) + #add the labels to the plots xlab('Sepal length') + ylab('Sepal width') + ggtitle('Decision boundaries of KNN')+ #remove grey border from the tile scale_x_continuous(expand=c(0,0)) + scale_y_continuous(expand=c(0,0))
And this gives us this cool plot:
Possible extensions
Our current KNN is basic, but you can improve and test it in several ways:
 What is the influence of the number of neighbors ? (You should see some overfitting/underfitting)
 Can you implement other metrics than distance ? Can you create kernel KNNs ?
 Instead of doing estimations using only the mean, could you use a more complex mapping ?
Thanks for reading ! To find more posts on Machine Learning, Python and R, you can follow us on Facebook or Twitter.
The post Create your Machine Learning library from scratch with R ! (3/5) – KNN appeared first on Enhance Data Science.
Rbloggers.com offers daily email updates about R news and tutorials on topics such as: Data science, Big Data, R jobs, visualization (ggplot2, Boxplots, maps, animation), programming (RStudio, Sweave, LaTeX, SQL, Eclipse, git, hadoop, Web Scraping) statistics (regression, PCA, time series, trading) and more...