Clustering and classification ======================================================== type: sub-section For IODS by Emma Kämäräinen



Powered by Rpresentation. The code for this presentation is [here.](https://raw.githubusercontent.com/TuomoNieminen/Helsinki-Open-Data-Science/master/docs/cluster_classification.Rpres) ```{r, include = F} # setup knitr::opts_chunk$set(echo = F, comment = NA) mytheme <- function(axt_text = 18, axt_title = 22 ) { ggplot2::theme(axis.text=element_text(size = axt_text), axis.title=element_text(size = axt_title, face="bold")) } ``` Clustering and classification ======================================================== incremental: false Classification: - You know the classes - The classification model is trained based on data - Classify new observations *** Clustering: - Unknown classes / number of classes is unknown - Find groups within data based on similarity of the observations Clustering and classification ======================================================== incremental: false autosize: true - Linear discriminant analysis - Distance measures - K-means Linear discriminant analysis ================================================== autosize: true Linear discriminant analysis (LDA) is a classification method. It can be used to model binary variables, like in logistic regression, or multiple class variables. The target variable needs to be categorical. It can be used to - Find the variables that discriminate/separate the classes best - Predict the classes of new data - Dimension reduction (not covered here) [This](http://machinelearningmastery.com/linear-discriminant-analysis-for-machine-learning/) is a good and simple blog post about LDA. R-Bloggers also have a post about LDA, see it [here](https://www.r-bloggers.com/computing-and-visualizing-lda-in-r/). Linear discriminant analysis ================================================== autosize: true Linear discriminant analysis produces results based on the assumptions that - variables are normally distributed (on condition of the classes) - the normal distributions for each class share the same covariance matrix Because of the assumptions, the data might need scaling before fitting the model. The variables also need to be continuous. Let's see an example next to wrap our heads around what LDA is really doing. LDA example ================================================== autosize: true ```{r, echo=FALSE, } library(MASS) data(iris) d <- scale(iris[,1:4]) Species <- as.factor(iris$Species) d<-data.frame(d, Species) fit <- lda(Species ~., data = d) fit ``` LDA predictions ================================================== autosize: true Classifying new observations: - Based on the trained model LDA calculates the probabilities for the new observation for belonging in each of the classes - The observation is classified to the class of the highest probability - The math behind the probabilities can be seen [here](http://scikit-learn.org/stable/modules/lda_qda.html) for those who are interested. [Bayes theorem](https://en.wikipedia.org/wiki/Bayes'_theorem) is used to estimate the probabilities. - You'll see how the predicting is done in the DataCamp exercises. LDA example ================================================== autosize: true ```{r, echo=FALSE, fig.height=15, fig.width=19} library(RColorBrewer) col <- brewer.pal(8, 'Set1') palette(col) lda.arrows <- function(x, myscale = 1, arrow_heads = 0.2, color = "orange", tex = 2, choices = c(1,2), lwd= 2){ heads <- coef(x) arrows(x0 = 0, y0 = 0, x1 = myscale * heads[,choices[1]], y1 = myscale * heads[,choices[2]], col=color, length = arrow_heads, lwd=2) text(myscale * heads[,choices], labels = row.names(heads), cex = tex, col=color, pos=3, lwd=2.5) } par(cex.label=2, cex.lab=2, cex =1.5, mar=c(5,6,3,3)) plot(fit, col=as.numeric(d$Species), cex =1.5) lda.arrows(fit) ``` Distance measures ================================================== incremental: false autosize: true How to determine if observations are similar or dissimilar with each others? - [Euklidean distance](https://en.wikipedia.org/wiki/Euclidean_distance) - [Manhattan/Taxicab distance](https://en.wikipedia.org/wiki/Taxicab_geometry) (axis-aligned directions) - [Jaccard index](https://en.wikipedia.org/wiki/Jaccard_index) (binary/categorical distance measure) - [Hamming distance](https://en.wikipedia.org/wiki/Hamming_distance) (distance measure for words/strings) K -means ================================================== incremental: false autosize: true - *K*-means is possibly the oldest and used clustering method in many fields of study - Pro: Easy to use and often finds a solution - Con: Small change in the dataset can produce very different results - Many variations of *k*-means: *k*-means++, *k*-medoids, *k*-medians... K-means algorithm ================================================== incremental: false autosize: true 1. Choose the number of clusters you want to have and pick initial cluster centroids. 2. Calculate distances between centroids and datapoints. 3. For all the data points: Assign data point to cluster based on which centroid is closest. 4. Update centroids: within each cluster, calculate new centroid 5. Update clusters: Calculate distances between data points and updated centroids. If some other centroid is closer than the cluster centroid where the data point belongs, the data point changes cluster. Continue updating steps until the centroids or the clusters do not change K-means example ================================================== ```{r, echo=F} data(iris) library(ggplot2) g <- ggplot(iris, aes(Sepal.Length, Sepal.Width, color = Species)) g + geom_point(size = 4) + mytheme() ``` *** ```{r, echo=F} set.seed(20) km <- kmeans(iris[,1:4],3) cluster<- as.factor(km$cluster) iris <- data.frame(iris, cluster) g <- ggplot(iris, aes(Sepal.Length, Sepal.Width, color = cluster)) g + geom_point(size = 4) + mytheme() ``` Source: [This R-Bloggers Post](https://www.r-bloggers.com/k-means-clustering-in-r/) K-means notes ================================================== incremental: false autosize: true Remarks about *k*-means: - Distance measure in the algorithm: - Different distance measures produce different output - Deciding the best distance is not always easy - Number of clusters as input - Many ways to find the optimal number of clusters - One way is to look at the total within cluster sum of squares (see next slide) - [Other ways](https://en.wikipedia.org/wiki/Determining_the_number_of_clusters_in_a_data_set): hierarchical clustering, silhouette method, cross validation ... K-means: Total within sum of squares ================================================== incremental: false autosize: true Total within sum of squares is calculated by adding the within cluster sum of squares (WCSS) of every cluster together. The WCSS can be calculated with the pattern $WCSS = \sum_i^N (x_i - centroid)^2$ So you are searching for the number of clusters, where the observations are closest to the cluster center.