### How does the EM algorithm (in the context of GMM) compare to K-Means?

K-Means aims to minimize the Within Cluster Sum of Squares, while EM aims to maximize the likelihood of an underlying probability distribution.

- Machine Learning 101 (30)
- Statistics 101 (38)
- Supervised Learning (108)
- Regression (36)
- Classification (46)
- Logistic Regression (10)
- Support Vector Machine (10)
- Naive Bayes (4)
- Discriminant Analysis (5)
- Classification Evaluations (9)

- Classification & Regression Trees (CART) (23)

- Unsupervised Learning (46)
- Clustering (17)

- Regularization (6)
- Deep Learning (23)
- Data Preparation (43)
- General (5)
- Standardization (6)
- Missing data (7)
- Textual Data (16)
- Dimensionality Reduction (9)

K-Means aims to minimize the Within Cluster Sum of Squares, while EM aims to maximize the likelihood of an underlying probability distribution.

Pros: Do not have to specify the number of clusters before running the algorithm

Pros: Easy to implement

Cons: Must specify number of clusters in advance

Being that clustering is a distance-based algorithm, outliers can have multiple undesired effects on the quality of the clusters produced.

K-Means ++ has been generally shown to be the best initialization approach to use when performing K-Means clustering.

Using an objective function that minimizes the within-cluster sum of squares (WCSS) causes K-Means to produce spherically shaped clusters.

K-Means minimizes the total within-cluster sum of squares (WCSS)

The final cluster assignments of the K-Means algorithm can be sensitive to the location of the initial centroids.

The most common way to choose k is to run the algorithm over a range of values

K-Means starts by selecting initial centroids for the k-clusters by randomly choosing k observations

Find out all the ways

that you can