You are on page 1of 63

Data Mining (DM)

GTU #3160714

Unit-5
Clustering

Prof. Naimish R. Vadodariya


Computer Engineering Department
Darshan Institute of Engineering & Technology, Rajkot
naimish.vadodariya@darshan.ac.in
8866215253
 Topics
Loopingto be covered

• Clustering
• Problem Definition, Clustering Overview, Evaluation of Clustering
Algorithms
• Partitioning Clustering -K-Means Algorithm
• PAM Algorithm
• Hierarchical Clustering – Agglomerative Methods and divisive methods
• Basic Agglomerative Algorithm
Clustering : Introduction
Section - 1
What is Clustering?
Clustering is one of the most important research areas in the field of data mining.
Cluster analysis or clustering is the task of grouping a set of objects in such a way that
objects in the same group (called a cluster) are more similar (in some sense or another) to
each other than to those in other groups (clusters).
It is an unsupervised learning technique.
Data clustering is the subject of active research in several fields such as statistics, pattern
recognition and machine learning.
From a practical perspective clustering plays an outstanding role in data mining applications
in many domains.
The main advantage of clustering is that interesting patterns and structures can be
found directly from very large data sets with little or none of the background
knowledge.
Clustering algorithms an be applied in many areas, like marketing, biology, libraries,
insurance, city-planning, earthquake studies and www document classification.
Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 4
Applications of Clustering
Real life examples where we use clustering:
 Marketing
 Finding group of customers with similar behavior given a large data-base of customers.
 Data containing their properties and past buying records (Conceptual Clustering).
 Biology
 Classification of Plants and Animals Based on the properties under observation (Conceptual
Clustering).
 Insurance
 Identifying groups of car insurance policy holders with a high average claim cost (Conceptual
Clustering).
 City-Planning
 Groups of houses according to their house type, value and geographical location it can be both
(Conceptual Clustering and Distance Based Clustering)
 Libraries
 It is used in clustering different books on the basis of topics and information.
 Earthquake studies
 By learning the earthquake-affected areas we can determine the dangerous zones.
Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 5
Partitioning Clustering
K-Means Algorithm
Section - 2
What is Partitioning?
Clustering is a division of data into groups of similar objects.
Each group, called cluster, consists of objects that are similar between themselves and
dissimilar to objects of other groups.
It represents many data objects by few clusters and hence, it models data by its clusters.
A cluster is a set of points such that any point in a cluster is closer (or more similar) to every
other point in the cluster than to any point not in the cluster.

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 7


K-MEANS Algorithm
K-Means is one of the simplest unsupervised learning algorithm that solve the well known
clustering problem.
The procedure follows a simple and easy way to classify a given data set through a certain
number of clusters (k-clusters).
The main idea is to define k centroids, one for each cluster.
A centroid is “the center of mass of a geometric object of uniform density”, though here,
we'll consider mean vectors as centroids.
It is a method of classifying/grouping items into k groups (where k is the number of pre-
chosen groups).
The grouping is done by minimizing the sum of squared distances between items or objects
and the corresponding centroid.

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 8


K-MEANS Algorithm Cont..

A clustered scatter plot.


The black dots are data points.
The red lines illustrate the partitions created
by the k-means algorithm.
The blue dots represent the centroids which
define the partitions.

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 9


K-MEANS Algorithm Cont..
The initial partitioning can be done in a variety of ways.
Dynamically Chosen
 This method is good when the amount of data is expected to grow.
 The initial cluster means can simply be the first few items of data from the set.
 For instance, if the data will be grouped into 3 clusters, then the initial cluster means will be the first 3
items of data.
Randomly Chosen
 Almost self-explanatory, the initial cluster means are randomly chosen values within the same range as the
highest and lowest of the data values.
Choosing from Upper and Lower Bounds
 Depending on the types of data in the set, the highest and lowest of the data range are chosen as the initial
cluster means.

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 10


K-Means Algorithm - Example
Section - 3
K-Means Algorithm - Example
Sr. Height Weight
1 185 72
2 170 56
K First we take K=2 So, two clusters or groups.
3 168 60
(185,72)
1 We choose first (185,72) & second (170,56) row as
4 179 68
centroid of each cluster or group.
5 182 72
Now, we have to find Euclidean Distance,
6 188 77
7 180 71
Where
8 180 70 K  & = Observed Value
9 183 84 2  & = Centroid Value
(170,56)
10 180 88
11 180 67
12 177 76

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 12


K-Means Algorithm – Example Cont..
Sr. Height Weight
1 185 72
2 170 56
K
3 168 60
(185,72)
1
4 179 68
5 182 72
6 188 77
7 180 71
8 180 70 K
9 183 84 2
(170,56)
10 180 88
11 180 67 Now, data (168,60) nearer to K2, so it belongs to K2.
12 177 76 K1 = {1}
K2 = {2,3}
Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 13
K-Means Algorithm – Example Cont..
Sr. Height Weight
1 185 72
2 170 56
K
3 168 60
(185,72)
1
4 179 68
5 182 72 Now, New Centroid Calculation
For K2 = {2,3}
6 188 77
So, K2 = {(170,56),(168,60)}
7 180 71 = 170+168/2 & 56+60/2
8 180 70 K We get new centroid C = (169,58)
9 183 84 2
(169,58)
10 180 88
11 180 67
12 177 76

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 14


K-Means Algorithm – Example Cont..
Sr. Height Weight
1 185 72
2 170 56
K
3 168 60
(185,72)
1
4 179 68
5 182 72
6 188 77
7 180 71
8 180 70 K
9 183 84 2
(169,58)
10 180 88
11 180 67 Now, data (179,68) nearer to K1, so it belongs to K1.
12 177 76 K1 = {1,4}
K2 = {2,3}
Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 15
K-Means Algorithm – Example Cont..
Sr. Height Weight
1 185 72
2 170 56
K
3 168 60
(182,70)
1
4 179 68 Now, New Centroid Calculation
5 182 72 For K1 = {1,4}
So, K2 = {(185,72),(179,68)}
6 188 77
= 185+179/2 & 72+68/2
7 180 71 We get new centroid C = (182,70)
8 180 70 K
9 183 84 (169,58)
2
10 180 88
11 180 67
12 177 76

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 16


K-Means Algorithm – Example Cont..
Sr. Height Weight
1 185 72
2 170 56
K
3 168 60
(182,70)
1
4 179 68
5 182 72
6 188 77
7 180 71
8 180 70 K
9 183 84 2
(169,58)
10 180 88 = 0

11 180 67 Now, data (182,72) nearer to K1, so it belongs to K1.


12 177 76 K1 = {1,4,5}
K2 = {2,3}
Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 17
K-Means Algorithm – Example Cont..
Sr. Height Weight
1 185 72
2 170 56
K
3 168 60
(182,71)
1
4 179 68 Now, New Centroid Calculation
5 182 72 For K1 = {1,4,5}
So, K2 = {(185,72),(179,68),(182,72)}
6 188 77
= 185+179+182/3 & 72+68+72/3
7 180 71 We get new centroid C = (182,70.666) ~
8 180 70 K (182,71)
9 183 84 (169,58)
2
10 180 88
11 180 67
12 177 76

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 18


K-Means Algorithm – Example Cont..
Sr. Height Weight
1 185 72
2 170 56
K
3 168 60
(182,71)
1
4 179 68
5 182 72
6 188 77
7 180 71
8 180 70 K
9 183 84 2
(169,58)
10 180 88
11 180 67 Now, data (188,77) nearer to K1, so it belongs to K1.
12 177 76 K1 = {1,4,5,6}
K2 = {2,3}
Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 19
K-Means Algorithm – Example Cont..
Sr. Height Weight
1 185 72
2 170 56
K
3 168 60
(183,72)
1
4 179 68 Now, New Centroid Calculation
5 182 72 For K1 = {1,4,5,6}
So, K2 = {(185,72),(179,68),(182,72),(188,77)}
6 188 77
= 185+179+182+188/4 & 72+68+72+77/7
7 180 71 We get new centroid C = (183.50,72.25) ~ (183,72)
8 180 70 K
9 183 84 (169,58)
2
10 180 88
11 180 67
12 177 76

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 20


K-Means Algorithm – Example Cont..
Sr. Height Weight
1 185 72
2 170 56
K
3 168 60
(183,72)
1
4 179 68
5 182 72
6 188 77
7 180 71
8 180 70 K
9 183 84 2
(169,58)
10 180 88
11 180 67 Now, data (180,71) nearer to K1, so it belongs to K1.
12 177 76 K1 = {1,4,5,6,7}
K2 = {2,3}
Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 21
K-Means Algorithm – Example Cont..
Sr. Height Weight
1 185 72
2 170 56
K
3 168 60 Cluster K1 = {1,4,5,6,7,8,9,10,11,12}
1
4 179 68
5 182 72
6 188 77
7 180 71
8 180 70 K
9 183 84 2 Cluster K2 = {2,3}

10 180 88
11 180 67
12 177 76

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 22


K-Means Algorithm Cont..
Let us assume two clusters, and each individual's scores include two variables.
Step-1
 Choose the number of clusters.
Step-2
 Set the initial partition, and the initial mean vectors for each cluster.
Step-3
 For each remaining individual...
Step-4
 Get averages for comparison to the Cluster 1:
 Add individual's A value to the sum of A values of the individuals in Cluster 1, then divide by the total number of
scores that were summed.
 Add individual's B value to the sum of B values of the individuals in Cluster 1, then divide by the total number of
scores that were summed.

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 23


K-Means Algorithm Cont..
Step-5
 Get averages for comparison to the Cluster 2:
 Add individual's A value to the sum of A values of the individuals in Cluster 2, then divide by the total number of
scores that were summed.
 Add individual's B value to the sum of B values of the individuals in Cluster 2, then divide by the total number of
scores that were summed.
Step-6
 If the averages found in Step 4 are closer to the mean values of Cluster 1, then this individual belongs to
Cluster 1, and the averages found now become the new mean vectors for Cluster 1.
 If closer to Cluster 2, then it goes to Cluster 2, along with the averages as new mean vectors.
Step-7
 If there are more individual's to process, continue again with Step 4. Otherwise go to Step 8.
Step-8
 Now compare each individual’s distance to its own cluster's mean vector, and to that of the opposite
cluster.
 The distance to its cluster's mean vector should be smaller than it distance to the other vector.
 If not, relocate the individual to the opposite cluster.
Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 24
K-Means Algorithm Cont..
Step-9
 If any relocations occurred in Step 8, the algorithm must continue again with Step 3, using all individuals
and the new mean vectors.
 If no relocations occurred, stop. Clustering is complete.

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 25


K-Medoids Clustering
Algorithm (PAM)
Section - 4
What is Medoid?
Medoids are similar in concept to means or centroids, but medoids are always restricted to
be members of the data set.
Medoids are most commonly used on data when a mean or centroid cannot be defined, such
as graphs.
Note: A medoid is not equivalent to a median.

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 27


K-Medoids Clustering Algorithm (PAM)
The k-medoids algorithm is a clustering algorithm related to the k-means algorithm also
called as the medoid shift algorithm.
Both the k-means and k-medoids algorithms are partitional (breaking the dataset up into
groups).
In contrast to the k-means algorithm, k-medoids chooses datapoints as centers (medoids or
exemplars).
K-medoids is also a partitioning technique of clustering that clusters the data set of n objects
into k clusters with k known a priori.
It could be more robust to noise and outliers as compared to k-means because it minimizes a
sum of general pairwise dissimilarities instead of a sum of squared Euclidean distances.
A medoid of a finite dataset is a data point from this set, whose average dissimilarity to all
the data points is minimal i.e. it is the most centrally located point in the set.

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 28


K-Medoids Clustering Algorithm (PAM) Cont..
It was proposed in 1987 by Kaufman and Rousseeuw.
A medoid can be defined as the point in the cluster, whose dissimilarities with all the other
points in the cluster is minimum.
The dissimilarity of the medoid(Ci) and object(Pi) or items is calculated by using E = |Pi - Ci|
The cost in K-Medoids algorithm is given as
Steps for k-medoid clustering (Partitioning Around Medoids (PAM)) algorithm follows as..
1. Initialize: randomly select k of the n data points as the medoids.
2. Assignment step: Associate each data point to the closest medoid.
3. Update step:
 For each medoid m and each data point o associated to m swap m and o and compute the total cost of the
configuration (that is, the average dissimilarity of o to all the data points associated to m).
 Select the medoid o with the lowest cost of the configuration.
4. Repeat alternating steps 2 and 3 until there is no change in the assignments.

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 29


K-Medoids Clustering Algorithm - Example
Step 1:
Sr. X Y Let the randomly selected 2 medoids, so select k = 2 and let C1 -(4, 5) and C2 -(8, 5) are
the two medoids.
0 8 7
The dissimilarity of each non-medoid point with the medoids is calculated and tabulated:
1 3 7
Sr. X Y Dissimilarity From C1 Dissimilarity From C2
2 4 9
3 9 6 0 8 7 |(8-4)|+|(7-5)| = 6 |(8-8)|+|(7-5)| = 2
4 8 5 1 3 7 3 7
5 5 8 2 4 9 4 8
6 7 3 3 9 6 6 2
7 8 4 5 5 8 4 6
8 7 5 6 7 3 5 3
9 4 5 7 8 4 5 1
8 7 5 3 1

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 30


K-Medoids Clustering Algorithm – Example Cont..

Dissimilarit Dissimilarit • Each point is assigned to the cluster of that medoid


Sr. X Y
y From C1 y From C2 whose dissimilarity is less.
0 8 7 6 2 • The points 1, 2, 5 go to cluster C1 and 0, 3, 6, 7, 8
go to cluster C2.
1 3 7 3 7 • The Cost = (3 + 4 + 4) + (2 + 2 + 3 + 1 + 1) = 20
2 4 9 4 8
3 9 6 6 2
5 5 8 4 6
6 7 3 5 3
7 8 4 5 1
8 7 5 3 1

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 31


K-Medoids Clustering Algorithm – Example Cont..

Dissimilarit Dissimilarit • Step 3: randomly select one non-medoid point


Sr. X Y
y From C1 y From C2 and recalculate the cost.
0 8 7 6 3 • Let the randomly selected point be (8, 4).
• The dissimilarity of each non-medoid point with
1 3 7 3 8 the medoids – C1 (4, 5) and C2 (8, 4) is calculated
2 4 9 4 9 and tabulated.
3 9 6 6 3
4 8 5 4 1
5 5 8 4 7
6 7 3 5 2
8 7 5 3 2

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 32


K-Medoids Clustering Algorithm – Example Cont..

Dissimilarit Dissimilarit • Each point is assigned to that cluster whose


Sr. X Y
y From C1 y From C2 dissimilarity is less. So, the points 1, 2, 5 go to
0 8 7 6 3 cluster C1 and 0, 3, 4, 6, 8 go to cluster C2.
• The New cost,
1 3 7 3 8 = (3 + 4 + 4) + (3 + 3 + 1 + 2 + 2) = 22
2 4 9 4 9 • Swap Cost = New Cost – Previous Cost
3 9 6 6 3 = 22 – 20
=2
4 8 5 4 1 • So, 2>0 that is positive, now our previous medoid is
5 5 8 4 7 best.
• The total cost of Medoid (8,4) > the total cost
6 7 3 5 2
when (8,5) was the medoid earlier & it generates
8 7 5 3 2 the same clusters as earlier.
• If you get negative then you have to take new
medoid and recalculate again.

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 33


K-Medoids Clustering Algorithm – Example Cont..

Sr. X Y • As the swap cost is not less than zero, we undo the swap.
• Hence (4, 5) and (8, 5) are the final medoids.
0 8 7 • The clustering would be in the following way
1 3 7
2 4 9
3 9 6
4 8 5
5 5 8
6 7 3
7 8 4
8 7 5
9 4 5

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 34


K-Medoids Clustering Algorithm (Try Yourself!!)

Sr. X Y
0 2 6
1 3 4
2 3 8
3 4 7
4 6 2
5 6 4
6 7 3
7 7 4
8 8 5
9 7 6

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 35


Hierarchical Clustering
Techniques
Section - 5
Hierarchical Clustering
Hierarchical Clustering is a technique to group objects based on distance or similarity.
Hierarchical Clustering is called as unsupervised learning.
Because, the machine (computer) learns mere from objects with their features and then the
machine will automatically categorize those objects into groups.
This clustering technique is divided into two types:
 Agglomerative
 In this technique, initially each data point is considered as an individual cluster.
 At each iteration, the similar clusters merge with other clusters until one cluster or K clusters are formed.
 Divisive
 Divisive Hierarchical clustering is exactly the opposite of the Agglomerative Hierarchical clustering.
 In Divisive Hierarchical clustering, we consider all the data points as a single cluster and in each iteration, we
separate the data points from the cluster which are not similar.
 Each data point which is separated is considered as an individual cluster. In the end, we’ll be left with n clusters.

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 37


Agglomerative Hierarchical Clustering Technique
In this technique, initially each data point is considered as an individual cluster.
At each iteration, the similar clusters merge with other clusters until one cluster or K clusters
are formed.
The basic algorithm of Agglomerative is straight forward.
 Compute the proximity matrix
 Let each data point be a cluster
 Repeat: Merge the two closest clusters and update the proximity matrix
 Until only a single cluster remains
Key operation is the computation of the proximity of two clusters.

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 38


Agglomerative Hierarchical Clustering - Example
To understand better let’s see a pictorial representation of the Agglomerative Hierarchical
clustering Technique.
Lets say we have six data points {A,B,C,D,E,F}.
Step- 1:
 In the initial step, we calculate the proximity of individual points and consider all the six data points as
individual clusters as shown in the image below.

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 39


Agglomerative Hierarchical Clustering - Example
Step- 2:
 In step two, similar clusters are merged together and formed as a single cluster.
 Let’s consider B,C, and D,E are similar clusters that are merged in step two.
 Now, we’re left with four clusters which are A, BC, DE, F.
Step- 3:
 We again calculate the proximity of new clusters and merge the similar clusters to form new clusters A,
BC, DEF.
Step- 4:
 Calculate the proximity of the new clusters.
 The clusters DEF and BC are similar and merged together to form a new cluster.
 We’re now left with two clusters A, BCDEF.
Step- 5:
 Finally, all the clusters are merged together and form a single cluster.

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 40


Agglomerative Hierarchical Clustering - Example
The Hierarchical clustering Technique can be visualized using a Dendrogram.
A Dendrogram is a tree-like diagram that records the sequences of merges or splits.

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 41


Divisive Hierarchical Clustering Technique
Divisive Hierarchical clustering is exactly the opposite of the Agglomerative Hierarchical
clustering.
In Divisive Hierarchical clustering, we consider all the data points as a single cluster and in
each iteration, we separate the data points from the cluster which are not similar.
Each data point which is separated is considered as an individual cluster.
In the end, we’ll be left with n clusters.
As we are dividing the single clusters into n clusters, it is named as Divisive Hierarchical
clustering.
It is not much used in the real world.

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 42


Divisive Hierarchical Clustering Technique Cont..
Calculating the similarity between two clusters is important to merge or divide the clusters.
There are certain approaches which are used to calculate the similarity between two clusters:
 MIN
 MAX
 Group Average
 Distance Between Centroids
 Ward’s Method

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 43


MIN
Min is known as single-linkage algorithm can be defined as the similarity of two clusters
C1 and C2 is equal to the minimum of the similarity between points Pi and Pj such that Pi
belongs to C1 and Pj belongs to C2.
Mathematically this can be written as,
 Sim(C1,C2) = Min Sim(Pi,Pj) such that Pi ∈ C1 & Pj ∈ C2
In simple words, pick the two closest points such that one point lies in cluster one and the
other point lies in cluster 2 and takes their similarity and declares it as the similarity between
two clusters.

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 44


MAX
Max is known as the complete linkage algorithm, this is exactly opposite to the MIN
approach.
The similarity of two clusters C1 and C2 is equal to the maximum of the similarity between
points Pi and Pj such that Pi belongs to C1 and Pj belongs to C2.
Mathematically this can be written as,
 Sim(C1,C2) = Max Sim(Pi,Pj) such that Pi ∈ C1 & Pj ∈ C2
In simple words, pick the two farthest points such that one point lies in cluster one and the
other point lies in cluster 2 and takes their similarity and declares it as the similarity between
two clusters.

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 45


Group Average
Take all the pairs of points and compute their similarities and calculate the average of the
similarities.
Mathematically this can be written as,
 Sim(C1,C2) = ∑ Sim(Pi, Pj)/|C1|*|C2|, where, Pi ∈ C1 & Pj ∈ C2

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 46


Distance between centroids
Compute the centroids of two clusters C1 & C2 and take the similarity between the two
centroids as the similarity between two clusters.
This is a less popular technique in the real world.

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 47


Ward’s Method
This approach of calculating the similarity between two clusters is exactly the same as Group
Average except that Ward’s method calculates the sum of the square of the distances Pi and
Pj.
Mathematically this can be written as,
 Sim(C1,C2) = ∑ (dist(Pi, Pj))²/|C1|*|C2|

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 48


Hierarchical Clustering
Example
Agglomerative Hierarchical Clustering - Example
X Y Calculate Euclidean distance, create the distance matrix.
P1 0.40 0.53 Distance [(x,y),(a,b)] =
P2 0.22 0.38
P3 0.35 0.32
P4 0.26 0.19
P5 0.08 0.41
P6 0.45 0.30
P1 P2 P3 P4 P5 P6
P1 0
0.2 0
P2
3
P3 0
P4 0
P5 0
P6 0
Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 50
Agglomerative Hierarchical Clustering - Example
X Y
P1 0.40 0.53
P2 0.22 0.38
P3 0.35 0.32
P4 0.26 0.19
P5 0.08 0.41
P6 0.45 0.30 P1 P2 P3 P4 P5 P6
P1 0
0.2 0
P2
3
0.2 0
P3
2
P4 0
P5 0
P6 0
Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 51
Agglomerative Hierarchical Clustering - Example
X Y P1 P2 P3 P4 P5 P6
P1 0.40 0.53 P1 0
P2 0.22 0.38 P2 0.23 0
P3 0.35 0.32 P3 0.22 0.15 0
P4 0.26 0.19 P4 0.37 0.20 0.15 0
P5 0.08 0.41 P5 0.34 0.14 0.28 0.29 0
P6 0.45 0.30 P6 0.23 0.25 0.11 0.22 0.39 0

3 6

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 52


Agglomerative Hierarchical Clustering - Example
To Update the distance matrix MIN[dist(P3,P6),P1]
MIN (dist(P3,P1),(P6,P1))
Min[(0.22,0.23)]
0.22 P1 P2 P3 P4 P5 P6
P1 0
P2 0.23 0
P3 0.22 0.15 0
To Update the distance matrix MIN[dist(P3,P6),P2]
P4 0.37 0.20 0.15 0
MIN (dist(P3,P2),(P6,P2))
P5 0.34 0.14 0.28 0.29 0
Min[(0.15,0.25)]
P6 0.23 0.25 0.11 0.22 0.39 0
0.15

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 53


Agglomerative Hierarchical Clustering - Example
To Update the distance matrix MIN[dist(P3,P6),P4]
MIN (dist(P3,P4),(P6,P4))
Min[(0.15,0.22)]
0.15 P1 P2 P3 P4 P5 P6
P1 0
P2 0.23 0
P3 0.22 0.15 0
To Update the distance matrix MIN[dist(P3,P6),P5]
P4 0.37 0.20 0.15 0
MIN (dist(P3,P5),(P6,P5))
P5 0.34 0.14 0.28 0.29 0
Min[(0.28,0.39)]
P6 0.23 0.25 0.11 0.22 0.39 0
0.28

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 54


Agglomerative Hierarchical Clustering - Example
The Updated distance matrix for cluster P3, P6
P1 P2 P3,P6 P4 P5 P1 P2 P3 P4 P5 P6
P1 0 P1 0
P2 0.23 0 P2 0.23 0
P3,P6 0.22 0.15 0 P3 0.22 0.15 0
P4 0.37 0.20 0.15 0 P4 0.37 0.20 0.15 0
P5 0.34 0.14 0.28 0.29 0 P5 0.34 0.14 0.28 0.29 0
P6 0.23 0.25 0.11 0.22 0.39 0

2 5

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 55


Agglomerative Hierarchical Clustering - Example
To Update the distance matrix MIN[dist(P2,P5),P1]
MIN (dist(P2,P1),(P5,P1))
Min[(0.23,0.34)]
0.23 P1 P2 P3,P6 P4 P5
P1 0
P2 0.23 0
To Update the distance matrix MIN[dist(P2,P5),(P3,P6)]P3,P6 0.22 0.15 0
P4 0.37 0.20 0.15 0
MIN [(dist(P2,(P3,P6)),(P5,(P3,P6))]
P5 0.34 0.14 0.28 0.29 0
Min[(0.15,0.28)]
0.15

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 56


Agglomerative Hierarchical Clustering - Example
To Update the distance matrix MIN[dist(P2,P5),P4] P1 P2 P3,P6 P4 P5
MIN (dist(P2,P4),(P5,P4)) P1 0
Min[(0.20,0.29)] P2 0.23 0
0.20 P3,P6 0.22 0.15 0
P4 0.37 0.20 0.15 0
P5 0.34 0.14 0.28 0.29 0

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 57


Agglomerative Hierarchical Clustering - Example

P1 P2 P3,P6 P4 P5
P1 P2,P5 P3,P6 P4
P1 0
P1 0
P2 0.23 0
P2,P5 0.23 0
P3,P6 0.22 0.15 0
P3,P6 0.22 0.15 0
P4 0.37 0.20 0.15 0
P4 0.37 0.20 0.15 0
P5 0.34 0.14 0.28 0.29 0

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 58


Agglomerative Hierarchical Clustering - Example
P1 P2,P5 P3,P6 P4 P1 P2,P5 P3,P6 P4
P1 0 P1 0
P2,P5 0.23 0 P2,P5 0.23 0
P3,P6 0.22 0.15 0 P3,P6 0.22 0.15 0
P4 0.37 0.20 0.15 0 P4 0.37 0.20 0.15 0

2 5
3 6

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 59


Agglomerative Hierarchical Clustering - Example
P1 P2,P5 P3,P6 P4 To Update the distance matrix MIN[dist(P2,P5),(P3,P6)),P1]
MIN (dist(P2,P5),P1),((P3,P6),P1)]
P1 0
Min[(0.23,0.22)]
P2,P5 0.23 0 0.22
P3,P6 0.22 0.15 0
P4 0.37 0.20 0.15 0 To Update the distance matrix MIN[dist(P2,P5),(P3,P6)),P4]
MIN (dist(P2,P5),P4),((P3,P6),P4)]
Min[(0.20,0.15)]
0.15

P1 P2,P5,P3,P6 P4
P1 0
P2,P5,P3,P 0
0.22
6
2 5 4 P4 0.37 0.15 0
3 6

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 60


Agglomerative Hierarchical Clustering - Example
P1 P2,P5,P3,P6 P4 To Update the distance matrix MIN[dist(P2,P5,P3,P6),P4]
MIN (dist(P2,P5,P3,P6),P1),(P4,P1)]
P1 0
Min[(0.22,0.37)]
P2,P5,P3,P 0 0.22
0.22
6
P4 0.37 0.15 0
P1 P2,P5,P3,P6,P4
P1 0
P2,P5,P3,P6,P 0
0.22
4

2 5 4
3 6

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 61


Agglomerative Hierarchical Clustering - Example

X Y
P1 0.40 0.53
P2 0.22 0.38
P3 0.35 0.32
P4 0.26 0.19
P5 0.08 0.41
P6 0.45 0.30
3 6 2 5 4 1

Prof. Naimish R Vadodariya #3160714 (DM)  Unit 5 – Cluster Analysis 62


Data Mining (DM)
GTU #3160714

Unit – 5
Cluster
Thank Any
Questions ?
Analysis
You

Prof. Naimish R. Vadodariya


Computer Engineering
Darshan Institute of Engineering & Technology, Rajkot
naimish.Vadodariya@darshan.ac.in
8866215253

You might also like