(Minimum Spanning Tree Methods (MST))
Line 22: Line 22:
  
 
Use for visualization
 
Use for visualization
 +
 
[[Image:Lecture24Fig2_Old Kiwi.jpg]] Fig.2
 
[[Image:Lecture24Fig2_Old Kiwi.jpg]] Fig.2
  

Revision as of 11:52, 10 April 2008

Minimum Spanning Tree Methods (MST)

One can find MST using Prim's Algorithm (fast test '57) or Krusal's Algorithm

Zahn's clustering algorithm (1971)

  • Find MST
  • Identify "inconsistent" edges in MST and remove them.

e.g. remove the longest edge

1 edge removed=> 2 clusters

2 edges removed=> 3 clusters

What if you have?

Lecture24Fig1 Old Kiwi.jpg Fig.1

Instead, use "local" inconsistency remove edges significantly larger than their neighborhood edges.

Use for visualization

Lecture24Fig2 Old Kiwi.jpg Fig.2

Visualizations of hierarchical clustering

Consider the following set of five 2D data points, which we seek to cluster hierarchically.

Lecture23ClustersRaw Old Kiwi.jpg

We may represent the hierarchical clustering in various ways. One is by a Venn diagram, in which we circle the data points which belong to a cluster, then subsequently circle any clusters that belong to a larger cluster in the hierarchy.

Lecture23VennClusters Old Kiwi.jpg

Another representation is a dendogram. A dendogram represents the clustering as a tree, with clusters that are more closely grouped indicated as siblings "earlier" in the tree. The dendogram also includes a "similarity scale," which indicates the distance between the data points (clusters) which were grouped to form a larger cluster. For the example dataset above (with distances calculated as Euclidian distance), we have the following dendogram:

Lecture23DendogramCluster Old Kiwi.jpg

A third representation of hierarchical clustering is by using brackets. We bracket data points/clusters which are grouped into a cluster in a hierarchical fashion as follows: $ \{\{\{X_1,X_2\},\{X_3,X_4\}\},X_5\} $

Alumni Liaison

Followed her dream after having raised her family.

Ruth Enoch, PhD Mathematics