Title :
Cluster validity analysis using subsampling
Author :
Abul, Osman ; Lo, Anthony ; Alhajj, Reda ; Polat, Faruk ; Barker, Ken
Author_Institution :
Dept. Comput. Sci., Calgary Univ., Alta., Canada
Abstract :
Cluster validity investigates whether generated clusters are true clusters or due to chance. This is usually done based on subsampling stability analysis. Related to this problem is estimating true number of clusters in a given dataset. There are a number of methods described in the literature to handle both purposes. In this paper, we propose three methods for estimating confidence in the validity of clustering result. The first method validates clustering result by employing supervised classifiers. The dataset is divided into training and test sets and the accuracy of the classifier is evaluated on the test set. This method computes confidence in the generalization capability of clustering. The second method is based on the fact that if a clustering is valid then each of its subsets should be valid as well. The third method is similar to second method; it takes the dual approach, i.e., each cluster is expected to be stable and compact. Confidence is estimated by repeating the process a number of times on subsamples. Experimental results illustrate effectiveness of the proposed methods.
Keywords :
data analysis; generalisation (artificial intelligence); learning (artificial intelligence); pattern clustering; sampling methods; statistical analysis; cluster validity analysis; confidence estimation; dataset; generalization; subsampling stability analysis; supervised classifiers; test sets; training set; Clustering algorithms; Computer science; Humans; Organizing; Pattern analysis; Sampling methods; Stability analysis; Testing; Visualization;
Conference_Titel :
Systems, Man and Cybernetics, 2003. IEEE International Conference on
Print_ISBN :
0-7803-7952-7
DOI :
10.1109/ICSMC.2003.1244614