is achieved by clustering the w_co_assoc matrix using the K-means, SL, CL, AL or
WR algorithms. Experimental results with both synthetic and real data show that
SWEACS lead in general to better best results than the EAC and Strehl methods.
However, no individual WEACS configuration leads systematically to the best results
in all data sets. As a complementary step to the WEACS approach we combine all the
final data partitions obtained by the use of the ALL clustering ensemble construction
method. We use the EAC approach to do this combination and we use the Ward Link
algorithm to obtain the final data partition. We reach almost in all data sets the best
results or values very close to the best results.
These results show that the association of the subsampling and the weighting
mechanisms with cluster combination techniques lead to good results.
References
1. A. Fred, “Finding consistent clusters in data partitions,”. in Multiple Classifier Systems,
Josef Kittler and Fabio Roli editors, vol. LNCS 2096, Springer, 2001, pp. 309-318.
2. Fred A., Jain A. K., “Evidence accumulation clustering based on the k-means algorithm,”
in S.S.S.P.R, T.Caelli et al., editor,., Vol. LNCS 2396, Springer-Verlag, 2002, pp. 442 –
451
3. Fred and A.K. Jain, “Combining Multiple Clusterings using Evidence Accumulation,”
IEEE Transactions on Pattern analysis and Machine Intelligence, Vol. 27, No.6, June 2005,
pp. 835-850.
4. F.Jorge Duarte, Ana L.N. Fred, André Lourenço and M. Fátima C. Rodrigues, “Weighting
Cluster Ensembles in Evidence Accumulation Clustering”, Workshop on Extraction of
Knowledge from Databases and Warehouses, EPIA 2005.
5. F.Jorge F.Duarte, Ana L.N. Fred, André Lourenço and M. Fátima C. Rodrigues, “Weighted
Evidence Accumulation Clustering”, Fourth Australasian Conference on Knowledge Dis-
covery and Data Mining 2005.
6. A. Strehl and J. Ghosh, “Cluster ensembles - a knowledge reuse framework for combining
multiple partitions,” Journal of Machine Learning Research 3, 2002.
7. B. Park and H. Kargupta, Data Mining Handbook, chapter: Distributed Data Mining. Law-
rence Erlbaum Associates, 2003.
8. M. Meila and D. Heckerman, “An Experimental Comparison of Several Clustering and
Initialization Methods”, Proc. 14th Conf. Uncertainty in Artificial Intelligence, p.p. 386-
395, 1998.
9. M. Halkidi, Y. Batistakis, M. Vazirgiannis, "Clustering algorithms and validity measures",
Tutorial paper in the proceedings of the SSDBM 2001 Conference.
10. Theodorodis, S., Koutroubas, K., Pattern Recognition. Academic Press, 1999.
11. Hubert L.J., Schultz J., “Quadratic assignment as a general data-analysis strategy,” British
Journal of Mathematical and Statistical Psychology, Vol.29, 1975, pp. 190-241.
12. Dunn, J.C., “Well separated clusters and optimal fuzzy partitions,” J. Cybern, Vol. 4, 1974,
pp. 95-104.
13. Davies, D.L., Bouldin, D.W., “A cluster separation measure,”. IEEE Transaction on Pattern
Analysis and Machine Intelligence, Vol. 1, No2, 1979.
14. S.C. Sharma, Applied Multivariate Techniques, John Willwy & Sons, 1996.
15. Calinski, R.B.& Harabasz, J, “A dendrite method for cluster analysis,” Communications in
statistics 3, 1974, pp.1-27.
16. Kaufman, L. & Roussesseeuw, P., Finding groups in data: an introduction to cluster analy-
sis, New York, Wiley, 1990.
115