Abstract
The K-means clustering algorithm has had successful application in sufficient dimension reduction. Unfortunately, the algorithm does have reproducibility and nestness, which will be discussed in this paper. These are clear deficits for the K-means clustering algorithm; however, the hierarchical clustering algorithm has both reproducibility and nestness, but intensive comparison between K-means and hierarchical clustering algorithm has not yet been done in a sufficient dimension reduction context. In this paper, we rigorously study the two clustering algorithms for two popular sufficient dimension reduction methodology of inverse mean and clustering mean methods throughout intensive numerical studies. Simulation studies and two real data examples confirm that the use of hierarchical clustering algorithm has a potential advantage over the K-means algorithm.
Original language | English |
---|---|
Pages (from-to) | 431-443 |
Number of pages | 13 |
Journal | Communications for Statistical Applications and Methods |
Volume | 27 |
Issue number | 4 |
DOIs | |
State | Published - 1 Jul 2020 |
Bibliographical note
Publisher Copyright:© 2020 The Korean Statistical Society, and Korean International Statistical Society.
Keywords
- Central subspace
- Hierarchical clustering
- Informative predictor subspace
- K-means clustering
- Multivariate slicing
- Sufficient dimension reduction