Super-sparse principal component analyses for high-throughput genomic data

Donghwan Lee, Woojoo Lee, Youngjo Lee, Yudi Pawitan

Research output: Contribution to journalArticlepeer-review

36 Scopus citations

Abstract

Background: Principal component analysis (PCA) has gained popularity as a method for the analysis of high-dimensional genomic data. However, it is often difficult to interpret the results because the principal components are linear combinations of all variables, and the coefficients (loadings) are typically nonzero. These nonzero values also reflect poor estimation of the true vector loadings; for example, for gene expression data, biologically we expect only a portion of the genes to be expressed in any tissue, and an even smaller fraction to be involved in a particular process. Sparse PCA methods have recently been introduced for reducing the number of nonzero coefficients, but these existing methods are not satisfactory for high-dimensional data applications because they still give too many nonzero coefficients.Results: Here we propose a new PCA method that uses two innovations to produce an extremely sparse loading vector: (i) a random-effect model on the loadings that leads to an unbounded penalty at the origin and (ii) shrinkage of the singular values obtained from the singular value decomposition of the data matrix. We develop a stable computing algorithm by modifying nonlinear iterative partial least square (NIPALS) algorithm, and illustrate the method with an analysis of the NCI cancer dataset that contains 21,225 genes.Conclusions: The new method has better performance than several existing methods, particularly in the estimation of the loading vectors.

Original languageEnglish
Article number296
JournalBMC Bioinformatics
Volume11
DOIs
StatePublished - 2 Jun 2010

Fingerprint

Dive into the research topics of 'Super-sparse principal component analyses for high-throughput genomic data'. Together they form a unique fingerprint.

Cite this