Font Size: a A A

Bayesian and Information-Theoretic Learning of High Dimensional Data

Posted on:2013-06-22Degree:Ph.DType:Dissertation
University:Duke UniversityCandidate:Chen, MinhuaFull Text:PDF
GTID:1458390008479827Subject:Statistics
Abstract/Summary:
The concept of sparseness is harnessed to learn a low dimensional representation of high dimensional data. This sparseness assumption is exploited in multiple ways. In the Bayesian Elastic Net, a small number of correlated features are identified for the response variable. In the sparse Factor Analysis for biomarker trajectories, the high dimensional gene expression data is reduced to a small number of latent factors, each with a prototypical dynamic trajectory. In the Bayesian Graphical LASSO, the inverse covariance matrix of the data distribution is assumed to be sparse, inducing a sparsely connected Gaussian graph. In the nonparametric Mixture of Factor Analyzers, the covariance matrices in the Gaussian Mixture Model are forced to be low-rank, which is closely related to the concept of block sparsity. Finally in the information-theoretic projection design, a linear projection matrix is explicitly sought for information-preserving dimensionality reduction. All the methods mentioned above prove to be effective in learning both simulated and real high dimensional datasets.
Keywords/Search Tags:High dimensional, Bayesian
Related items