Show simple item record

dc.contributor.advisorKeyes, David E.
dc.contributor.authorCharara, Ali
dc.date.accessioned2018-05-24T08:04:04Z
dc.date.available2019-05-24T00:00:00Z
dc.date.issued2018-05-24
dc.identifier.citationCharara, A. (2018). Exploiting Data Sparsity In Covariance Matrix Computations on Heterogeneous Systems. KAUST Research Repository. https://doi.org/10.25781/KAUST-5M8Z4
dc.identifier.doi10.25781/KAUST-5M8Z4
dc.identifier.urihttp://hdl.handle.net/10754/627948
dc.description.abstractCovariance matrices are ubiquitous in computational sciences, typically describing the correlation of elements of large multivariate spatial data sets. For example, covari- ance matrices are employed in climate/weather modeling for the maximum likelihood estimation to improve prediction, as well as in computational ground-based astronomy to enhance the observed image quality by filtering out noise produced by the adap- tive optics instruments and atmospheric turbulence. The structure of these covariance matrices is dense, symmetric, positive-definite, and often data-sparse, therefore, hier- archically of low-rank. This thesis investigates the performance limit of dense matrix computations (e.g., Cholesky factorization) on covariance matrix problems as the number of unknowns grows, and in the context of the aforementioned applications. We employ recursive formulations of some of the basic linear algebra subroutines (BLAS) to accelerate the covariance matrix computation further, while reducing data traffic across the memory subsystems layers. However, dealing with large data sets (i.e., covariance matrices of billions in size) can rapidly become prohibitive in memory footprint and algorithmic complexity. Most importantly, this thesis investigates the tile low-rank data format (TLR), a new compressed data structure and layout, which is valuable in exploiting data sparsity by approximating the operator. The TLR com- pressed data structure allows approximating the original problem up to user-defined numerical accuracy. This comes at the expense of dealing with tasks with much lower arithmetic intensities than traditional dense computations. In fact, this thesis con- solidates the two trends of dense and data-sparse linear algebra for HPC. Not only does the thesis leverage recursive formulations for dense Cholesky-based matrix al- gorithms, but it also implements a novel TLR-Cholesky factorization using batched linear algebra operations to increase hardware occupancy and reduce the overhead of the API. Performance reported of the dense and TLR-Cholesky shows many-fold speedups against state-of-the-art implementations on various systems equipped with GPUs. Additionally, the TLR implementation gives the user flexibility to select the desired accuracy. This trade-off between performance and accuracy is, currently, a well-established leading trend in the convergence of the third and fourth paradigm, i.e., HPC and Big Data, when moving forward with exascale software roadmap.
dc.language.isoen
dc.subjectdata sparse
dc.subjectHierarchical
dc.subjectcovariance matrix
dc.subjectGPU
dc.subjecttile low-rank
dc.subjectDense Linear Algebra
dc.titleExploiting Data Sparsity In Covariance Matrix Computations on Heterogeneous Systems
dc.typeDissertation
dc.contributor.departmentComputer, Electrical and Mathematical Science and Engineering (CEMSE) Division
dc.rights.embargodate2019-05-24
thesis.degree.grantorKing Abdullah University of Science and Technology
dc.contributor.committeememberGenton, Marc G.
dc.contributor.committeememberHadwiger, Markus
dc.contributor.committeememberLtaief, Hatem
dc.contributor.committeememberElster, Ann C.
thesis.degree.disciplineComputer Science
thesis.degree.nameDoctor of Philosophy
dc.rights.accessrightsAt the time of archiving, the student author of this dissertation opted to temporarily restrict access to it. The full text of this dissertation became available to the public after the expiration of the embargo on 2019-05-24.
refterms.dateFOA2019-05-24T00:00:00Z


Files in this item

Thumbnail
Name:
Charara-Thesis-180522.pdf
Size:
9.966Mb
Format:
PDF

This item appears in the following Collection(s)

Show simple item record