Font Size: a A A

Optimizing performance on massively parallel computers using a remote memory access programming model

Posted on:2011-01-05Degree:Ph.DType:Dissertation
University:Washington State UniversityCandidate:Krishnan, ManojkumarFull Text:PDF
GTID:1448390002469397Subject:Computer Science
Abstract/Summary:
Parallel programming models are of paramount importance because they affect both the performance delivered by massively parallel systems and the productivity of the programmer seeking that performance. Advancements in networks, multicore chips, and related technology continue to improve the efficiency of modern supercomputers. However, the average application efficiency is a small fraction of the peak system efficiency.;This research proposes techniques for optimizing application performance on supercomputers using remote memory access (RMA) parallel programming model. The growing gaps between CPU-network and CPU-memory timescales are fundamental problems that require attention in the design of communication models as well as scalable parallel algorithms. This research validates the RMA model because of its simplicity, its good hardware support on modern networks, and its posession of certain characteristics important for reducing the performance gap between system peak and application performance.;The effectiveness of these optimizations is evaluated in the context of parallel linear algebra kernels. The current approach differs from the other linear algebra algorithms by the explicit use of shared memory and remote memory access communication rather than message passing. It is suitable for clusters and scalable shared memory systems. The experimental results on large scale systems (Linux-Infiniband cluster, Cray XT) demonstrate consistent performance advantages over the ScaLAPACK suite, the leading implementation of parallel linear algebra algorithms used today.
Keywords/Search Tags:Performance, Parallel, Remote memory access, Programming, Linear algebra
Related items