Now showing items 1-6 of 6
Function Shipping in a Scalable Parallel Programming Model
Increasingly, a large number of scientific and technical applications exhibit dynamically generated parallelism or irregular data access patterns. These applications pose significant challenges to achieving scalable ...
Compiling for software distributed-shared memory systems
In this thesis, we explore the use of software distributed shared memory (SDSM) as a target communication layer for parallelizing compilers. We explore how to effectively exploit compiler-derived knowledge of sharing and ...
RCC: A compiler for the R language for statistical computing
R is a programming language for statistics that enables users to express computation at a high level of abstraction. Until now, its only implementation has been the R interpreter. Though interpretation is convenient for ...
Semi-automatic synthesis of parameterized performance models for scientific programs
Building parameterized performance models of applications in an automatic way is difficult because of the large number of variables that affect performance, including architecture-dependent factors, algorithmic choices and ...
Efficient call path profiles on unmodified, optimized code
Identifying performance bottlenecks and their associated calling contexts is critical for tuning high-performance applications. This thesis presents a new approach to measuring resource utilization and its calling context. ...
Exploring the potential for accelerating sparse matrix-vector product on a Processing-in-Memory architecture
As the importance of memory access delays on performance has mushroomed over the past few decades, researchers have begun exploring Processing-in-Memory (PIM) technology, which offers higher memory bandwidth, lower memory ...