Techniques for Shared Resource Management in Systems with Throughput Processors

03/19/2018
by   Rachata Ausavarungnirun, et al.
0

The continued growth of the computational capability of throughput processors has made throughput processors the platform of choice for a wide variety of high performance computing applications. Graphics Processing Units (GPUs) are a prime example of throughput processors that can deliver high performance for applications ranging from typical graphics applications to general-purpose data parallel (GPGPU) applications. However, this success has been accompanied by new performance bottlenecks throughout the memory hierarchy of GPU-based systems. We identify and eliminate performance bottlenecks caused by major sources of interference throughout the memory hierarchy. We introduce changes to the memory hierarchy for systems with GPUs that allow the memory hierarchy to be aware of both CPU and GPU applications' characteristics. We introduce mechanisms to dynamically analyze different applications' characteristics and propose four major changes throughout the memory hierarchy. We propose changes to the cache management and memory scheduling mechanisms to mitigate intra-application interference in GPGPU applications. We propose changes to the memory controller design and its scheduling policy to mitigate inter-application interference in heterogeneous CPU-GPU systems. We redesign the MMU and the memory hierarchy in GPUs to be aware of ddress-translation data in order to mitigate the inter-address-space interference. We introduce a hardware-software cooperative technique that modifies the memory allocation policy to enable large page support in order to further reduce the inter-address-space interference at the shared Translation Lookaside Buffer (TLB). Our evaluations show that the GPU-aware cache and memory management techniques proposed in this dissertation are effective at mitigating the interference caused by GPUs on current and future GPU-based systems.

READ FULL TEXT

page 27

page 28

page 31

page 33

research
08/16/2017

Improving Multi-Application Concurrency Support Within the GPU Memory System

GPUs exploit a high degree of thread-level parallelism to hide long-late...
research
04/04/2017

Tackling Diversity and Heterogeneity by Vertical Memory Management

Existing memory management mechanisms used in commodity computing machin...
research
05/15/2018

Predictable Performance and Fairness Through Accurate Slowdown Estimation in Shared Main Memory Systems

This paper summarizes the ideas and key concepts in MISE (Memory Interfe...
research
06/27/2019

State-of-the-Art on Query Transaction Processing Acceleration

The vast amount of processing power and memory bandwidth provided by mod...
research
09/12/2021

Ohm-GPU: Integrating New Optical Network and Heterogeneous Memory into GPU Multi-Processors

Traditional graphics processing units (GPUs) suffer from the low memory ...
research
08/02/2023

IRQ Coloring and the Subtle Art of Mitigating Interrupt-generated Interference

Integrating workloads with differing criticality levels presents a formi...
research
07/22/2017

Optimizations of Management Algorithms for Multi-Level Memory Hierarchy

In the near future the SCM is predicted to modify the form of new progra...

Please sign up or login with your details

Forgot password? Click here to reset