Instructions' Latencies Characterization for NVIDIA GPGPUs

05/21/2019
by   Yehia Arafa, et al.
0

The last decade has seen a shift in the computer systems industry where heterogeneous computing has become prevalent. Nowadays, Graphics Processing Units (GPUs) are in a variety of systems from supercomputers to mobile phones and tablets. They are not only used for graphics operations but rather as general-purpose special hardware (GPGPUs) to boost the performance of compute-intensive applications. However, the percentage of undisclosed characteristics beyond what vendors provide is small. In this paper, we propose a very low overhead and portable analysis for exposing the hidden latency of each individual instruction executing in the pipeline and different access latencies of the various memory hierarchies at the microarchitecture level. We also show the impact of the possible optimizations a CUDA compiler have over the various latencies. We run our evaluation on seven different high-end NVIDIA GPUs from five different generations/architectures namely: Kepler, Maxwell, Pascal, Volta, and Turing. We believe that this work would help architects have an accurate characterization of the latencies of these GPUs, which would subsequently help in modeling the hardware accurately. In addition, this would also make application developers more aware of how to optimize their applications.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset