Abstract

The management of separate memory spaces of CPUs and GPUs brings an additional burden to the development of software for GPUs. To help with this, CUDA unified memory provides a single address space that can be accessed from both CPU and GPU. The automatic data transfer mechanism is based on page faults generated by the memory accesses. This mechanism has a performance cost, that can be with explicit memory prefetch requests. Various hints on the inteded usage of the memory regions can also be given to further improve the performance. The overall effect of unified memory compared to an explicit memory management can depend heavily on the application. In this paper we evaluate the performance impact of CUDA unified memory using the heterogeneous pixel reconstruction code from the CMS experiment as a realistic use case of a GPU-targeting HEP reconstruction software. We also compare the programming model using CUDA unified memory to the explicit management of separate CPU and GPU memory spaces.

Details

Title
Performance of CUDA Unified Memory in CMS Heterogeneous Pixel Reconstruction
Author
Kortelainen, Matti J; Kwok, Martin
Section
Offline Computing
Publication year
2021
Publication date
2021
Publisher
EDP Sciences
ISSN
21016275
e-ISSN
2100014X
Source type
Conference Paper
Language of publication
English
ProQuest document ID
2582847754
Copyright
© 2021. This work is licensed under https://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and conditions, you may use this content in accordance with the terms of the License.