Content area

Abstract

Using NVIDIA’s Compute Unified Device Architecture (CUDA) C/C++ as well as python libraries accelerated using a Graphics Processing Unit (GPU), GPU-accelerated computing was tested against traditional Central Processing Unit (CPU) computing to determine whether it’s feasible for GPU computing to replace traditional CPU computing. Three experiments were designed to compare the computational speed increase of the GPU versus the added overhead of the required memory transfers to and from the GPU. These experiments include purely computational tasks, machine learning model training, large-scale data processing, and cloud service creation. The goal was to use these experiments to optimize GPU kernels to fit into the GPU architecture to minimize execution time and to get a fair comparison against its CPU counterpart. In cloud services specifically, autoscaling is a major feature to handle varying workloads without wasting resources. The novel contribution for this thesis comes in the form of an intelligent autoscaling feature that schedules multiple kernels on a single GPU to maximize the resources available before autoscaling to more GPU resources.

Details

1010268
Title
Dynamic GPU Kernel Scaling in Service-Oriented Architecture
Number of pages
38
Publication year
2025
Degree date
2025
School code
0010
Source
MAI 86/11(E), Masters Abstracts International
ISBN
9798314880050
Advisor
Committee member
De Luca, Gennaro; Zhao, Ming
University/institution
Arizona State University
Department
Computer Science
University location
United States -- Arizona
Degree
M.S.
Source type
Dissertation or Thesis
Language
English
Document type
Dissertation/Thesis
Dissertation/thesis number
32001804
ProQuest document ID
3202720340
Document URL
https://www.proquest.com/dissertations-theses/dynamic-gpu-kernel-scaling-service-oriented/docview/3202720340/se-2?accountid=208611
Copyright
Database copyright ProQuest LLC; ProQuest does not claim copyright in the individual underlying works.
Database
ProQuest One Academic