Kernel Mechanisms for Efficient GPU Accelerated Deep Neural Network Inference on Embedded Devices
| dc.contributor.advisor | Stiber, Michael | |
| dc.contributor.author | Nigam, Hemant | |
| dc.date.accessioned | 2018-04-24T22:16:45Z | |
| dc.date.issued | 2018-04-24 | |
| dc.date.submitted | 2018 | |
| dc.description | Thesis (Master's)--University of Washington, 2018 | |
| dc.description.abstract | Embedded platforms with integrated graphics processing units (GPUs) are popular choices, for use-cases, like Autonomous machines, to run the Deep Neural Networks (DNNs) inference workload. However, due to a rapid increase in data volume, DNN inference is becoming even more computationally intensive and memory sensitive, which necessitates a mechanism for improving DNN inference efficiency on existing embedded systems. This Master’s thesis investigates the memory sensitivity of DNN inference – specifically, the impact of off-chip memory (DRAM) contention on DNN inference performance. It demonstrates a prototype GPU aware memory isolation mechanism: a locking mechanism in the GPU driver to reduce DRAM contention caused by multicore CPUs, thus improving DNN inference efficiency. Experiments performed on a Jetson TX2 board running the Linux4Tegra OS shows the benefits of our proposed mechanism, with up to 13.5% speedup of a micro-benchmark and up to 41% and 86% speedup of two object detection benchmarks. | |
| dc.embargo.lift | 2023-03-29T22:16:45Z | |
| dc.embargo.terms | Restrict to UW for 5 years -- then make Open Access | |
| dc.format.mimetype | application/pdf | |
| dc.identifier.other | Nigam_washington_0250O_18302.pdf | |
| dc.identifier.uri | http://hdl.handle.net/1773/41727 | |
| dc.language.iso | en_US | |
| dc.rights | none | |
| dc.subject | Deep Neural Network | |
| dc.subject | DRAM | |
| dc.subject | Edge Computing | |
| dc.subject | GPU Acceleration | |
| dc.subject | Inference | |
| dc.subject | Computer science | |
| dc.subject.other | To Be Assigned | |
| dc.title | Kernel Mechanisms for Efficient GPU Accelerated Deep Neural Network Inference on Embedded Devices | |
| dc.type | Thesis |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- Nigam_washington_0250O_18302.pdf
- Size:
- 5.31 MB
- Format:
- Adobe Portable Document Format
