Applying Source Level Auto-Vectorization to Aparapi Java
dc.contributor.author | Albert, Frank Curtis | en |
dc.contributor.committeechair | Ravindran, Binoy | en |
dc.contributor.committeemember | Broadwater, Robert P. | en |
dc.contributor.committeemember | Wang, Chao | en |
dc.contributor.department | Electrical and Computer Engineering | en |
dc.date.accessioned | 2014-06-20T08:00:15Z | en |
dc.date.available | 2014-06-20T08:00:15Z | en |
dc.date.issued | 2014-06-19 | en |
dc.description.abstract | Ever since chip manufacturers hit the power wall preventing them from increasing processor clock speed, there has been an increased push towards parallelism for performance improvements. This parallelism comes in the form of both data parallel single instruction multiple data (SIMD) instructions, as well as parallel compute cores in both central processing units (CPUs) and graphics processing units (GPUs). While these hardware enhancements offer potential performance enhancements, programs must be re-written to take advantage of them in order to see any performance improvement Some lower level languages that compile directly to machine code already take advantage of the data parallel SIMD instructions, but often higher level interpreted languages do not. Java, one of the most popular programming languages in the world, still does not include support for these SIMD instructions. In this thesis, we present a vector library that implements all of the major SIMD instructions in functions that are accessible to Java through JNI function calls. This brings the benefits of general purpose SIMD functionality to Java. This thesis also works with the data parallel Aparapi Java extension to bring these SIMD performance improvements to programmers who use the extension without any additional effort on their part. Aparapi already provides programmers with an API that allows programmers to declare certain sections of their code parallel. These parallel sections are then run on OpenCL capable hardware with a fallback path in the Java thread pool to ensure code reliability. This work takes advantage of the knowledge of independence of the parallel sections of code to automatically modify the Java thread pool fallback path to include the vectorization library through the use of an auto-vectorization tool created for this work. When the code is not vectorizable the auto-vectorizer tool is still able to offer performance improvements over the default fallback path through an improved looped implementation that executes the same code but with less overhead. Experiments conducted by this work illustrate that for all 10 benchmarks tested the auto-vectorization tool was able to produce an implementation that was able to beat the default Aparapi fallback path. In addition it was found that this improved fallback path even outperformed the GPU implementation for several of the benchmarks tested. | en |
dc.description.degree | Master of Science | en |
dc.format.medium | ETD | en |
dc.identifier.other | vt_gsexam:3019 | en |
dc.identifier.uri | http://hdl.handle.net/10919/49022 | en |
dc.publisher | Virginia Tech | en |
dc.rights | In Copyright | en |
dc.rights.uri | http://rightsstatements.org/vocab/InC/1.0/ | en |
dc.subject | Auto-Vectorization | en |
dc.subject | Aparapi | en |
dc.subject | Java | en |
dc.subject | GPGPU Computing | en |
dc.subject | SIMD | en |
dc.subject | Parallelism | en |
dc.subject | Threaded | en |
dc.title | Applying Source Level Auto-Vectorization to Aparapi Java | en |
dc.type | Thesis | en |
thesis.degree.discipline | Computer Engineering | en |
thesis.degree.grantor | Virginia Polytechnic Institute and State University | en |
thesis.degree.level | masters | en |
thesis.degree.name | Master of Science | en |