MARC보기
LDR00000nam u2200205 4500
001000000435859
00520200228110144
008200131s2019 ||||||||||||||||| ||eng d
020 ▼a 9781392502648
035 ▼a (MiAaPQ)AAI27712087
035 ▼a (MiAaPQ)OhioLINKosu1563472235437977
040 ▼a MiAaPQ ▼c MiAaPQ ▼d 247004
0820 ▼a 004
1001 ▼a Zhu, Gangyi.
24510 ▼a Predicting Performance of Parallel Analytics and Irregular Computations.
260 ▼a [S.l.]: ▼b The Ohio State University., ▼c 2019.
260 1 ▼a Ann Arbor: ▼b ProQuest Dissertations & Theses, ▼c 2019.
300 ▼a 182 p.
500 ▼a Source: Dissertations Abstracts International, Volume: 81-06, Section: B.
500 ▼a Advisor: Agrawal, Gagan.
5021 ▼a Thesis (Ph.D.)--The Ohio State University, 2019.
506 ▼a This item must not be sold to any third party vendors.
520 ▼a Scientific simulation and computation are becoming more complicated in recent years due to multiple reasons including new data processing paradigms, irregular computation pattern, evolving hardware development, and other factors. Since there is an increasingly large gap between the I/O performance and compute power so that the cost of writing and reading the vast amount of simulated data from disk is expensive, more attentions have been paid to in-situ analytics. Irregular computation, usually arising in domains that use sparse matrices, graphs, or other irregular data structures, introduces a variety of uncertainty to the performance of scientific computation. Moreover, the newly developed hardware platforms such as MIC and GPUs also impose challenges for understanding and improving performance of scientific applications. Hence, there is clearly a need for analyzing the impacts of these factors to scientific computations.We start from predicting performance of the disk-based and in-situ parallel data analytics implemented in MapReduce-like frameworks. We take two distinct approaches towards performance prediction. We first expand SKOPE (a SKeleton framewOrk for Performance Exploration) with performance models for disk data read, cache performance, and page fault penalty. Second, an analytical performance model is also developed.Next, we take irregular computations into consideration for performance prediction. Cache performance of irregular computations is highly input-dependent. Based on the sparse matrix view of irregular computation as well as the cache locality analysis, we propose a novel sampling approach named Adaptive Stratified Row sampling -- this method is capable of generating a representative sample that delivers cache performance similar to the original input. On top of our sampling method, we incorporate reuse distance analysis to accommodate different cache configurations with high efficiency. Besides, we modify SKOPE, a code skeleton framework, to predict the execution time for irregular applications with the predicted cache performance.We extend the work of modeling irregular computations to the SIMD scenario. Our first insight is that developing a universal sampling approach for all sparse matrices is unpractical. According to the non-zero distribution of the sparse matrix, we propose two novel sampling strategies: Stride Average sampling and Random Tile sampling, which are suitable for uniform and skewed sparse matrices respectively. To help categorize a sparse matrix as uniform or skewed, we introduce clustering coefficient as an important feature which can be propagated into the decision tree model. We also adapt Random Node Neighbor sampling approach for efficient estimation of clustering coefficient.Finally, we target another topic of irregular computation on GPUs: sparse matrix format selection for Sparse Matrix-Vector Multiplication (SpMV). Based on the storage properties and processing granularity of different formats, we develop three novel sampling schemes: Row Crop sampling, Random Warp sampling, and Diagonal Align sampling. Then we obtain the base performance for each format by executing SpMV over the generated samples. The best format is predicted by scaling the base performance based on the difference of parallelism between the original matrix and the sampled one.
590 ▼a School code: 0168.
650 4 ▼a Computer engineering.
650 4 ▼a Computer science.
690 ▼a 0984
690 ▼a 0464
71020 ▼a The Ohio State University. ▼b Computer Science and Engineering.
7730 ▼t Dissertations Abstracts International ▼g 81-06B.
773 ▼t Dissertation Abstract International
790 ▼a 0168
791 ▼a Ph.D.
792 ▼a 2019
793 ▼a English
85640 ▼u http://www.riss.kr/pdu/ddodLink.do?id=T15494734 ▼n KERIS ▼z 이 자료의 원문은 한국교육학술정보원에서 제공합니다.
980 ▼a 202002 ▼f 2020
990 ▼a ***1008102
991 ▼a E-BOOK