TY - JOUR
T1 - Predicting parallel application performance via machine learning approaches
AU - Singh, Karan
AU - Ipek, Engin
AU - McKee, Sally A.
AU - De Supinski, Bronis R.
AU - Schulz, Martin
AU - Caruana, Rich
PY - 2007/12/10
Y1 - 2007/12/10
N2 - Consistently growing architectural complexity and machine scales make the creation of accurate performance models for large-scale applications increasingly challenging. Traditional analytic models are difficult and time consuming to construct, and are often unable to capture full system and application complexity. To address these challenges, we automatically build models based on execution samples. We use multilayer neural networks, because they can represent arbitrary functions and handle noisy inputs robustly. In this paper we focus on two well-known parallel applications whose variations in execution times are not well understood: SMG 2000, a semicoarsening multigrid solver, and HPL, an open-source implementation of LINPACK. We sparsely sample performance data on two radically different platforms across large, multidimensional parameter spaces and show that our models based on these data can predict performance within 2% to 7% of actual application runtimes.
AB - Consistently growing architectural complexity and machine scales make the creation of accurate performance models for large-scale applications increasingly challenging. Traditional analytic models are difficult and time consuming to construct, and are often unable to capture full system and application complexity. To address these challenges, we automatically build models based on execution samples. We use multilayer neural networks, because they can represent arbitrary functions and handle noisy inputs robustly. In this paper we focus on two well-known parallel applications whose variations in execution times are not well understood: SMG 2000, a semicoarsening multigrid solver, and HPL, an open-source implementation of LINPACK. We sparsely sample performance data on two radically different platforms across large, multidimensional parameter spaces and show that our models based on these data can predict performance within 2% to 7% of actual application runtimes.
KW - Artificial neural networks
KW - High-performance computing
KW - Performance modeling
UR - http://www.scopus.com/inward/record.url?scp=35948986416&partnerID=8YFLogxK
U2 - 10.1002/cpe.1171
DO - 10.1002/cpe.1171
M3 - Article
AN - SCOPUS:35948986416
SN - 1532-0626
VL - 19
SP - 2219
EP - 2235
JO - Concurrency and Computation: Practice and Experience
JF - Concurrency and Computation: Practice and Experience
IS - 17
ER -