35
Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Embed Size (px)

Citation preview

Page 1: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Presented by: Marlon Bright

1 August 2008

Advisor: Masoud Sadjadi, Ph.D.

REU – Florida International University

Page 2: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Outline

Grid Enablement of Weather Research and Forecasting Code (WRF)

Profiling and Prediction Tools Research Goals Project Timeline Project Status Challenges Overcome Remaining Work

2REU - Florida International University

Page 3: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Motivation – Weather Research and Forecasting Code (WRF) Goal – Improved Weather Prediction

Accurate and Timely Results Precise Location Information

WRF Status Over 160,000 lines (mostly FORTRAN and C) Single Machine/Cluster compatible Single Domain Fine Resolution -> Resource Requirements

How to Overcome this? Through Grid Enablement

Expected Benefits to WRF More available resources – Different Domains Faster results Improved Accuracy

3REU - Florida International University

Page 4: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

System Overview Web-Based Portal Grid Middleware (Plumbing)

Job-Flow ManagementMeta-Scheduling

○ Performance Prediction

Profiling and Benchmarking Development Tools and Environments

Transparent Grid Enablement (TGE)○ TRAP: Static and Dynamic adaptation of programs○ TRAP/BPEL, TRAP/J, TRAP.NET, etc.

GRID superscalar: Programming Paradigm for parallelizing a sequential application dynamically in a Computational Grid

4REU - Florida International University

Page 5: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Performance PredictionIMPORTANT part of Meta-Scheduling

Allows for: Optimal usage of grid resources through

“smarter” meta-schedulingMany users overestimate job requirementsReduced idle time for compute resourcesCould save costs and energy

Optimal resource selection for most expedient job return time

Tools: Amon /Aprof and Paraver/Dimemas

5REU - Florida International University

Page 6: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Research Goals

1. Extend Amon/Aprof research to larger number of nodes, different archtitecture, and different version of WRF (Version 2.2.1).

2. Compare/contrast Aprof predictions to Dimemas predictions in terms of accuracy and prediction computation time.

3. Analyze if/how Amon/Aprof could be used in conjunction with Dimemas/Paraver for optimized application performance prediction and, ultimately, meta-scheduling

6REU - Florida International University

Page 7: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Timeline End of June: Get MPItrace linking properly with WRF Version Compiled on GCB, then Mind

COMPLETE a) Install Amon and Aprof on MareNostrum and ensure proper functioning

AMON COMPLETE; APROF FINAL STAGES

b) Run Amon benchmarks on MareNostrum COMPLETE Early/Mid July:

Use and analyze Aprof predictions within MareNostrum (and possibly between MareNostrum, GCB, and Mind) MN COMPLETE

Use generated MPI/ OpenMP tracefiles (Paraver/Dimemas) to predict within (and possibly between) Mind, GCB, and MareNostrum IN PROGRESS

Late July/Early August: Experiment with how well Amon and Aprof relate to/could possibly be

combined with Dimemas IN PROGRESS Compose paper presenting significant findings. IN PROGRESS Analyze how findings relate to bigger picture. Make optimizations on grid-

enablement of WRF.

7REU - Florida International University

Page 8: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University
Page 9: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Amon / Aprof

Amon – monitoring program that runs on each compute node recording new processes

Aprof – regression analysis program running on head node; receives input from Amon to make execution time predictions (within cluster & between clusters)

9REU - Florida International University

Page 10: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Amon / Aprof Monitoring and Prediction

8REU - Florida International University

Page 11: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Amon / Aprof Approach to Modeling Resource Usage

9

WRF

REU - Florida International University

Page 12: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Previous Findings for Amon / AprofExperiments were performed on two clusters at FIU

—Mind (16 nodes) and GCB (8 nodes) Experiments were run to predict for different

number of nodes and cpu loads (i.e. 2,3,…,14,15 and 20%, 30%,…,90%, 100%)

Aprof predictions were within 10% error versus actual recorded runtimes within Mind and GCB and between Mind and GCB

Conclusion: first step assumption was valid. -> Move to extending research to higher number of nodes.

12REU - Florida International University

Page 13: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

How’d they do that? Developed a benchmarking script that edits and submits a job

file to MareNostrum (MN) scheduler Runs for each number of nodes (8, 16, 32, 64, 96, 128) Runs for each cpu percentage (100, 75, 50, 25) Records execution time, average cpu utilization, participating

nodes, etc. Job file:

○ Requests desired number of nodes from MN○ Starts Amon on each returned node to monitor and return processes○ Starts cpulimit on each returned node limiting the effective power given

to the WRF process○ Executes WRF as parallel job across the returned nodes

Developed modification script Combines Amon output to one file Filters processes to solely WRF processes Edits processes to Aprof friendly format

REU - Florida International University 13

Page 14: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

How’d they do that ? (cont’d) Start Aprof loading input file as data Executed Aprof Query Automation script

Starts telnet session querying Aprof for benchmarked scenarios

Compares predicted values to actual values returned in run

Outputs text file and graphing plot file of comparison statistics

REU - Florida International University 14

Page 15: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Experimental Process

REU - Florida International University 15

Page 16: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Extreme? Makeover

Amon Output Process Aprof Input Process

--- (464) ---

name: wrf.exe

cpus: 4

cpu MHz: 1/0.000 [MHz]

cache size: 1/0 [KB]

elapsed time: 957952 [msec]

utime: 956370 [msec] 957810 [msec]

stime: 570 [msec] 860 [msec]

intr: 18783

ctxt switch: 58290

fork: 95

storage R: 0 [blocks] 0 [blocks]

storage W: 0 [blocks]

network Rx: 19547308 [bytes]

network Tx: 1434925 [bytes]

--- (464) ---

name: wrf.exe

inv #cpu: 1/16

inv clock: 1/574

cache size: 1/1024 [KB]

elapsed time: 1990992 [msec]

inv clock*#cpu: 1/(36763)

Why: Version of Linux on MN does not report some characteristics (i.e. cache size). From its initial design, Amon reports in different format than Aprof reads.

REU - Florida International University 10

Page 17: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Aprof Predictionname: wrf.exe

elapsed time:

5.783787e+06

===========================================================

explanatory: value parameter std.dev

----------------- ------------- ------------- -------------

: 1.000000e+00 5.783787e+06 1.982074e+05

===========================================================

predicted: value residue rms std.dev

----------------- ------------- ------------- -------------

elapsed time: 5.783787e+06 4.246451e+06 1.982074e+05

===========================================================

REU - Florida International University 17

Page 18: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Query Automation Script Output

adj. cpu speed, processors, actual, predicted, rms, std. dev, actual difference,

3591.363, 1, 5222, 5924.82, 1592.459, 415.3491, 13.4588280352

3591.363, 2, 2881, 3246.283, 1592.459, 181.5382, 12.6790350573

3591.363, 3, 2281, 2353.438, 1592.459, 105.334, 3.17571240684

3591.363, 4, 1860, 1907.015, 1592.459, 69.19778, 2.52768817204

3591.363, 5, 1681, 1639.161, 1592.459, 49.83672, 2.48893515764

3591.363, 6, 1440, 1460.592, 1592.459, 39.5442, 1.43

3591.363, 7, 1380, 1333.043, 1592.459, 34.76459, 3.40268115942

3591.363, 8, 1200, 1237.381, 1592.459, 33.27651, 3.11508333333

3591.363, 9, 1200, 1162.977, 1592.459, 33.56231, 3.08525

3591.363, 10, 1080, 1103.454, 1592.459, 34.68943, 2.17166666667

3591.363, 11, 1200, 1054.753, 1592.459, 36.15324, 12.1039166667

3591.363, 12, 1080, 1014.169, 1592.459, 37.70271, 6.09546296296

3591.363, 13, 1200, 979.8292, 1592.459, 39.22018, 18.3475666667

3591.363, 14, 1021, 950.3947, 1592.459, 40.65455, 6.91530852106

3591.363, 15, 1020, 924.8848, 1592.459, 41.9872, 9.32501960784

REU - Florida International University 18

Page 19: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Paraver / Dimemaso Dimemas - simulation tool for the

parametric analysis of the behavior of message-passing applications on a configurable parallel platform.

o Paraver – tool that allows for performance visualization and analysis of trace files generated from actual executions and by Dimemas

Tracefiles generated by MPItrace that is linked into execution code

19REU - Florida International University

Page 20: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Dimemas Simulation Process Overview

1. Link MPItrace into application source code—dynamically generates tracefiles for each node application running on

2. Identify computation iterations in Paraver compose a smaller trace file by selecting a few iterations, preserving communications and eliminating initialization phases

3. Convert the new tracefile to Dimemas format (.trf) using CEPBA provided ‘prv2trf’ tool

4. Load tracefile into Dimemas simulator, configure target machine, and with information generate Dimemas configuration file

5. Call simulator with or without option of generating a Paraver (.prv) tracefile for viewing.

REU - Florida International University 20

Page 21: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Paraver/Dimemas – DiP Environment

REU - Florida International University 21

Page 22: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

How’d they do that? Generated Paraver tracefiles, Dimemas tracefiles, and

simulation configuration files for each number of nodes Developed Dimemas simulation script

–”simulation_automater.sh” Selects configuration file for desired number of nodes Edits configuration file for desired cpu percentage Records execution time, average cpu utilization

Finalizing development of prediction validation script. Will: Compare Dimemas predicted values to actual run values Outputs text file and graphing plot file of comparison

statistics

REU - Florida International University 22

Page 23: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Dimemas Prediction

23REU - Florida International University

Execution time: 36.354146

Speedup: 5.34

CPU Time: 194.066431

Id. Computation %time Communication

1 31.224017 91.21 3.008552

2 20.089440 78.20 5.599083

3 19.305673 76.84 5.818317

4 28.672368 83.27 5.762332

5 29.058603 85.36 4.982049

6 19.488003 77.63 5.614155

7 18.727851 78.57 5.108366

8 27.500476 84.29 5.123971

Id. Mess.sent Bytes sent Immediate recv Waiting recv Bytes re cv Coll.op. Block time Comm. time Wait link time Wait bus es time I/O time

1 7.577000e+03 1.583659e+08 3.539000e+03 4.080000e+03 1.671666 e+08 1.475000e+03 0.247092 0.383663 0.319859 0.000000 0.000000

2 8.948000e+03 2.200029e+08 8.797000e+03 1.440000e+02 2.186629 e+08 1.475000e+03 3.710867 0.383663 0.098868 0.000000 0.000000

3 8.948000e+03 2.176712e+08 6.904000e+03 2.037000e+03 2.163992 e+08 1.475000e+03 3.453668 0.383663 0.243052 0.000000 0.000000

Page 24: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

24REU - Florida International University

Page 25: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Amon / Aprof

Software installed and tailored to MareNostrum

Proficient in executing software Amon benchmarking completed Aprof query automation complete and

results generated Lessons learned on extending

Amon/Aprof to different architecture

REU - Florida International University 25

Page 26: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Dimemas / Paraver

Proficient in executing software Paraver and Dimemas tracefiles

generated for each number of nodes (8, 16, 32, 64, 96, 128)

Benchmarking script complete Simulations generated Comparison script being finalized

26REU - Florida International University

Page 27: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Quick ComparisonAmon / Aprof Dimemas / Paraver

Pros: Simpler to deploy in comparison Scalability of model is promising

with first results Feasible solution for performance

prediction purposes

Cons: Requires more base executions

for accurate performance in comparison

Pros: More features—could be more

useful to experienced user (i.e. adjustment of system characteristics)

Visualization and analysis of execution for analysis purposes

Graphical User Interface

Cons: Requires special compilation of

applications Requires non-trivial-to-install

kernel patch Large tracefiles (sometimes

gigabytes)

REU - Florida International University27

Page 28: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Aprof Results 100% CPU Utilization

REU - Florida International University 28

Page 29: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Aprof Results 100% CPU Utilization

REU - Florida International University 29

Page 30: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Significant Challenges Overcome Amon:

Adjustment of source code to proper functioning on MareNostrum (MN)

Development of benchmarking script to conform to system architecture of MareNostrum (i.e. going through its scheduler; one process per node; etc.)

Proper functioning of CPU limit for accurate cpu percentage

Job termination by MN Scheduler due to execution surpassing wall clock limit

30REU - Florida International University

Page 31: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Significant Challenges Overcome(cont’d) Aprof:

Adjustment of source code for less complex, more consistent data input

Development of prediction and comparison scripts for MareNostrum

Dimemas/ParaverMPItrace properly linked in with WRF on GCB and

MindGeneration of trace and configuration files

WRFVersion 2.2 installed and compiled on Mind

31REU - Florida International University

Page 32: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Challenges Remaining Lengthy Amon benchmarking runs due to job

times spent in queue Complexities in preparing Dimemas tracefiles for

simulation purposes Extracting accurate predictions from Dimemas –

trace files are reduced in order to speed up prediction process; therefore, predicted times must be multiplied by a determined factor

REU - Florida International University 32

Page 33: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Remaining Work Next Week:

Finalizing scripting of Dimemas prediction simulations for the same scenarios of those of Amon and Aprof

Fall 2008: Experiment with how well Amon and Aprof relate to/could

possibly be combined with Dimemas Decide if and how to compare results from MareNostrum,

GCB, and Mind (i.e. the same versions of WRF would have to be running in all three locations)

Compose paper presenting significant results and submit paper to conference.

Future Work: Work with metascheduling team on implementation of tools.

33REU - Florida International University

Page 34: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

References S. Masoud Sadjadi, Liana Fong, Rosa

M. Badia, Javier Figueroa, Javier Delgado, Xabriel J. Collazo-Mojica, Khalid Saleem, Raju Rangaswami, Shu Shimizu, Hector A. Duran Limon, Pat Welsh, Sandeep Pattnaik, Anthony Praino, David Villegas, Selim Kalayci, Gargi Dasgupta, Onyeka Ezenwoye, Juan Carlos Martinez, Ivan Rodero, Shuyi Chen, Javier Muñoz, Diego Lopez, Julita Corbalan, Hugh Willoughby, Michael McFail, Christine Lisetti, and Malek Adjouadi. Transparent grid enablement of weather research and forecasting. In Proceedings of the Mardi Gras Conference 2008 - Workshop on Grid-Enabling Applications, Baton Rouge, Louisiana, USA, January 2008.

http://www.cs.fiu.edu/~sadjadi/Presentations/Mardi-Gras-GEA-2008-TGE-WRF.ppt

S. Masoud Sadjadi, Shu Shimizu, Javier Figueroa, Raju Rangaswami, Javier Delgado, Hector Duran, and Xabriel Collazo. A modeling approach for estimating execution time of long-running scientific applications. In Proceedings of the 22nd IEEE International Parallel & Distributed Processing Symposium (IPDPS-2008), the Fifth High-Performance Grid Computing Workshop (HPGC-2008), Miami, Florida, April 2008.

http://www.cs.fiu.edu/~sadjadi/Presentations/HPGC-2008-WRF%20Modeling%20Paper%20Presentationl.ppt “Performance/Profiling”. Presented by

Javier Figueroa in Special Topics in Grid Enablement of Scientific Applications Class. 13 May 2008

34REU - Florida International University

Page 35: Presented by: Marlon Bright 1 August 2008 Advisor: Masoud Sadjadi, Ph.D. REU – Florida International University

Acknowledgements

REU Partnerships for International Research

and Education (PIRE) The Barcelona SuperComputing Center

(BSC) Masoud Sadjadi, Ph. D. - FIU Rosa Badia, Ph.D. - BSC Javier Delgado – FIU Javier Figueroa – Univ. of Miami

35REU - Florida International University