15
1 HPC Middleware on GRID HPC Middleware on GRID as a material for discussion as a material for discussion of WG5 of WG5 GeoFEM/RIST GeoFEM/RIST August 2nd, 2001, ACES/GEM at MHPCC August 2nd, 2001, ACES/GEM at MHPCC Kihei, Maui, Hawaii Kihei, Maui, Hawaii

1 HPC Middleware on GRID … as a material for discussion of WG5 GeoFEM/RIST August 2nd, 2001, ACES/GEM at MHPCC Kihei, Maui, Hawaii

Embed Size (px)

Citation preview

Page 1: 1 HPC Middleware on GRID … as a material for discussion of WG5 GeoFEM/RIST August 2nd, 2001, ACES/GEM at MHPCC Kihei, Maui, Hawaii

1

HPC Middleware on GRIDHPC Middleware on GRID… … as a material for discussion of as a material for discussion of

WG5WG5

GeoFEM/RISTGeoFEM/RISTAugust 2nd, 2001, ACES/GEM at MHPCCAugust 2nd, 2001, ACES/GEM at MHPCC

Kihei, Maui, HawaiiKihei, Maui, Hawaii

Page 2: 1 HPC Middleware on GRID … as a material for discussion of WG5 GeoFEM/RIST August 2nd, 2001, ACES/GEM at MHPCC Kihei, Maui, Hawaii

2

BackgroundBackground• Various Types of HPC Platforms– MPP, VPP– PC Clusters, Distributed Parallel MPPs, SMP Clusters– 8-Way SMP, 16-Way SMP, 256-Way SMP– Power, HP-RISC, Alpha/Itanium, Pentium, Vector PE• Parallel/Single PE Optimization is Important Issue for Efficiency– Everyone knows that ... but it's a big task especially for application experts such as geophysics people in ACES community.– Machine dependent optimization/tuning required.• Simulation Methods such as FEM/FDM/BEM/LSM/DEM etc. have Typical Processes for Computation.• How about "Hiding" these Processes from Users ?– code development : efficient, reliable, portable, maintenance-free• line number of the source codes will be reduced– accelerates advancement of the applications (= physics)

Page 3: 1 HPC Middleware on GRID … as a material for discussion of WG5 GeoFEM/RIST August 2nd, 2001, ACES/GEM at MHPCC Kihei, Maui, Hawaii

3

Background (cont.)Background (cont.)• Current GeoFEM provides this environment

– limited to FEM– not necessarily perfect

• GRID as next generation HPC infrastructure– Currently, middlewares and protocols are being developed t

o enable unified interface to treat various OS, computers, ultra-speed network and database.

– What are expected to GRID ?• Meta-computing : simultaneous use of supercomputers in the world• Volunteer-computing : efficient use of idling computers• Access Grid : research collaboration environment• Data Intensive Computing : computation with large-scale data• Grid ASP : application services on WEB

Page 4: 1 HPC Middleware on GRID … as a material for discussion of WG5 GeoFEM/RIST August 2nd, 2001, ACES/GEM at MHPCC Kihei, Maui, Hawaii

4

Similar Research GroupsSimilar Research Groups• ALICEALICE (( ANLANL ))

• CCAforumCCAforum (( Common Component ArchitectureCommon Component Architecture,,DOEDOE ))• DOE/ASCI/Distributed Computing Research TeamDOE/ASCI/Distributed Computing Research Team

– ESI(Equation Solver Interface Standards)ESI(Equation Solver Interface Standards)– FEI(The Finite Element/Equation Solver Interface Specification)FEI(The Finite Element/Equation Solver Interface Specification)

• ADRADR (( Active Data RepositoryActive Data Repository )()( NPACINPACI ))

Page 5: 1 HPC Middleware on GRID … as a material for discussion of WG5 GeoFEM/RIST August 2nd, 2001, ACES/GEM at MHPCC Kihei, Maui, Hawaii

5

Are they successful ? Are they successful ? It seems NOIt seems NO

• Very limited targets, processesVery limited targets, processes– Mainly for Optimization of Linear SolversMainly for Optimization of Linear Solvers

• Where are Interfaces between Applications and Libraries ?Where are Interfaces between Applications and Libraries ?– Approach from Computer/Computational Science PeopleApproach from Computer/Computational Science People– Not Really Easy to Use by Application PeopleNot Really Easy to Use by Application People

-Linear solvers -Numerical Algorithms -Parallel Programming -Optimization

Computer/Computational

Science

-FEM -FDM -Spectral Methods -MD, MC -BEM

Applications

Page 6: 1 HPC Middleware on GRID … as a material for discussion of WG5 GeoFEM/RIST August 2nd, 2001, ACES/GEM at MHPCC Kihei, Maui, Hawaii

6

Example of HPC Middleware Example of HPC Middleware (1)(1)

Simulation Methods include Some Simulation Methods include Some Typical ProcessesTypical Processes

Sparse Mat. Mult.Sparse Mat. Mult.

Nonlinear ProcedureNonlinear Procedure

FFTFFT

Eward TermsEward Terms

O(N) Ab Initio MDO(N) Ab Initio MD

Page 7: 1 HPC Middleware on GRID … as a material for discussion of WG5 GeoFEM/RIST August 2nd, 2001, ACES/GEM at MHPCC Kihei, Maui, Hawaii

7Example of HPC Middleware Example of HPC Middleware (2)(2)

Individual Process could be optimized Individual Process could be optimized for Various Types of MPP for Various Types of MPP

ArchitecturesArchitectures

Sparse Mat. Mult.

Nonlinear Proc.

FFT

Eward Terms

O(N) Ab Initio MD MPP-A

MPP-B

MPP-C

Page 8: 1 HPC Middleware on GRID … as a material for discussion of WG5 GeoFEM/RIST August 2nd, 2001, ACES/GEM at MHPCC Kihei, Maui, Hawaii

8

Sparse Matrix Mult.Sparse Matrix Mult.

Nonlinear Proc.Nonlinear Proc.

FFTFFT

Eward TermsEward Terms

Example of HPC Middleware Example of HPC Middleware (3)(3)

Use Optimized LibrariesUse Optimized Libraries

OO(N)(N)

abab--

initioinitio

MMDD

Sparse Matrix Mult.Sparse Matrix Mult.

Nonlinear Proc.Nonlinear Proc.

FFTFFT

Eward TermsEward Terms

Sparse Matrix Mult.Sparse Matrix Mult.

Nonlinear Alg.Nonlinear Alg.

FFTFFT

Eward TermsEward Terms

Sparse Matrix Mult.Sparse Matrix Mult.

Nonlinear Proc.Nonlinear Proc.

FFTFFT

Eward TermsEward Terms

Page 9: 1 HPC Middleware on GRID … as a material for discussion of WG5 GeoFEM/RIST August 2nd, 2001, ACES/GEM at MHPCC Kihei, Maui, Hawaii

9Example of HPC Middleware Example of HPC Middleware (4)(4)- Optimized code is generated by special language/ - Optimized code is generated by special language/ compiler based on analysis data and H/W compiler based on analysis data and H/W information.information.- Optimum algorithm can be adopted- Optimum algorithm can be adopted

OO(N)(N)

abab--

initioinitio

MMDD

MPP-AMPP-A

MPP-BMPP-B

MPP-CMPP-C

Data for Analysis Model

ParametersParametersof H/Wof H/W

SpecialSpecialCompilerCompiler

Sparse Matrix Mult.Sparse Matrix Mult.

Nonlinear Proc.Nonlinear Proc.

FFTFFT

Eward TermsEward Terms

Page 10: 1 HPC Middleware on GRID … as a material for discussion of WG5 GeoFEM/RIST August 2nd, 2001, ACES/GEM at MHPCC Kihei, Maui, Hawaii

10Example of HPC Middleware Example of HPC Middleware (5)(5)- On network-connected H/W's (meta-computing)- On network-connected H/W's (meta-computing)- Optimized for individual architecture- Optimized for individual architecture- Optimum load-balancing- Optimum load-balancing

OO(N)(N)

abab--

initioinitio

MMDD

analysisanalysismodelmodelspacespace

Page 11: 1 HPC Middleware on GRID … as a material for discussion of WG5 GeoFEM/RIST August 2nd, 2001, ACES/GEM at MHPCC Kihei, Maui, Hawaii

11Example of HPC Middleware Example of HPC Middleware (6)(6)

Multi Module Coupling through Multi Module Coupling through PlatformPlatform

HPC Platform/MiddlewareHPC Platform/Middleware

Ab-Initio MDAb-Initio MD Classical MDClassical MD FEMFEM

HPC Platform/MiddlewareHPC Platform/Middleware

Ab-Initio MDAb-Initio MD Classical MDClassical MD FEMFEM

ModelingModeling VisualizationVisualization LoadLoadBalancingBalancing

ResourceResourceManagementManagementOptimizationOptimizationDataData

AssimilationAssimilation

Page 12: 1 HPC Middleware on GRID … as a material for discussion of WG5 GeoFEM/RIST August 2nd, 2001, ACES/GEM at MHPCC Kihei, Maui, Hawaii

12

PETAFLOPS on GRIDPETAFLOPS on GRIDfrom GeoFEM's Point of Viewfrom GeoFEM's Point of View

• Why? When?Why? When?– Datasets (mesh, observation, result) Datasets (mesh, observation, result)

could be distributed.could be distributed.– Problem size could be too large for sProblem size could be too large for s

ingle MPP system.ingle MPP system.• according to G.C.Fox, according to G.C.Fox, (TOP500) is (TOP500) is

about 100 TFLOPS now ...about 100 TFLOPS now ...

• LegionLegion– Prof.Grimshaw (U.Virginia)Prof.Grimshaw (U.Virginia)– Grid OS, Global OSGrid OS, Global OS– Can handle MPP's connected througCan handle MPP's connected throug

h network as one huge MPP (= Supeh network as one huge MPP (= Super MPP)r MPP)

MPP-AMPP-A

MPP-BMPP-B

MPP-CMPP-C

– Optimization on Individual Optimization on Individual Architecture (H/W)Architecture (H/W)

– Load balancing according to Load balancing according to machine performance and machine performance and resource availabilityresource availability

Page 13: 1 HPC Middleware on GRID … as a material for discussion of WG5 GeoFEM/RIST August 2nd, 2001, ACES/GEM at MHPCC Kihei, Maui, Hawaii

13

PETAFLOPS on GRID (cont.)PETAFLOPS on GRID (cont.)

• GRID + (OS) + HPC MW/PFGRID + (OS) + HPC MW/PF• Environment for "Electronic CollaborationEnvironment for "Electronic Collaboration

Page 14: 1 HPC Middleware on GRID … as a material for discussion of WG5 GeoFEM/RIST August 2nd, 2001, ACES/GEM at MHPCC Kihei, Maui, Hawaii

14

Page 15: 1 HPC Middleware on GRID … as a material for discussion of WG5 GeoFEM/RIST August 2nd, 2001, ACES/GEM at MHPCC Kihei, Maui, Hawaii

15

"Parallel" FEM Procedure"Parallel" FEM Procedure

Initial Mesh DataInitial Mesh Data

PartitioningPartitioning

Post ProcPost Proc..Data Input/OutputData Input/Output

Domain Specific Domain Specific Algorithms/ModelsAlgorithms/Models

Matrix AssembleMatrix Assemble

Linear SolversLinear Solvers

VisualizationVisualization

Pre-ProcessingPre-Processing MainMain Post-ProcessingPost-Processing