26
SuperMUC – PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller Munich Network Management Team Ludwig-Maximilians-Universität München (LMU) & Leibniz Supercomputing Centre (LRZ)

SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

  • Upload
    others

  • View
    1

  • Download
    0

Embed Size (px)

Citation preview

Page 1: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

SuperMUC –PetaScale HPC at the Leibniz Supercomputing Centre (LRZ)Dieter Kranzlmüller

Munich Network Management TeamLudwig-Maximilians-Universität München (LMU) &Leibniz Supercomputing Centre (LRZ)

Page 2: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

SGI UV @ Leibniz Supercomputing Centre

D. Kranzlmüller Inbetriebnahme MACH 2

Page 3: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

Leibniz Supercomputing Centreof the Bavarian Academy of Sciences and Humanities

D. Kranzlmüller Inbetriebnahme MACH 3

Cuboidcontaining

computing systems72 x 36 x 36 meters

Institute Building

Institute Building

LectureHalls

VisualisationCentre

With 156 employees + 38 extra stafffor more than 90.000 students andfor more than 30.000 employees including 8.500 scientists

Page 4: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

Leibniz Supercomputing Centreof the Bavarian Academy of Sciences and Humanities

Computer Centre for all Munich Universities

D. Kranzlmüller Inbetriebnahme MACH 4

IT Service Provider:• Munich Scientific Network (MWN)• Web servers• e-Learning• E-Mail• Groupware• Special equipment:

• Virtual Reality Laboratory• Video Conference• Scanners for slides and large

documents• Large scale plotters

IT Competence Centre:• Hotline and support• Consulting (security, networking,

scientifc computing, …)• Courses (text editing, image processing,

UNIX, Linux, HPC, …)

Page 5: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

Leibniz Supercomputing Centreof the Bavarian Academy of Sciences and Humanities

European Supercomputing Centre

National Supercomputing Centre

Regional Computer Centre for all Bavarian Universities

Computer Centre for all Munich Universities

D. Kranzlmüller Inbetriebnahme MACH 5

Page 6: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

Leibniz Supercomputing Centreof the Bavarian Academy of Sciences and Humanities

European Supercomputing Centre

National Supercomputing Centre

Regional Computer Centre for all Bavarian Universities

Computer Centre for all Munich Universities

D. Kranzlmüller Inbetriebnahme MACH 6

SGI UV

SGI Altix

Linux Clusters

SuperMUC

Linux Hosting and Housing

Page 7: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

PRACE Timeline

7

HPCEUR HET

2004 2005 2006 2007 2008

PRACE – Partnership for Advanced Computing in Europe

D. Kranzlmüller Inbetriebnahme MACH

Page 8: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

The European Scientific Case• Weather, Climatology, Earth Science

– degree of warming, scenarios for our future climate.– understand and predict ocean properties and variations– weather and flood events

• Astrophysics, Elementary particle physics, Plasma physics– systems, structures which span a large range of different length and time scales– quantum field theories like QCD, ITER

• Material Science, Chemistry, Nanoscience– understanding complex materials, complex chemistry, nanoscience– the determination of electronic and transport properties

• Life Science– system biology, chromatin dynamics, large scale protein dynamics, protein

association and aggregation, supramolecular systems, medicine• Engineering

– complex helicopter simulation, biomedical flows, gas turbines and internal combustion engines, forest fires, green aircraft,

– virtual power plant8

Page 9: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

PRACE Timeline

D. Kranzlmüller Inbetriebnahme MACH 9

HPCEUR HET

2004 2005 2006 2007 2008

PRACE MoU

PRACE – Partnership for Advanced Computing in Europe

Page 10: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

PRACE Timeline

10

HPCEUR HET

2004 2005 2006 2007 2008

PRACE OperationPRACE Implementation Phase (1IP, 2IP)

2009 2010 2011 2012 2013

Phase

PRACE MoU PRACE Preparatory

EU-Grant: INFSO-RI-211528, 10 Mio. €

PRACE (AISBL), a legal entitywith (current) seat location in Brussels

23.4. 2010

PRACE – Partnership for Advanced Computing in Europe

D. Kranzlmüller Inbetriebnahme MACH

Page 11: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

PRACE Research Infrastructure Created• Establishment of the legal framework

– PRACE AISBL created with seat in Brussels in April(Association Internationale Sans But Lucratif)

– 20 members representing 20 European countries– Inauguration in Barcelona on June 9

11

Page 12: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

Gauss Centre for Supercomputing (GCS)

Joint activity of the 3 German National HPC Centres– John von Neumann Institut für Computing (NIC), Jülich– Leibniz Supercomputing Centre (LRZ), Garching near Munich– Höchstleistungsrechenzentrum Stuttgart (HLRS), Stuttgart

Largest and most powerful supercomputer infrastructure in Europe

Foundation of GCS (e.V.) April, 13th, 2007. Principal Partner in PRACE

(Partnership for Advanced Computing in Europe)

12D. Kranzlmüller Inbetriebnahme MACH

Page 13: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

PRACE Research Infrastructure Created• Establishment of the legal framework

– PRACE AISBL created with seat in Brussels in April(Association Internationale Sans But Lucratif)

– 20 members representing 20 European countries– Inauguration in Barcelona on June 9

• Funding secured for 2010 - 2015– 400 Million € from France, Germany, Italy, Spain

Provided as Tier-0 services on TCO basis– Funding decision for 100 Million € in The Netherlands

expected soon– 70+ Million € from EC FP7 for preparatory and implementation

Grants INFSO-RI-211528 and 261557 Complemented by ~ 60 Million € from PRACE members

13

Page 14: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

PRACE Tier-0 Systems• 1st Tier-0 System

– Jugene: BlueGene/P in GCS@Juelich– 72 Racks, 1 PFlop/s peak– 35% of capacity provided to PRACE

• 2nd Tier-0 System– Curie: Bull Cluster with Intel CPUs operated by CEA– 1.6 PFlop/s peak in Oct. 2011 (1st step in 10/2010)– Largest fraction of capacity provided to PRACE

• Next Procurements (in alphabetical order)– BSC, CINECA, GCS@HLRS, GCS@LRZ– Procurement plan based on analysis of user requirements and market

14

Page 15: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

PRACE Tier-0 Access• Single pan-European Peer Review• Early Access Call in May 2010

– 68 proposals asked for 1870 Million Core hours– 10 projects granted with 328 Million Core hours– Principal Investigators from D (5), UK (2) NL (1), I (1), PT (1)– Involves researchers from 31 institutions in 12 countries

• 1st Regular Call closed on August 2010– 58 proposals received asked for 2900 million core hours– 33 proposals have fulfilled the technical assessment– 360 million core hours available

for a 12 months allocation period starting November 2010• Further calls being scheduled (every 6 months)

– 2nd regular call will include both Jugene and Curie15

Page 16: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

I/O

nodes

NAS80 Gbit/s

18 Thin node islands (each >8000 cores)

1 Fat node island (8200 cores) SuperMIG

$HOME1.5 PB / 10 GB/s

Snapshots/Replika1.5 PB

(separate fire section)

non blocking non blocking

pruned tree (4:1)

SB-EP16 cores/node

2 GB/core

WM-EX40cores/node6.4 GB/core

10 PB

200 GB/s

Peak: 3 PFexpected: <3 MW>150,000 cores

GPFS for$WORK

$SCRATCH

InternetAchive and Backup ~ 30 PB

Desaster Recovery Site

Compute nodes Compute nodes

PRACE Tier-0 @ LRZ: SuperMUC

16D. Kranzlmüller Inbetriebnahme MACH

Page 17: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

System Parameters SuperMUC

17

Number of Islands (thin+fat) 18+1Number of nodes >9000Number of cores >150000Processor types (thin + fat) SB-EP + WM-EXTotal size of memory (TByte) >300Expected electrical power consumption of total system (kW) 2800Inlet temperature of compute node coolant (°C) > 30Outlet temperature of compute node coolant (range) (°C) 33 to 50Topology within an island fully nonblockingTopology between islands pruned tree (4:1)IB technology FDR10Theoretical bisection bandwith of the entire system (GByte/s) >11000Parallel file system type GPFSNAS user storage NetAppSize of parallel storage (Pbyte) 10Size of NAS user storage (PByte) 2 + (2 for Replica)Aggregate theoretical bandwidth to/from SAN/DAS storage (GByte/s) 200Aggregate theoretical bandwidth to/from NAS storage (GByte/s) 10

D. Kranzlmüller Inbetriebnahme MACH

Page 18: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

Energy Efficient Supercomputing at LRZ

18

1

10

100

1,000

10,000

100,000

1,000,000

10,000,000

GFl

op/s

, GB

yte

GFlop/s GByte

HLRB1: Hitachi SR8000+

HLRB2: SGI Altix 4700+ Linux-Cluster

10-fold every 4 yearsDouble every 14.1 Montths

Cray Y-MP8

IBM SP2Cray T90

Fujitsu

KSR

SuperMUCIBM

= 1Tera...

= 1 Peta...

= 1Giga

Cray Y-MP2

D. Kranzlmüller Inbetriebnahme MACH

Page 19: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

Financing Scheme

19

2010-2014Phase 1

2014-2016Phase 2

High End System

Investment Costs (Hardware and Software)

53 Mio € ~ 19 Mio €

Operating Costs (Electricity costs and maintenance for hardware und software, some additional personnel)

32 Mio € ~ 29 Mio €

SUM 85 Mio € ~ 48 Mio €Extension Buildings (construction and infrastructure) 49 Mio €

Funding for Phase 2 is announced but not legally secured

D. Kranzlmüller

for SuperMUC Investment and Operating Costs (gross, incl. VAT)

Page 20: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

Energy Consumption at LRZ

20

0

5.000

10.000

15.000

20.000

25.000

30.000

35.000

Pow

erco

nsum

ptio

nin

MW

h

C-HPC Workshop, Shanghai, September 29, 2011D. Kranzlmüller Inbetriebnahme MACH

Page 21: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

System Parameters SuperMUC

21

Number of Islands (thin+fat) 18+1Number of nodes >9000Number of cores >150000Processor types (thin + fat) SB-EP + WM-EXTotal size of memory (TByte) >300Expected electrical power consumption of total system (kW) 2800Inlet temperature of compute node coolant (°C) > 30Outlet temperature of compute node coolant (range) (°C) 33 to 50Topology within an island fully nonblockingTopology between islands pruned tree (4:1)IB technology FDR10Theoretical bisection bandwith of the entire system (GByte/s) >11000Parallel file system type GPFSNAS user storage NetAppSize of parallel storage (Pbyte) 10Size of NAS user storage (PByte) 2 + (2 for Replica)Aggregate theoretical bandwidth to/from SAN/DAS storage (GByte/s) 200Aggregate theoretical bandwidth to/from NAS storage (GByte/s) 10

D. Kranzlmüller Inbetriebnahme MACH

Page 22: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

SuperMUC - Energy Efficient Supercomputing

Probably most powerful x86-system in Europe (3PetaFlops peak)

Use for science in Europe (PRACE), Germany (GCS) and Bavaria (KONWIHR)

System with >150.000 cores, 324 TeraByte Main Memory

Most energy efficient General Purpose Supercomputer in Europe in 2012

– Hot liquid cooling– Reuse of waste heat– Hardware and software tools for clock scaling and optimization

(„dynamic frequency scaling“, „CPU throttling“ WIKIPEDIA)

22D. Kranzlmüller Inbetriebnahme MACH

Page 23: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

SuperMUC Thin Node Cooling

23D. Kranzlmüller Inbetriebnahme MACH

Page 24: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

Energy Efficiency at LRZ

Measures around SuperMUC– New Contract (Spot Market / Evaluation of alternative Technologies)– Optimization of Building and Cooling Infrastructure (additional cooling loop)– Hot liquid cooling PUE < 1,1– Cooperation LRZ / TUM / LMU / IBM on Tools and Provider / User Strategies– Cooperation with Building Management YIT

PRACE– 1 IP – Evaluation Prototype SGI Ultraviolet– 2 IP – Evaluation Prototype T-Platforms

Exascale EU Project DEEP:– System based on „Accelerator – Architecture“ (Intel MIC)– Cooling and Prototype Evaluation

Exascale EU Project Mont-Blanc– System based on low-power commercially available embedded CPUs– Next-generation HPC machine with a range of embedded technology – Software applications to run on this new generation of HPC systems

D. Kranzlmüller Inbetriebnahme MACH 24

Page 25: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

The MNM-Team AppleTV Cluster

4+2 nodes, connected by a 100 MBit Ethernet switch

www.appletvcluster.com

ATV2 Benchmarks: 4 nodes achieve Rmax

of 160.4 MFlops/Watt Power consumption: about 10 Watts (all 4 nodes) → Energy efficiency of 16 MFlops/Watt Green500 list #500: 21 MFlops/Watt

D. Kranzlmüller Inbetriebnahme MACH 25

Page 26: SuperMUC – PetaScale HPC at the Leibniz Supercomputing ...kranzlm/vortraege/2011-10-17 Linz - SuperMUC... · PetaScale HPC at the Leibniz Supercomputing Centre (LRZ) Dieter Kranzlmüller

SuperMUC Procurement Goals

26

To establish an integrated system and a programming environment which interact to enable the solution of the most challenging

scientific problems from widely varying scientific areas in the least amount of time.

D. Kranzlmüller Inbetriebnahme MACH