15
Center for Computational Sciences OAK RIDGE NATIONAL LABORATORY U. S. DEPARTMENT OF ENERGY Vision for OSC Computing and Computational Sciences http:// www.ccs.ornl.gov/ Thomas Zacharia Associate Laboratory Director Computing and Computational Sciences Oak Ridge National Laboratory Earth Simulator Rapid Response Meeting May 15-16, 2002

Vision for OSC Computing and Computational Sciences

  • Upload
    lukas

  • View
    45

  • Download
    0

Embed Size (px)

DESCRIPTION

Vision for OSC Computing and Computational Sciences. Thomas Zacharia Associate Laboratory Director Computing and Computational Sciences Oak Ridge National Laboratory. http://www.ccs.ornl.gov/. Earth Simulator Rapid Response Meeting May 15-16, 2002. Charge from Dr. Orbach. - PowerPoint PPT Presentation

Citation preview

Page 1: Vision for OSC Computing and Computational Sciences

Cen

ter f

or C

ompu

tatio

nal S

cien

ces

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Vision for OSC Computing and Computational Sciences

http://www.ccs.ornl.gov/

Thomas ZachariaAssociate Laboratory Director

Computing and Computational SciencesOak Ridge National Laboratory

Earth Simulator Rapid Response Meeting May 15-16, 2002

Page 2: Vision for OSC Computing and Computational Sciences

Cen

ter f

or C

ompu

tatio

nal S

cien

ces

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Charge from Dr. Orbach

• Review “. . . current state of the national computer vendor community relative to high performance computing”

• “. . . Vision for what realistically should be accomplished in the next five years within the Office of Science in high performance computing”

Page 3: Vision for OSC Computing and Computational Sciences

Cen

ter f

or C

ompu

tatio

nal S

cien

ces

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Dr. Orbach’s Vision for OSC ComputingStatement to ASCAC Committee, May 8, 2002

• “… there is a centrality of computation in everything that we do”

• “… large scale computation is the future of every program in the Office of Science”

• “… we want to have our own computing program in non-defense computational science”

Page 4: Vision for OSC Computing and Computational Sciences

Cen

ter f

or C

ompu

tatio

nal S

cien

ces

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

FY 03 Budget Request for OSC Computing Considerably Lower

than Required to Meet Goals

0

100,000,000

200,000,000

300,000,000

400,000,000

500,000,000

600,000,000

700,000,000

800,000,000

900,000,000

DOE-SC NNSA NSF

Fiscal Year

Bud

get D

olla

rs

Page 5: Vision for OSC Computing and Computational Sciences

Cen

ter f

or C

ompu

tatio

nal S

cien

ces

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

As Fraction of Total Budget, OSC is Half NNSA and NSF and Needs Significant

Increase to Meet Goals

0

2

4

6

8

10

12

DOE-SC NNSA NSF

Com

putin

g B

udge

t / T

otal

Bud

get (

%)

Page 6: Vision for OSC Computing and Computational Sciences

Cen

ter f

or C

ompu

tatio

nal S

cien

ces

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Earth Simulator has Heightened Urgency for Infrastructure Strategy for

Scientific ComputingCritical Steps:• Invest in critical software

with integrated science, and computer science development teams

• Deploy scientific computing hardware infrastructure in support of “large scale computation”– Cray, HP, IBM, SGI– IBM is the largest US

installation• Develop new initiative to

support advanced architecture research

Top 500 SupercomputersUS has been #1 in 12 of 19 lists

A concerted effort will be required to regain US leadership in high performancecomputing. The LINPACK benchmark generally overestimates the effectivenessof an architecture for applications such as climate by a substantial factor. Stability and reliability are also important system properties.

Page 7: Vision for OSC Computing and Computational Sciences

Cen

ter f

or C

ompu

tatio

nal S

cien

ces

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Invest in Critical Software with Integrated Science and Computer Science

Development Teams SciDAC: a Good Start Towards Scientific Computing Software

• Scientific Applications– Climate Simulation– Computational Chemistry– Fusion – 5 Topics– High Energy Nuclear

Physics – 5 Topics• Collaboratories

– Four Projects• Middleware & Network

Research– Six Projects

• Computer Science– Scalable Systems Software– Common Component

Architecture– Performance Science and

Engineering– Scientific Data Management

• Applied Mathematics– PDE Linear/Nonlinear

Solvers and Libraries– Structured Grids/AMR– Unstructured Grids

Dave Bader, SciDAC PI Meeting, Jan 15, 2002, Washington DC

Page 8: Vision for OSC Computing and Computational Sciences

Cen

ter f

or C

ompu

tatio

nal S

cien

ces

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Deploy Scientific Computing Hardware Infrastructure to Support

“Large-Scale Computation”• Provide most effective and efficient computing resources for a set of

scientific applications• Serve as focal point for scientific research community as it adapts to

new computing technologies• Provide organizational framework needed for multidisciplinary

activities– Addressing software challenges requires strong, long term collaborations

among disciplinary computational scientists, computer scientists, and applied mathematicians

• Provide organizational framework needed for development of community codes– Implementing many scientific codes requires wide range of disciplinary

expertise• Organizational needs will continue to grow as computers advance to

petaflops scale

Dave Bader, SciDAC PI Meeting, Jan. 15, 2002, Washington, DC

Page 9: Vision for OSC Computing and Computational Sciences

Cen

ter f

or C

ompu

tatio

nal S

cien

ces

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Earth Simulator has Widened Gap with DOE Scientific Computing Hardware Infrastructure

Top left: comparison between ES and SC resources – highlights widening gap between SC capabilities and others

Top right: comparison between ES and US resources of comparable peak performance – highlights architectural difference and need for new initiative to close the gap

Right: comparison between ES and US resources of comparable cost

1,000

3,000

5,000

7,000

Earth Simulator SEABORG CHEETAH

Sim

ulat

ions

yea

rs/d

ay

WideningGap

1,000

3,000

5,000

7,000

Earth Simulator POWER4 H+ (40TFlops) Power5 (50 TFlops)

Sim

ulat

ions

yea

rs/d

ay TechnologyGap

0

2,000

4,000

6,000

8,000

10,000

Earth Simulator POWER4 H+ (3*40TFlops) Power5 (3*50 TFlops)

Sim

ulat

ions

yea

rs/d

ay

Page 10: Vision for OSC Computing and Computational Sciences

Cen

ter f

or C

ompu

tatio

nal S

cien

ces

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Possible U.S. Response in the Near Term for Increased Computing Capacity

• 40 TFlops Peak• 5120 Vector Processors• 8 GFlops Processor• 8 Processors per Node• $500 M Procurement• $50M/yr Maintenance• Limited Software Investment to date• Significant Ancillary Impact on

Biology, Nanoscience, Astrophysics, HENP, Fusion

• 40 TFlops Peak• 5120 Power5 Processors• 8 GFlops Processor• 64 Processors per Node• $100 M Procurement• $10M/yr Maintenance• SciDAC Investment in Computational

Science and related ISICs• Significant Ancillary Impact on

Biology, Nanoscience, Astrophysics, HENP, Fusion

Earth Simulator US Alternative

Page 11: Vision for OSC Computing and Computational Sciences

Cen

ter f

or C

ompu

tatio

nal S

cien

ces

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Best Performance of High Resolution Atmospheric Model

Inter-node bandwidth (Mb/s)

Performance of Hi-Resolution Atmospheric Model105

104

103

102

101

102 103 104 105

Earth Simulator (2560)

AlphaES45 (2048)

AlphaES40 (256)

SP3 WHII (512)T3E (512)

GFl

ops

Page 12: Vision for OSC Computing and Computational Sciences

Cen

ter f

or C

ompu

tatio

nal S

cien

ces

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

1996 1997 1998 1999 2000 2001 2002 2003 2004 2005

Year

100

1000

10000

100000

1000000D

olla

rs/G

Flop

s

C/L/DASCIT3E

BeowulfsCOTSJPL

QCDSPColumbia QCDOC

Columbia/IBM

Blue Gene/L

ASCI Blue

ASCI White ASCI Compaq

Develop New Initiative to Support Advanced Architecture:

BlueGene Offers Possible Option

Page 13: Vision for OSC Computing and Computational Sciences

Cen

ter f

or C

ompu

tatio

nal S

cien

ces

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

QCDSP (600GF based on Texas Instruments DSP C31)– Gordon Bell Prize for Most Cost Effective Supercomputer in '98– Columbia University Designed and Built– Optimized for Quantum Chromodynamics (QCD)– 12,000 50MF Processors– Commodity 2MB DRAM

QCDOC (20TF based on IBM System-on-a-Chip)– Collaboration between Columbia University and IBM Research– Optimized for QCD– IBM 7SF Technology (ASIC Foundry Technology)– 20,000 1GF processors (nominal)– 4MB Embedded DRAM + External Commodity DDR/SDR SDRAM

BlueGene L/D (180TF based on IBM System-on-a-Chip)– Designed by IBM Research in IBM CMOS 8SF Technology– 64,000 2.8GF processors (nominal)– 4MB Embedded DRAM + External Commodity DDR SDRAM

BlueGene Architecture is a (more) General Purpose Machine that builds on QCDOC

Page 14: Vision for OSC Computing and Computational Sciences

Cen

ter f

or C

ompu

tatio

nal S

cien

ces

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Host System:– Diagnostics, booting, archive– Application dependent requirements

System Organization (conceptual)

File Server Array– ~ 500 RAID PC servers– Gb Ethernet and/or Infiniband– Application dependent requirements

BlueGene/L Processing Nodes81920 Nodes– Two major partitions– 65536 nodes production– Platform (256 TFlops peak)– 16384 nodes partitioned into code

development platforms

Host Computer

Top View of system

50 Feet

Page 15: Vision for OSC Computing and Computational Sciences

Cen

ter f

or C

ompu

tatio

nal S

cien

ces

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Summary• Continue investment in critical software with integrated

science, and computer science development teams• Deploy scientific computing hardware infrastructure in

support of “large scale computation”• Develop new initiative to support advanced

architecture research • Develop a bold new facilities strategy for OSC

computing• Increase OSC computing budget to support outlined

strategy

Without sustained commitment to scientific computing, key computing and computational sciences capabilities,

including personnel, will erode beyond recovery.