41
GriPhyN Project Overview Paul Avery University of Florida [email protected] GriPhyN NSF Project Review 29-30 January 2003 Chicago

GriPhyN Project Overview Paul Avery University of Florida [email protected] GriPhyN NSF Project Review 29-30 January 2003 Chicago

Embed Size (px)

Citation preview

Page 1: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

GriPhyN Project Overview

Paul AveryUniversity of [email protected]

GriPhyN NSF Project Review29-30 January 2003

Chicago

Page 2: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

229 Jan 2003Paul Avery, University of Florida [email protected]

GriPhyN = Experiments + CS + Grids

• GriPhyN = Grid Physics Network– Computer Scientists (Globus, Condor, SRB, …)– Physicists from 4 frontier physics/astronomy expts.

• GriPhyN basics (2000 – 2005)– $11.9M (NSF) + $1.6M (matching)– 17 universities, SDSC, 3 labs, ~80 people– Integrated Outreach effort (UT Brownsville)

• Management– Paul Avery (Florida) co-Director– Ian Foster (Chicago) co-Director– Mike Wilde (Argonne) Project Coordinator– Rick Cavanaugh (Florida) Deputy Coordinator

Page 3: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

329 Jan 2003Paul Avery, University of Florida [email protected]

GriPhyN Institutions (Sep. 2000)– U Florida– U Chicago– Boston U– Caltech– U Wisconsin, Madison– USC/ISI– Harvard– Indiana– Johns Hopkins– Texas A&M– Stanford

– U Illinois at Chicago– U Penn– U Texas, Brownsville– UC Berkeley– U Wisconsin, Milwaukee– UC San Diego– SDSC– Lawrence Berkeley Lab– Argonne– Fermilab– Brookhaven

Funded by GriPhyN

Page 4: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

429 Jan 2003Paul Avery, University of Florida [email protected]

GriPhyN Vision• Create tools to enable collaborative research

– Large research teams

• … by global scientific communities– International distribution of people and resources

• … at petascale levels– PetaOps + PetaBytes + Performance

• … in a transparent way– Scientists think in terms of their science

Page 5: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

529 Jan 2003Paul Avery, University of Florida [email protected]

GriPhyN Science Drivers

• US-CMS & US-ATLAS– HEP experiments at LHC/CERN– 100s of Petabytes

• LIGO– Gravity wave experiment– 100s of Terabytes

• Sloan Digital Sky Survey– Digital astronomy (1/4 sky)– 10s of Terabytes

Data

gro

wth

Com

mu

nit

y g

row

th

2007

2002

2001

Massive CPU Large, distributed datasets Large, distributed communities

Page 6: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

629 Jan 2003Paul Avery, University of Florida [email protected]

GriPhyN Goals• Conduct CS research to achieve vision

– Virtual Data as unifying principle– Planning, execution, performance monitoring

• Disseminate through Virtual Data Toolkit

• Integrate into GriPhyN science experiments– Common Grid tools, services

• Impact other disciplines– HEP, biology, medicine, virtual astronomy, eng.– Other Grid projects

• Educate, involve, train students in IT research– Undergrads, grads, postdocs, – Underrepresented groups

Page 7: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

729 Jan 2003Paul Avery, University of Florida [email protected]

Goal: PetaScale Virtual-Data Grids

Virtual Data Tools

Request Planning &Scheduling Tools

Request Execution & Management Tools

Transforms

Distributed resources(code, storage, CPUs,networks)

ResourceManagement

Services

Security andPolicy

Services

Other GridServices

Interactive User Tools

Production TeamSingle Researcher Workgroups

Raw datasource

PetaOps Petabytes Performance

Page 8: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

829 Jan 2003Paul Avery, University of Florida [email protected]

CMS Experiment

Example: Global LHC Data Grid

Online System

CERN Computer Center > 20 TIPS

USAKorea RussiaUK

Institute

100-200 MBytes/s

2.5 Gbits/s

1 Gbits/s

2.5 - 10 Gbits/s

~0.6 Gbits/s

Tier 0

Tier 1

Tier 3

Tier 2

Physics cachePCs, other portals

Institute

Institute

Institute

Tier2 Center

Tier2 Center

Tier2 Center

Tier2 Center

Tier 4

Page 9: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

929 Jan 2003Paul Avery, University of Florida [email protected]

GriPhyN Project Challenges

• We balance and coordinate– CS research

with “goals, milestones & deliverables”

– GriPhyN schedule/priorities/riskswith those of the 4 experiments

– General tools developed by GriPhyNwith specific tools developed by 4 experiments

– Data Grid design, architecture & deliverableswith those of other Grid projects

• Appropriate balance requires– Tight management, close coordination, trust

• We have (so far) met these challenges– But requires constant attention, good will

Page 10: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

1029 Jan 2003Paul Avery, University of Florida [email protected]

External Advisory Committee

Physics Experiments

Project DirectorsPaul AveryIan Foster

Inte

rne

t 2

DO

E

Sc

ien

ce

NS

F P

AC

Is

Project CoordinationMike Wilde

Rick Cavanaugh

Outreach/EducationManuela Campanelli

Industrial Connections

Ian Foster / Paul Avery

EDG, LCG,Other Grid Projects

ArchitectureCarl Kesselman

VDT DevelopmentCoord.: M. Livny

Requirements, Definition & Scheduling(Miron Livny)

Integration, Testing, Documentation, Support

(Alain Roy)

Globus Project & NMI Integration

(Carl Kesselman)

CS ResearchCoord.: I. Foster

Virtual Data(Mike Wilde)

Request Planning & Scheduling

(Ewa Deelman)

Execution Management(Miron Livny)

Measurement, Monitoring & Prediction

(Valerie Taylor)

ApplicationsCoord.: R. Cavanaugh

ATLAS(Rob Gardner)

CMS(Rick Cavanaugh)

LIGO(Albert Lazzarini)

SDSS(Alexander Szalay)

Inter-Project Coordination:

R. Pordes

HICB(Larry Price)

HIJTB(Carl Kesselman)

PPDG(Ruth Pordes)

TeraGrid, NMI, etc.(TBD)

International (EDG, etc)(Ruth Pordes)

GriPhyNManagement

iVDGL

iVDGLRob Gardner

Page 11: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

1129 Jan 2003Paul Avery, University of Florida [email protected]

External Advisory Committee• Members

– Fran Berman (SDSC Director)– Dan Reed (NCSA Director)– Joel Butler (former head, FNAL Computing Division)– Jim Gray (Microsoft)– Bill Johnston (LBNL, DOE Science Grid)– Fabrizio Gagliardi (CERN, EDG Director)– David Williams (former head, CERN IT)– Paul Messina (former CACR Director)– Roscoe Giles (Boston U, NPACI-EOT)

• Met with us 3 times: 4/2001, 1/2002, 1/2003– Extremely useful guidance on project scope & goals

Page 12: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

1229 Jan 2003Paul Avery, University of Florida [email protected]

Integration of GriPhyN and iVDGL• International Virtual-Data Grid Laboratory

– A global Grid laboratory (US, EU, Asia, …)– A place to conduct Data Grid tests “at scale”– A mechanism to create common Grid infrastructure– A laboratory for Grid tests by other disciplines

• Tight integration with GriPhyN– Testbeds– VDT support– Outreach– Common External Advisory Committee

• International participation– DataTag (EU)– UK e-Science programme: support 6 CS Fellows

Page 13: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

1329 Jan 2003Paul Avery, University of Florida [email protected]

GriPhyN/iVDGL Basics

• Both NSF funded, overlapping periods– GriPhyN: $11.9M (NSF) + $1.6M (match) (2000–2005)– iVDGL: $13.7M (NSF) + $2M (match) (2001–2006)

• Basic composition– GriPhyN: 12 universities, SDSC, 3 labs (~82 people)– iVDGL: 16 institutions, SDSC, 3 labs (~84 people)– Large overlap: people, institutions, experiments

• GriPhyN (Grid research) vs iVDGL (Grid deployment)– GriPhyN: 2/3 “CS” + 1/3 “physics” ( 0% H/W)– iVDGL: 1/3 “CS” + 2/3 “physics” (20% H/W)– Virtual Data Toolkit (VDT) in common– Testbeds in common

Page 14: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

1429 Jan 2003Paul Avery, University of Florida [email protected]

U Florida CMSCaltech CMS, LIGOUC San Diego CMS, CSIndiana U ATLAS, iGOCBoston U ATLASU Wisconsin, Milwaukee LIGOPenn State LIGOJohns Hopkins SDSS, NVOU Chicago CSU Southern California CSU Wisconsin, Madison CSSalish Kootenai Outreach, LIGOHampton U Outreach, ATLASU Texas, Brownsville Outreach, LIGOFermilab CMS, SDSS, NVOBrookhaven ATLASArgonne Lab ATLAS, CS

iVDGL Institutions

T2 / Software

CS support

T3 / Outreach

T1 / Labs(not funded)

Page 15: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

1529 Jan 2003Paul Avery, University of Florida [email protected]

US-iVDGL Sites (Spring 2003)

Partners?EUCERNBrazilAustraliaKoreaJapan

UF

Wisconsin

BNL

Indiana

Boston USKC

Brownsville

Hampton

PSU

J. Hopkins

Caltech

Tier1Tier2Tier3

FIU

FSUArlington

Michigan

LBL

Oklahoma

Argonne

Vanderbilt

UCSD/SDSC

NCSA

Fermilab

Page 16: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

1629 Jan 2003Paul Avery, University of Florida [email protected]

Example: US-CMS Grid Testbed

UCSD

Florida

Wisconsin

Caltech

Fermilab

FIU

FSU

Brazil

Korea

CERN

Page 17: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

1729 Jan 2003Paul Avery, University of Florida [email protected]

iVDGL Management & Coordination

Project Coordination Group

US External Advisory Committee

GLUE Interoperability Team

Collaborating Grid Projects

TeraGrid

EDG Asia

DataTAG

BTEV

LCG?

BioALICE Geo

?

D0 PDC CMS HI ?

US ProjectDirectors

Outreach Team

Core Software Team

Facilities Team

Operations Team

Applications Team

International Piece

US Project Steering Group

U.S. Piece

GriPhyN Mike Wilde

Page 18: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

1829 Jan 2003Paul Avery, University of Florida [email protected]

Meetings in 2000-2001•GriPhyN/iVDGL meetings

– Oct. 2000 All-hands Chicago– Dec. 2000 Architecture Chicago– Apr. 2001 All-hands, EAC USC/ISI– Aug. 2001 Planning Chicago– Oct. 2001 All-hands, iVDGL USC/ISI

•Numerous smaller meetings– CS-experiment– CS research– Liaisons with PPDG and EU DataGrid– US-CMS and US-ATLAS computing reviews– Experiment meetings at CERN

Page 19: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

1929 Jan 2003Paul Avery, University of Florida [email protected]

Meetings in 2002

• GriPhyN/iVDGL meetings– Jan. 2002 EAC, Planning, iVDGL Florida– Mar. 2002 Outreach Workshop

Brownsville– Apr. 2002 All-hands Argonne– Jul. 2002 Reliability Workshop ISI– Oct. 2002 Provenance Workshop Argonne– Dec. 2002 Troubleshooting Workshop Chicago– Dec. 2002 All-hands technical ISI +

Caltech– Jan. 2003 EAC SDSC

• Numerous other 2002 meetings– iVDGL facilities workshop (BNL)– Grid activities at CMS, ATLAS meetings– Several computing reviews for US-CMS, US-ATLAS– Demos at IST2002, SC2002– Meetings with LCG (LHC Computing Grid) project– HEP coordination meetings (HICB)

Page 20: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

2029 Jan 2003Paul Avery, University of Florida [email protected]

Progress: CS, VDT, Outreach• Lots of good CS research (Later talks)

• Installation revolution: VDT + Pacman (Later talk)– Several major releases this year: VDT 1.1.5– VDT/Pacman vastly simplify Grid software installation– Used by all experiments– Agreement to use VDT by LHC Computing Grid Project

• Grid integration in experiment s/w (Later talks)

• Expanding education/outreach (Later talk)– Integration with iVDGL– Collaborations: PPDG, NPACI-EOT, SkyServer, QuarkNet– Meetings, brochures, talks, …

Page 21: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

2129 Jan 2003Paul Avery, University of Florida [email protected]

Progress: Student Participation• Integrated student involvement

– CS research– VDT deployment, testing, support– Integrating Grid tools in physics experiments– Cluster building, testing– Grid software deployment– Outreach, web development

• Integrated postdoc involvement– Involvement in all areas– Necessary when students not sufficient

Page 22: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

2229 Jan 2003Paul Avery, University of Florida [email protected]

Global Context: Data Grid Projects

• U.S. Infrastructure Projects– GriPhyN (NSF)– iVDGL (NSF)– Particle Physics Data Grid (DOE)– TeraGrid (NSF)– DOE Science Grid (DOE)

• EU, Asia major projects– European Data Grid (EDG) (EU, EC)– EDG related national Projects (UK, Italy, France, …)– CrossGrid (EU, EC)– DataTAG (EU, EC)– LHC Computing Grid (LCG) (CERN)– Japanese Project– Korea project

Page 23: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

2329 Jan 2003Paul Avery, University of Florida [email protected]

U.S. Project Coordination: Trillium• Trillium = GriPhyN + iVDGL + PPDG

– Large overlap in leadership, people, experiments

• Benefit of coordination– Common S/W base + packaging: VDT + PACMAN– Low overhead for collaborative or joint projects:

security, monitoring, newsletter, prod. grids, demos– Wide deployment of new technologies, e.g. Virtual Data– Stronger, more extensive outreach effort

• Forum for US Grid projects– Joint view, strategies, meetings and work– Unified entity to deal with EU & other Grid projects

• “Natural” collaboration across DOE and NSF projects– Funding agency interest?

Page 24: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

2429 Jan 2003Paul Avery, University of Florida [email protected]

GriPhyN = Experiments + CS + Grids

• GriPhyN = Grid Physics Network– Computer Scientists (Globus, Condor, SRB, …)– Physicists from 4 frontier physics/astronomy expts.

• GriPhyN basics (2000 – 2005)– $11.9M (NSF) + $1.6M (matching)– 17 universities, SDSC, 3 labs, ~80 people– Integrated Outreach effort (UT Brownsville)

• Management– Paul Avery (Florida) co-Director– Ian Foster (Chicago) co-Director– Mike Wilde (Argonne) Project Coordinator– Rick Cavanaugh (Florida) Deputy Coordinator

Page 25: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

2529 Jan 2003Paul Avery, University of Florida [email protected]

GriPhyN Institutions (Sep. 2000)– U Florida– U Chicago– Boston U– Caltech– U Wisconsin, Madison– USC/ISI– Harvard– Indiana– Johns Hopkins– Texas A&M– Stanford

– U Illinois at Chicago– U Penn– U Texas, Brownsville– UC Berkeley– U Wisconsin, Milwaukee– UC San Diego– SDSC– Lawrence Berkeley Lab– Argonne– Fermilab– Brookhaven

Funded by GriPhyN

Page 26: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

2629 Jan 2003Paul Avery, University of Florida [email protected]

GriPhyN Project Challenges

• We balance and coordinate– CS research

with “goals, milestones & deliverables”

– GriPhyN schedule/priorities/riskswith those of the 4 experiments

– General tools developed by GriPhyNwith specific tools developed by 4 experiments

– Data Grid design, architecture & deliverableswith those of other Grid projects

• Appropriate balance requires– Tight management, close coordination, trust

• We have (so far) met these challenges– But requires constant attention, good will

Page 27: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

2729 Jan 2003Paul Avery, University of Florida [email protected]

External Advisory Committee

Physics Experiments

Project DirectorsPaul AveryIan Foster

Inte

rne

t 2

DO

E

Sc

ien

ce

NS

F P

AC

Is

Project CoordinationMike Wilde

Rick Cavanaugh

Outreach/EducationManuela Campanelli

Industrial Connections

Ian Foster / Paul Avery

EDG, LCG,Other Grid Projects

ArchitectureCarl Kesselman

VDT DevelopmentCoord.: M. Livny

Requirements, Definition & Scheduling(Miron Livny)

Integration, Testing, Documentation, Support

(Alain Roy)

Globus Project & NMI Integration

(Carl Kesselman)

CS ResearchCoord.: I. Foster

Virtual Data(Mike Wilde)

Request Planning & Scheduling

(Ewa Deelman)

Execution Management(Miron Livny)

Measurement, Monitoring & Prediction

(Valerie Taylor)

ApplicationsCoord.: R. Cavanaugh

ATLAS(Rob Gardner)

CMS(Rick Cavanaugh)

LIGO(Albert Lazzarini)

SDSS(Alexander Szalay)

Inter-Project Coordination:

R. Pordes

HICB(Larry Price)

HIJTB(Carl Kesselman)

PPDG(Ruth Pordes)

TeraGrid, NMI, etc.(TBD)

International (EDG, etc)(Ruth Pordes)

GriPhyNManagement

iVDGL

iVDGLRob Gardner

Page 28: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

2829 Jan 2003Paul Avery, University of Florida [email protected]

External Advisory Committee• Members

– Fran Berman (SDSC Director)– Dan Reed (NCSA Director)– Joel Butler (former head, FNAL Computing Division)– Jim Gray (Microsoft)– Bill Johnston (LBNL, DOE Science Grid)– Fabrizio Gagliardi (CERN, EDG Director)– David Williams (former head, CERN IT)– Paul Messina (former CACR Director)– Roscoe Giles (Boston U, NPACI-EOT)

• Met with us 3 times: 4/2001, 1/2002, 1/2003– Extremely useful guidance on project scope & goals

Page 29: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

2929 Jan 2003Paul Avery, University of Florida [email protected]

Integration of GriPhyN and iVDGL• International Virtual-Data Grid Laboratory

– A global Grid laboratory (US, EU, Asia, …)– A place to conduct Data Grid tests “at scale”– A mechanism to create common Grid infrastructure– A laboratory for Grid tests by other disciplines

• Tight integration with GriPhyN– Testbeds– VDT support– Outreach– Common External Advisory Committee

• International participation– DataTag (EU)– UK e-Science programme: support 6 CS Fellows

Page 30: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

3029 Jan 2003Paul Avery, University of Florida [email protected]

iVDGL Management & Coordination

Project Coordination Group

US External Advisory Committee

GLUE Interoperability Team

Collaborating Grid Projects

TeraGrid

EDG Asia

DataTAG

BTEV

LCG?

BioALICE Geo

?

D0 PDC CMS HI ?

US ProjectDirectors

Outreach Team

Core Software Team

Facilities Team

Operations Team

Applications Team

International Piece

US Project Steering Group

U.S. Piece

GriPhyN Mike Wilde

Page 31: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

3129 Jan 2003Paul Avery, University of Florida [email protected]

Meetings in 2000-2001•GriPhyN/iVDGL meetings

– Oct. 2000 All-hands Chicago– Dec. 2000 Architecture Chicago– Apr. 2001 All-hands, EAC USC/ISI– Aug. 2001 Planning Chicago– Oct. 2001 All-hands, iVDGL USC/ISI

•Numerous smaller meetings– CS-experiment– CS research– Liaisons with PPDG and EU DataGrid– US-CMS and US-ATLAS computing reviews– Experiment meetings at CERN

Page 32: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

3229 Jan 2003Paul Avery, University of Florida [email protected]

Meetings in 2002

• GriPhyN/iVDGL meetings– Jan. 2002 EAC, Planning, iVDGL Florida– Mar. 2002 Outreach Workshop

Brownsville– Apr. 2002 All-hands Argonne– Jul. 2002 Reliability Workshop ISI– Oct. 2002 Provenance Workshop Argonne– Dec. 2002 Troubleshooting Workshop Chicago– Dec. 2002 All-hands technical ISI +

Caltech– Jan. 2003 EAC SDSC

• Numerous other 2002 meetings– iVDGL facilities workshop (BNL)– Grid activities at CMS, ATLAS meetings– Several computing reviews for US-CMS, US-ATLAS– Demos at IST2002, SC2002– Meetings with LCG (LHC Computing Grid) project– HEP coordination meetings (HICB)

Page 33: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

3329 Jan 2003Paul Avery, University of Florida [email protected]

Global Context: Data Grid Projects

• U.S. Infrastructure Projects– GriPhyN (NSF)– iVDGL (NSF)– Particle Physics Data Grid (DOE)– TeraGrid (NSF)– DOE Science Grid (DOE)

• EU, Asia major projects– European Data Grid (EDG) (EU, EC)– EDG related national Projects (UK, Italy, France, …)– CrossGrid (EU, EC)– DataTAG (EU, EC)– LHC Computing Grid (LCG) (CERN)– Japanese Project– Korea project

Page 34: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

3429 Jan 2003Paul Avery, University of Florida [email protected]

U.S. Project Coordination: Trillium• Trillium = GriPhyN + iVDGL + PPDG

– Large overlap in leadership, people, experiments

• Benefit of coordination– Common S/W base + packaging: VDT + PACMAN– Low overhead for collaborative or joint projects:

security, monitoring, newsletter, prod. grids, demos– Wide deployment of new technologies, e.g. Virtual Data– Stronger, more extensive outreach effort

• Forum for US Grid projects– Joint view, strategies, meetings and work– Unified entity to deal with EU & other Grid projects

• “Natural” collaboration across DOE and NSF projects– Funding agency interest?

Page 35: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

3529 Jan 2003Paul Avery, University of Florida [email protected]

International Grid Coordination• Close collaboration with EU DataGrid (EDG)

– Many connections with EDG activities

• HICB: HEP Inter-Grid Coordination Board– Non-competitive forum, strategic issues, consensus– Cross-project policies, procedures and technology– International joint projects

• HICB-JTB Joint Technical Board– Definition, oversight and tracking of joint projects– GLUE interoperability group

• Participation in LHC Computing Grid (LCG)– Software Computing Committee (SC2)– Project Execution Board (PEB)– Grid Deployment Board (GDB)

Page 36: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

3629 Jan 2003Paul Avery, University of Florida [email protected]

Creation of WorldGrid• Joint US-EU Grid deployment

• GriPhyN contribution: VDT– WorldGrid is major driver for VDT– Demonstrated at IST2002 (Copenhagen)– Demonstrated at SC2002 (Baltimore)

• Becoming major outreach tool in 2003– Meeting in February to continue development

Page 37: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

3729 Jan 2003Paul Avery, University of Florida [email protected]

WorldGrid Sites

Page 38: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

3829 Jan 2003Paul Avery, University of Florida [email protected]

What Coordination TakesSergio Andreozzi DataTAG Schema, Glue testbed Carl

Kesselman iVDGL, Globus

Schema

Olof Baring EDG WP4 Schema, Information providers

Peter Kunszt

EDG WP2 Schema, Data Movement and Replication

Rick Cavanaugh GriPhyN, iVDGL

Applications Doug Olson PPDG Authentication, Authorization

Roberto Cecchini EDG, DataTAG

Authentication, Authorization

Ruth Pordes PPDG, iVDGL

Testbeds, Applications

Vincenzo Ciaschini

DataTAG Glue testbed, job submission

David Rebatto

DataTAG Applications

Ben Clifford iVDGL, Globus

MDS development Alain Roy iVDGL, Condor

Virtual Data Toolkit packaging, support.

Ewa Deelman iVDGL, Globus

Schema, VO Operations

Dane Skow PPDG Authentication, Authorization

Luca Dell’Agnello

DataTAG Authentication, Authorization

Scott Gose iVDGL, Globus

Testbed operations, Glue validation tests

Alan DeSmet PPDG, Condor

Applications Massimo Sgaravatto

EDG WP1 Schema, Job Scheduling

Flavia Donno EDG, DataTAG, LCG

Applications, Job Submission, data Movement

Jenny Schopf

PPDG, iVDGL, Globus

Schema, Monitoring

Sergio Fantinel DataTAG Applications Arie Shoshani

PPDG, LBNL

Storage Interface (SRM)

Enrico Ferro DataTAG Distribution, Applications

Fabio Spataro

DataTAG Authentication, Authorization

Rob Gardner iVDGL Applications, Testbed Regina Tam EDG WP5 Schema Jerry Gieraltowski

PPDG Applications Brian Tierney

PPDG, LBNL

Schema, Monitoring

John Gordon EDG WP5 Storage Schema and Services

Luca Vaccarossa

DataTAG Applications

David Groep EDG Authorization Cristina Vistoli

DataTAG Schema, Coordination

Leigh Grunhoefer iVDGL Authentication, Testbed

Saul Youssef

iVDGL Software Distribution, Applications

Page 39: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

3929 Jan 2003Paul Avery, University of Florida [email protected]

Extending GriPhyN’s Reach• Dynamic workspaces proposal

– Expansion of virtual data technologies to global analysis communities

• FIU: Creation of “CHEPREO” in Miami area– HEP research, participation in WorldGrid– Strong minority E/O, coordinate with GriPhyN/iVDGL– Research & int’l network: Brazil / South America

• Also, MRI, SciDAC, other proposals

Page 40: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

4029 Jan 2003Paul Avery, University of Florida [email protected]

Summary• CS research

– Unified approach based around Virtual Data– Virtual Data, Planning, Execution, Monitoring

• Education/Outreach– Student & postdoc involvement at all levels– New collaborations with other E/O efforts, WorldGrid

• Organization and management– Clear management coordinating CS + experiments– Collaboration/coordination US and international

• Research dissemination, broad impact– Wide deployment of VDT (US, WorldGrid, EDG, LCG)– Demo projects, experiment testbeds, major productions– New projects extending virtual data technologies

Page 41: GriPhyN Project Overview Paul Avery University of Florida avery@phys.ufl.edu GriPhyN NSF Project Review 29-30 January 2003 Chicago

4129 Jan 2003Paul Avery, University of Florida [email protected]

Grid References• Grid Book

– www.mkp.com/grids

• GriPhyN– www.griphyn.org

• iVDGL– www.ivdgl.org

• PPDG– www.ppdg.net

• TeraGrid– www.teragrid.org

• Globus– www.globus.org

• Global Grid Forum– www.gridforum.org

• EU DataGrid– www.eu-datagrid.org