Upload
crystal-hudson
View
215
Download
0
Tags:
Embed Size (px)
Citation preview
GridPPBuilding a UK Computing Grid for Particle Physics
Professor Steve Lloyd, Queen Mary, University of London
Chair of the GridPP Collaboration Board
Steve Lloyd Public Service Summit - 22 September 2004 Slide 2
Outline
• Why? – The CERN LHC and the Data Deluge• What? - GridPP and the Grid
– What is the Grid?– Applications and Middleware– Tier-1 and Tier-2 Regional Centres
• How? - GridPP Management• Summary and Challenges
Steve Lloyd Public Service Summit - 22 September 2004 Slide 3
What is GridPP?
19 UK Universities, CCLRC (RAL & Daresbury) and CERN
Funded by the Particle Physics and Astronomy Research Council (PPARC)
GridPP1 - 2001-2004 £17m "From Web to Grid"
GridPP2 - 2004-2007 £15m "From Prototype to Production"
Steve Lloyd Public Service Summit - 22 September 2004 Slide 4
The CERN LHC
4 Large Experiments
The world’s most powerful particle accelerator - 2007
Steve Lloyd Public Service Summit - 22 September 2004 Slide 5
LHC Experiments
• > 108 electronic channels• 8x108 proton-proton collisions/sec• 2x10-4 Higgs per sec• 10 Petabytes of data a year • (10 Million GBytes = 14 Million CDs)
Searching for the Higgs Particle and exciting new Physics
Starting from this event
Looking for this ‘signature’
e.g. ATLAS
Steve Lloyd Public Service Summit - 22 September 2004 Slide 6
What is the Grid?
MIDDLEWARE
CPUDisks, CPU etc
PROGRAMS
OPERATING SYSTEM
Word/Excel
Email/Web
Your Progra
mGames
CPUCluste
r
UserInterfac
eMachine
CPUCluste
r
CPUCluste
r
Resource Broker
Information Service
Single PC
Grid
DiskCluste
r
Your Progra
m
Middleware is the Operating System of a distributed computing system
Steve Lloyd Public Service Summit - 22 September 2004 Slide 7
What is the Grid?
From this:
To this:
Steve Lloyd Public Service Summit - 22 September 2004 Slide 8
International Collaboration
• EU DataGrid (EDG) 2001-2004– Middleware Development
Project
• US and other Grid projects– Interoperability
• LHC Computing Grid (LCG)– Grid Deployment Project for
LHC • EU Enabling Grids for e-Science
in Europe (EGEE) 2004-2006– Grid Deployment Project for all
disciplines
Steve Lloyd Public Service Summit - 22 September 2004 Slide 9
The LCG Grid
Steve Lloyd Public Service Summit - 22 September 2004 Slide 10
Grid Snapshot
Steve Lloyd Public Service Summit - 22 September 2004 Slide 11
GridPP1 Areas
6/Feb/2004
£3.57m
£5.67m
£3.74m
£2.08m£1.84m
CERN
DataGrid
Tier - 1/A
ApplicationsOperations
LHC Computing Grid Project (LCG)Applications, Fabrics, Technology and Deployment
European DataGrid (EDG)Middleware Development
UK Tier-1/A Regional CentreHardware and Manpower
Grid Application DevelopmentLHC and US Experiments + Lattice QCD
Management Travel etc
Steve Lloyd Public Service Summit - 22 September 2004 Slide 12
GridPP2 Areas
July 2004
£0.75m
£2.71m
£2.88m
£0.87m
£0.69m
£2.71m
£2.89m
£1.00m
£2.40m
Tier-1/AHardware
Tier-2Operations
Applications
M/S/N
LCG-2
MgrTravel
Ops
Tier-1/AOperations
Management Travel etc
UK Tier-1/A Regional Centre Hardware
UK Tier-2 Regional Centres Manpower
UK Tier-1/A Manpower
LHC Computing Grid Project (LCG)Manpower
Middleware, Security and Networking Manpower
Grid Application Development LHC and US Experiments + Lattice QCD, Phenomenology and Generic Portal
Steve Lloyd Public Service Summit - 22 September 2004 Slide 13
Application Development
Fabric
TapeStorage
Elements
RequestFormulator and
Planner
Client Applications
ComputeElements
Indicates component that w ill be replaced
DiskStorage
Elements
LANs andWANs
Resource andServices Catalog
ReplicaCatalog
Meta-dataCatalog
Authentication and SecurityGSISAM-specific user, group , node, st at ion regis tration B bftp ‘cookie’
Connectivity and Resource
CORBA UDP File transfer protocol s - ftp, b bftp, rcp GridFTP
Mass Storage s ystems protocol se.g. encp, hp ss
Collective Services
C atalogproto co ls
Signi fi cant Event Log ger Naming Service Database ManagerC atalog Manager
SAM R es ource M an ag em entB atch Sys tems - LSF, FB S, PB S,
C ondorData Mov erJob Services
Storage ManagerJob ManagerCache ManagerRequest Manager
“Dataset Editor” “File Storage Server”“Project Master” “Station M aster” “Station M aster”
Web Python codes, Java codesCom mand line D0 Fram ework C++ codes
“Stager”“Optim iser”
CodeRepostory
Name in “quotes” is SAM-given software component name
or addedenhanced using PPDG and Grid tools
GANGA
SAMGridLattice QCD
AliEn → ARDA
CMS
BaBar
Steve Lloyd Public Service Summit - 22 September 2004 Slide 14
Middleware Development
Configuration Management
Storage Interfaces
Network Monitoring
Security
Information Services
Grid Data Management
Steve Lloyd Public Service Summit - 22 September 2004 Slide 15
UK Tier-1/A Centre
• High quality data services• National and International
Role• UK focus for International
Grid development
•700 Dual CPU•80 TB Disk•60 TB Tape (Capacity 1PB)
Grid Operations Centre
Steve Lloyd Public Service Summit - 22 September 2004 Slide 16
UK Tier-2 Centres
ScotGridDurham, Edinburgh, Glasgow NorthGridDaresbury, Lancaster, Liverpool,Manchester, Sheffield
SouthGridBirmingham, Bristol, Cambridge,Oxford, RAL PPD, Warwick
LondonGridBrunel, Imperial, QMUL, RHUL, UCL
Mostly funded by HEFCE
Steve Lloyd Public Service Summit - 22 September 2004 Slide 17
UK Core e-Science
Programme
Institutes
Tier-2 Centres
CERNLCG
EGEE
GridPP
GridPP in Context
Tier-1/A
Middleware, Security,
Networking
Experiments
GridSupportCentre
Not to scale!
Apps Dev
AppsInt
GridPP
Steve Lloyd Public Service Summit - 22 September 2004 Slide 18
Management
Collaboration Board
Project ManagementBoard
Project Leader
Project Manager
DeploymentBoard
UserBoard
Production Manager
Dissemination Officer
EGEE Leader
Tier-1 Board
Project Map
Risk Register
Tier-2 Board
CERN LCG Liaison
Deployment Team
Steve Lloyd Public Service Summit - 22 September 2004 Slide 19
Summary
BaBar
D0CDF
ATLAS
CMS
LHCb
ALICE
19 UK Institutes
RAL Computer Centre
CERN ComputerCentre
SAMGrid
BaBarGrid
LCG
EDGGANGA
EGEE
UK PrototypeTier-1/A Centre
CERN PrototypeTier-0 Centre
4 UK Tier-2 Centres
LCG
UK Tier-1/ACentre
CERN Tier-0Centre
200720042001
4 UK Prototype Tier-2 Centres
ARDA
Separate Experiments,
Resources, Multiple Accounts
'One' Production GridPrototype Grids
Steve Lloyd Public Service Summit - 22 September 2004 Slide 20
Challenges
Concorde(15 km)
CD stack with1 year LHC data(~ 20 km)
We are here(1 km)
• Scaling to full size ~10,000 → 100,000 CPUs
• Stability, Robustness etc• Security• Sharing resources (in RAE
environment!)• International Collaboration• Continued funding beyond
start of LHC!
Steve Lloyd Public Service Summit - 22 September 2004 Slide 21
Further Info
http://www.gridpp.ac.uk