Upload
caroline-adkins
View
213
Download
0
Tags:
Embed Size (px)
Citation preview
S.L.Lloyd ATSE e-Science Visit April 2004 Slide 1
GridPP – A UK Computing Grid for Particle Physics
GridPP19 UK Universities, CCLRC
(RAL & Daresbury) and CERN
Funded by the Particle Physics and Astronomy Research Council (PPARC)
GridPP1 - 2001-2004 £17m "From Web to Grid"
GridPP2 - 2004-2007 £15m "From Prototype to Production"www.gridpp.ac.
uk
S.L.Lloyd ATSE e-Science Visit April 2004 Slide 2
Outline
• Why? – The CERN LHC and the Data Deluge• What? - GridPP and the Grid
– Applications– Middleware– Tier-1 and Tier2 Regional Centres– Future Challenges
• How? - GridPP Management• Summary
S.L.Lloyd ATSE e-Science Visit April 2004 Slide 3
The CERN Large Hadron Collider
4 Large Experiments
The world’s most powerful particle accelerator - 2007
S.L.Lloyd ATSE e-Science Visit April 2004 Slide 4
The LHC Experiments
• > 108 electronic channels• 8x108 proton-proton interactions/sec• 2x10-4 Higgs per sec• 10 Petabytes of data a year • (10 Million GBytes = 14 Million CDs –
twice the height of Everest)
Searching for the Higgs Particle and exciting new Physics
Starting from this event
Looking for this ‘signature’
e.g. ATLAS
S.L.Lloyd ATSE e-Science Visit April 2004 Slide 5
What is GridPP?
A team that have built a working prototype Grid – hundreds of computers across 16 UK
institutions– Allows scientists to access data and
processing power seamlessly, wherever they are
– No need for accounts at many different universities
– No need for lots of passwords– ‘Middleware’ detects where the computing
power and data is available and deals with security
Working with other disciplines – e.g. UK e-Science (Grid deployment)
medical researchers (Grid security) – BT (Grid Info. Services, networking)
S.L.Lloyd ATSE e-Science Visit April 2004 Slide 6
GridPP International Collaboration
• CERN IT – PPARC funded £5.7m for staff and hardware– prototype LHC Computing Grid went live
in September 2003 in 12 countries– UK sites among the first to join– Grid Operations Centre based at RAL, UK
• EU DataGrid – PPARC one of 6 main partners– GridPP software central to the project– testbed at more than 40 sites across
Europe with 1,000+ computer processors– around 100,000 jobs successfully
submitted via the UK to the prototype EU-wide grid
– 3 main areas• high energy physics• biology and medical image processing• Earth observation
S.L.Lloyd ATSE e-Science Visit April 2004 Slide 7
What is a Grid?
III. Grid Middlewaree.g. Information Services
I. Experiment Layere.g. Portals
II. Application Middlewaree.g. Metadata
IV. Facilities and Fabricse.g. Storage Services
"Hour Glass"
S.L.Lloyd ATSE e-Science Visit April 2004 Slide 8
GridPP Work Areas
6/Feb/2004
£3.57m
£5.67m
£3.74m
£2.08m£1.84m
CERN
DataGrid
Tier - 1/A
ApplicationsOperations
LHC Computing Grid Project (LCG)Applications, Fabrics, Technology and Deployment
European DataGrid (EDG)Middleware Development
UK Tier-1/A Regional CentreHardware and Manpower
Grid Application DevelopmentLHC and US Experiments + Lattice QCD
Management Travel etc
S.L.Lloyd ATSE e-Science Visit April 2004 Slide 9
Application Development
Fabric
TapeStorage
Elements
RequestFormulator and
Planner
Client Applications
ComputeElements
Indicates component that w ill be replaced
DiskStorage
Elements
LANs andWANs
Resource andServices Catalog
ReplicaCatalog
Meta-dataCatalog
Authentication and SecurityGSISAM-specific user, group , node, st at ion regis tration B bftp ‘cookie’
Connectivity and Resource
CORBA UDP File transfer protocol s - ftp, b bftp, rcp GridFTP
Mass Storage s ystems protocol se.g. encp, hp ss
Collective Services
C atalogproto co ls
Signi fi cant Event Log ger Naming Service Database ManagerC atalog Manager
SAM R es ource M an ag em entB atch Sys tems - LSF, FB S, PB S,
C ondorData Mov erJob Services
Storage ManagerJob ManagerCache ManagerRequest Manager
“Dataset Editor” “File Storage Server”“Project Master” “Station M aster” “Station M aster”
Web Python codes, Java codesCom mand line D0 Fram ework C++ codes
“Stager”“Optim iser”
CodeRepostory
Name in “quotes” is SAM-given software component name
or addedenhanced using PPDG and Grid tools
GANGA
SAMGridLattice QCD
AliEn → ARDA
CMS
BaBar
S.L.Lloyd ATSE e-Science Visit April 2004 Slide 10
Middleware Development
Middleware is the Operating System of a distributed computing system
Configuration Management
Storage Interfaces
Network Monitoring
Security
Information Services
Grid Data Management
S.L.Lloyd ATSE e-Science Visit April 2004 Slide 11
DataGrid Releases
Efficiency (Successful Jobs / Jobs submitted)
0%
10%
20%
30%
40%
50%
60%
70%
80%
90%
100%
De
c-0
2
Jan
-03
Fe
b-0
3
Ma
r-0
3
Ap
r-0
3
Ma
y-0
3
Jun
-03
Jul-
03
Au
g-0
3
Se
p-0
3
Oct
-03
No
v-0
3
De
c-0
3
Jan
-04
Fe
b-0
4Su
cc
es
sfu
l Jo
bs
/ J
ob
s s
ub
mit
ed
CMS EDGv1.4 Altlas EDGv1.4 LHCb EDGv1.4 LCG1 (EDG v2.0) EDG appl. TB v2.x
S.L.Lloyd ATSE e-Science Visit April 2004 Slide 12
Real Time Grid Map
LCG215
April 2004
S.L.Lloyd ATSE e-Science Visit April 2004 Slide 13
The UK Tier-1/A Centre
• High quality data services• National and International Role• UK focus for International Grid
development
LHCb
ATLAS
CMS
BaBar
April 2004:•700 Dual CPU•80TB Disk•60 TB Tape (Capacity 1PB)
Grid Operations Centre
S.L.Lloyd ATSE e-Science Visit April 2004 Slide 14
UK Tier-2 Centres
ScotGridDurham, Edinburgh, Glasgow NorthGridDaresbury, Lancaster, Liverpool,Manchester, Sheffield
SouthGridBirmingham, Bristol, Cambridge,Oxford, RAL PPD, Warwick
LondonGridBrunel, Imperial, QMUL, RHUL, UCL
S.L.Lloyd ATSE e-Science Visit April 2004 Slide 15
Year 2004 2005 2006 2007
CPU [kSI2000] 2395 4066 6380 9965
Disk [TB] 369 735 1424 2285
Tape [TB] 376 752 1542 2623
The Challenges AheadCPU
0
2000
4000
6000
8000
10000
12000
2004 2005 2006 2007
Year
kS
I20
00
ye
ar
ATLAS
CMS
LHCb
ALICE
Phenomenology
ZEUS
UKQCD
UKDMC
MINOS
MICE
LISA
D0
CDF
BaBar
ANTARES
LHC
NonLHC
Disk
0
500
1000
1500
2000
2500
2004 2005 2006 2007
Year
TB
ATLASCMSLHCbALICEPhenomenologyUKQCDUKDMCMINOSMICED0CRESSTCDFBaBarANTARES
LHC
NonLHC
Total Requirement: In International
Context -Q2 2004 LCGResources:
CPU Disk
S.L.Lloyd ATSE e-Science Visit April 2004 Slide 16
UK Core e-Science
Programme
Institutes
Tier-2 Centres
CERNLCG
EGEE
GridPP
GridPP in Context
Tier-1/A
Middleware, Security,
Networking
Experiments
GridSupportCentre
Not to scale!
Apps Dev
AppsInt
GridPP
S.L.Lloyd ATSE e-Science Visit April 2004 Slide 17
GridPP Management
Collaboration Board
Project ManagementBoard
Project Leader
Project Manager
Technical (Deploymen
t)Board
Experiments (User)Board
(Production Manager)
Dissemination Officer
EDG (EGEE) Leader
(Tier-1 Board)
GridPP1 (GridPP2)
Project Map
Risk Register
(Tier-2 Board)
CERN LCG Liaison
S.L.Lloyd ATSE e-Science Visit April 2004 Slide 18
Summary
BaBar
D0CDF
ATLAS
CMS
LHCb
ALICE
19 UK Institutes
RAL Computer Centre
CERN ComputerCentre
SAMGrid
BaBarGrid
LCG
EDGGANGA
EGEE
UK PrototypeTier-1/A Centre
CERN PrototypeTier-0 Centre
4 UK Tier-2 Centres
LCG
UK Tier-1/ACentre
CERN Tier-0Centre
200720042001
4 UK Prototype Tier-2 Centres
ARDA
Separate Experiments,
Resources, Multiple Accounts
'One' Production GridPrototype Grids