1
The Evolution of the LHCb Grid Computing Model S. Roiser on behalf of the LHCb Distributed Computing Team Tier Levels and Activities (original model and now includes unpledged resources 2 if job has no input data if resources available, otherwise concentrated on Tier 2s Tier 0 / 1s Tier 2s 1 Origi nal 2012 Origi nal 2012 Data Reconstruction Y N Y Data Stripping Y N Y Data Merging Y N User Analysis Y Y 2 Monte Carlo Simulation Y 3 Y Working Group Productions - Y - N Data Swimming - Y - Y LHCb Computing Model This poster shows the evolution of the LHCb Computing Model from an originally very rigid setup, adding flexibility over time in terms of data processing and its association to Tier levels. The changes are supported by LHCbDirac [1], the extension of DIRAC, with little additional developments efforts. DISK D1 / T0 TAPE D0 / T1 RAW - 2 1 SDST - 1 2 DST 4 2 3 MC-DST 3 1 3 Data Transfer and Storage Model 1 one copy at CERN, one at another T1 2 uploaded to site of RAW processing 3 archive copies, not accessible by users Data is stored only at CERN and T1 storage elements [4] RAW: data from detector SDST: reconstructed events (MC-)DST: physics streams DIRAC Dirac is the tool used by LHCb for all activities on distributed computing resources on behalf of the experiment. It handles data processing for reconstruction and simulation, physicists analysis activities and data management for file transfer, replication and removal. 2012 Prompt Data Processing • Standard prompt data processing at CERN + Tier 1 sites • Extending the experience of 2011 reprocessing, with offloading new CPU intensive workflow (“Data Swimming”) [2] to one Tier 2 site attached to each storage element New • “Free attachment” of Tier 2 sites to download input data from an available Tier 0/1 storage for data processing • Only for Reconstruction (CPU intensive, lower data I/O) • Tier 2 job needs to upload his output again to the same Tier 0/1 storage element, next input file can be taken from other storage • Increase of network traffic to be monitored closely • Usage of Event Filter Farm [3] for data processing (12k CPU Cores) • Tested so far for Monte Carlo Simulation Possible Future Scenarios for Data Processing Activ CERN CNAF FZK IN2P3 NL-T1 PIC RAL RAW Any T2 Site attac hed to T1 stora ge RAW SDST New Monte Carlo Simulation Simulation of particle interaction with the detector and their subsequent reconstruction up to a simulated analysis MC-DST Working Group Productions Large scale data analysis carried out by production team Data Swimming CPU intensive detector acceptance jobs with little data I/O Data Stripping / Merging The reconstructed events are selected into individual physics streams during stripping. Once a threshold is reached, the files of a given stream are merged into the final analysis DST file, ready for physics analysis. Data Processing Workflows Data Reconstruction The RAW data after processing by the high level trigger is reconstructed into physical quantities corresponding to identified particles and written into a SDST file. User Analysis Unorganized physics analysis jobs by individual users New 2011 New 2012 O r g a n i z e d b y t h e P r o d u c t i o n T e a m [1] LHCbDirac: distributed computing in LHCb, F. Stagni, Poster, Tuesday [2] “Swimming”: a data driven acceptance correction algorithm, M. Cattaneo, Parallel, Tuesday 14 20 -14 45 , Room 905/907 [3] Offline Processing in the Online Computer Farm, L. Granado, Parallel, Monday 17 -18 , E & L 2011 Data Reprocessing • Reprocessing of the 2011 data set in October and November, in parallel to prompt processing • Selected number of T2 sites attached to the “closest” T0/1 storage element, forming a “regional cloud” • 25 % of RAW reconstruction carried out at “attached Tier 2 sites” • Processing finished 2 weeks ahead of schedule New 2011 Prompt Data Processing • Prompt data processing according to initial computing model • Done at CERN and T1 sites • Full processing chain from RAW to DST done at a single site • Subsequent replication of DST file to 3 other T1 storage el • 1 fb -1 of data collected and processed CERN FZK RAL PIC NL- T1 IN2P 3 CNAF RAW RAW D ST D S T 3 more replicas to other storage + 2 archives 1 more replica RAW Reconstr uct SDST StripMer ge DST Every site does Glasgow Manchester Liverpool RAL QMUL SARA NIKHEF DESY FZK Krakow CERN CSCS YANDEX LAL LPC IN2P3 LAPP Legnaro Pisa CNAF CNAF-T2 CPPM Barcelona PIC USC IN2P3-T2

S. Roiser on behalf of the LHCb Distributed Computing Team

Embed Size (px)

DESCRIPTION

Data Processing Workflows. LHCb Computing Model. - PowerPoint PPT Presentation

Citation preview

Page 1: S. Roiser on behalf of the  LHCb  Distributed Computing Team

The Evolution of the LHCb

Grid Computing Model

S. Roiser on behalf of the LHCb Distributed Computing Team

Tier Levels and Activities (original model and now)

1 includes unpledged resources2 if job has no input data3 if resources available, otherwise concentrated on Tier 2s

Tier 0 / 1s Tier 2s1

Original 2012 Original 2012

Data Reconstruction Y N Y

Data Stripping Y N Y

Data Merging Y N

User Analysis Y Y2

Monte Carlo Simulation Y3 Y

Working Group Productions - Y - N

Data Swimming - Y - Y

LHCb Computing Model

This poster shows the evolution of the LHCb Computing Model from an originally very rigid setup, adding flexibility over time in terms of data processing and its association to Tier levels. The changes are supported by LHCbDirac [1], the extension of DIRAC, with little additional developments efforts.

DISKD1 / T0

TAPED0 / T1

RAW - 21

SDST - 12

DST 4 23

MC-DST 3 13

Data Transfer and Storage Model

1 one copy at CERN, one at another T12 uploaded to site of RAW processing3 archive copies, not accessible by users

Data is stored only at CERN and T1 storage elements [4]

RAW: data from detectorSDST: reconstructed events(MC-)DST: physics streams

DIR

AC

Dir

ac i

s th

e to

ol u

sed

by L

HC

b fo

r al

l ac

tiviti

es o

n di

stri

bute

d co

mpu

ting

reso

urce

s on

be

half

of

the

expe

rim

ent.

It

hand

les

data

pro

cess

ing

for

reco

nstr

uctio

n an

d si

mul

atio

n,

phys

icis

ts a

naly

sis

activ

ities

and

dat

a m

anag

emen

t fo

r fi

le tr

ansf

er, r

eplic

atio

n an

d re

mov

al.

2012 Prompt Data Processing

• Standard prompt data processing at CERN + Tier 1 sites • Extending the experience of 2011 reprocessing, with offloading

new CPU intensive workflow (“Data Swimming”) [2] to one Tier 2 site attached to each storage element New

• “Free attachment” of Tier 2 sites to download input data from an available Tier 0/1 storage for data processing• Only for Reconstruction (CPU intensive, lower data I/O)• Tier 2 job needs to upload his output again to the same Tier 0/1

storage element, next input file can be taken from other storage• Increase of network traffic to be monitored closely

• Usage of Event Filter Farm [3] for data processing (12k CPU Cores)• Tested so far for Monte Carlo Simulation

Possible Future Scenarios for Data Processing Activities

CERNCNAFFZK

IN2P3NL-T1

PICRAL

RAW

AnyT2 Site

attachedtoT1

storage

RAW

SDST

New

Monte Carlo SimulationSimulation of particle interaction with the detector and their subsequent reconstruction up to a simulated analysis MC-DST

Working Group ProductionsLarge scale data analysis carried out by production team

Data SwimmingCPU intensive detector acceptance jobs with little data I/O

Data Stripping / MergingThe reconstructed eventsare selected into individual physics streams during stripping. Once a threshold is reached, the files of a given stream are merged into the final analysis DST file, ready for physics analysis.

Data Processing Workflows

Data Reconstruction The RAW data after processing by the high level trigger is reconstructed into physical quantities corresponding to identified particles and written into a SDST file.

User AnalysisUnorganized physics analysis jobs by individual users

New 2011

New 2012

O r

g a

n i

z e

d

b

y

t

h e

P r

o d

u c

t i

o n

T e

a m

[1] LHCbDirac: distributed computing in LHCb, F. Stagni, Poster, Tuesday

[2] “Swimming”: a data driven acceptance correction algorithm, M. Cattaneo, Parallel, Tuesday 1420-1445, Room 905/907

[3] Offline Processing in the Online Computer Farm, L. Granado, Parallel, Monday 1750-1815, E & L Auditorium

[4] The LHCb Data Management System, P. Charpentier, Parallel, Monday 1445-1510, E & L Auditorium

2011 Data Reprocessing

• Reprocessing of the 2011 data set in October and November, in parallel to prompt processing

• Selected number of T2 sites attached to the “closest” T0/1 storage element, forming a “regional cloud”

• 25 % of RAW reconstruction carried out at “attached Tier 2 sites”• Processing finished 2 weeks ahead of schedule

New

2011 Prompt Data Processing

• Prompt data processing according to initial computing model • Done at CERN and T1 sites

• Full processing chain from RAW to DST done at a single site• Subsequent replication of DST file to 3 other T1 storage elements• 1 fb-1 of data collected and processed

CERN

FZK RAL

PIC

NL-T1 IN2P3

CNAF

RA

W

RAW

DST

DS

T

3 more replicas to other storage + 2 archives

1 morereplica

RAW

Reconstruct

SDST

StripMerge

DST

Every site does

Glasgow

Manchester

Liverpool

RAL QMUL SARA

NIKHEF

DESY

FZK

Krakow

CERN

CSCS

YANDEX

LAL

LPC

IN2P3

LAPPLegnaro

Pisa

CNAFCNAF-T2

CPPM

BarcelonaPIC

USC IN2P3-T

2