23
Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

Embed Size (px)

Citation preview

Page 1: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

Grid Update

Henry Nebrensky

Brunel University

MICE Collaboration Meeting CM23

Page 2: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

The Awesome Power of Grid Computing

The Grid provides seamless interconnection between thousands of computers.

It therefore generates new acronyms and jargon at

superhuman speed.

Page 3: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

MICE and the Grid (1)

● Computing: 10 UK sites & Sofia supporting MICE● Storage: 9 sites & RAL Tier 1 supporting MICE● Job submission: 2 each at Glasgow and RAL● UK centric!

Even though we don't yet need to ask to ask for more resources, does everyone know whom they should ask and how long the bureaucracy takes?

Page 4: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

MICE and the Grid (2)

● g4Beamline tested on Grid over a year ago● G4MICE in regular use

– ~1000 simulation jobs/day● Job submission: recent RB->WMS change● I will make a bundle of config files for the gLite

CLI available, for people to install their own UI – gLite User Guide:

http://glite.web.cern.ch/glite/documentation/default.asp● Ganga GUI (as used by Atlas) also works for

MICE, see D. Forrest MICE Note soon.

Page 5: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

MICE and the Grid (3)

● Grid interface to CASTOR storage at RAL Tier1 has been deployed (not sure what MICE quota is)

● Space unused – need to test (both interface and storage). We should ensure that the Tier 1 is aware of our activities; check:

– https://www.gridpp.ac.uk/wiki/RAL_Tier1_CASTOR_planning

and route requests via Paul Kyberd. ● Storage elsewhere has been used (e.g. staging of

simulation output) but as yet we are not yet formally storing data on the Grid.

Page 6: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

MICE and the Grid (4)

● We are currently using EGEE/WLCG middleware and resources, as they are receiving significant development effort and are a reasonable match for our needs (shared with various minor experiments such as LHC)

● Outside Europe other software may be expected –

e.g. the OSG stack in the US. Interoperability is

from our perspective a “known unknown”...

Page 7: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

MICE and Grid Data Storage

● The Grid can provide provide MICE not only with computing (number-crunching) power, but also with a secure global framework allowing users access to data

● Good news: storing development data on the Grid

keeps it available to the collaboration – not stuck

on an old PC in the corner of the lab ● Bad news: loss of ownership – who picks up the

data curation responsibilities?

Page 8: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

Grid File Management

● Each file is given a unique, machine-generated, GUID when stored on the Grid

● The file is physically uploaded to one (or more) SEs (Storage Elements) where it is given a machine-generated SURL (Storage URL)

● Machine-generated names are not (meant to be) human-usable

● A “replica catalogue” tracks the multiple SURLs of a GUID

Page 9: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

File Catalogue

● For sanity's sake we would like to associate nice sensible filenames with each file (LFN, Logical File Name)

● A “file catalogue” is basically a database that translates between something that looks like a Unix filesystem and the GUIDs and SURLs needed to actually access the data on the Grid

● MICE has an instance of LFC (LCG File Catalogue) run by the Tier 1 at RAL

● The LFC service can do both the replica and LFN cataloguing

Page 10: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

File Catalogue Namespace

● We need to agree on a consistent namespace for the file catalogue

● The aim is NOT to replicate the experiment structure, instead we want an easy-to-understand structure that is compact but without too many entries in each low-level directory (aim for 20-50)

Page 11: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

Concepts (1)

1) At present it is hard-to-impossible to delete a directory out of the LFC namespace. Avoid excess complexity – prevents creating dead branches

2) Ideally a directory should contain either only

more subdirectories, or only some files

Page 12: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

Concepts (2)

1) Don't assume it will be possible to browse this from a graphical interface with thumbnails – if you have to search through the LFC by hand, it will be painful even with the ideal structure

2) Moving things will cause confusion (though LFC

allows multiple “soft” links)

3) MC simulations close to that which they model

Page 13: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

Namespace (half-baked proposal)

● We get given /grid/mice/ by the server● Four upper-level directories:

– Construction/historical data from detector development and QA

– Calibration/needed during analysis (large datasets, c.f. DB)

– TestBeam/test beam data

– MICE/DAQ output and corresponding MC simulation

Page 14: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

Namespace proposal (1)

/grid/mice/MICE/StepN/ RefMomentum? / MC or data

● Split into MICE Step. ● Do we need more subdivision e.g. by momentum?● Separate directory for MC results, or in filename?

/TestBeam/place/...● KEK, Fermi, RAL (Tracker Cosmics)?● The subdivisions to be place-specific

Page 15: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

Namespace proposal (2)

/grid/mice/Construction/module/data

/Construction/Tracker/TrackerQA/Construction/Tracker/OpticalQA/Construction/Tracker/FibreDamage

● What about the other modules?● Should we bother with raw files, or tarball/zip

each subset up?

Page 16: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

Namespace proposal (3)

/grid/mice/Calibration/Tracker/VLPC/Calibration/BeamLine/FieldMaps

● How should this be split up = what else will be in

here – e.g. are the solenoid field maps part of the

spectrometer/grid/mice/Calibration/Spectrometer/FieldMaps

or part of the beamline/grid/mice/Calibration/BeamLine/FieldMaps/SpectrometerSolenoid

or do we put the field maps together/grid/mice/Calibration/FieldMaps/SpectrometerSolenoid/grid/mice/Calibration/FieldMaps/Quads

Page 17: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

Namespace proposal (3)/grid/mice

/Users/name

● For people to use as scratch space for their own

purposes● Encourage people to do this through LFC – helps

avoid “dark data”● LFC allows Unix-style access permissions

Page 18: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

Namespace

We are starting to deploy this now, so we need a (crude) idea of – what calibration data will be needed– how it will be provided - O(# of files), format, size

I have no ( = zero ) idea of what to expect from most of MICE for this !!!

What info should be represented by the directory structures and what within the filename?

Need to be consistent in use of upper and lower case in LFNs, else will create bogus entries.

Page 19: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

Metadata Catalogue (1)

● For many applications – such as analysis – you will want to identify the list of files containing the data that matches some parameters

● This is done by a “metadata catalogue”.For MICE this doesn't yet exist

● A metadata catalogue can in principle return either the GUID or an LFN

Page 20: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

Metadata Catalogue (2)

● We need to select a technology to use for this– use the conditions database– gLite AMGA (who else uses it – will it remain

supported?)● Need to implement – i.e. register metadata to files

– What metadata will be needed for

analysis?● Should the catalogue include the file format and

compression scheme (gzip ≠ PKzip)?

Page 21: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

Metadata Catalogue (2) for Humans

or, in non-Gridspeak:● we have several databases (conditions DB, EPICS,

e-Logbook) where we should be able to find all sorts of information about a run/timestamp.

● but how do we know which runs to be interested in, for our analysis?

● we need an “index” to the MICE data, and for this we need to define the set of “index terms” that will be used to search for relevant datasets.

Page 22: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

Grid Data Access ProtocolsOur current tools are based around the transfer of

whole files to/from a local disk on the processing machine.

The Grid also allows “POSIX I/O” (i.e. random access) directly to files on the local SE, using a secure version of the RFIO protocol. This would require compiling libshift into G4MICE.

This would be useful in cases where we need to access only a small part of the data in a file.

We currently don't see any need for this in MICE.

Page 23: Grid Update Henry Nebrensky Brunel University MICE Collaboration Meeting CM23

Last Slide

● “My laptop battery is flat” is no longer an excuse for not getting some simulation/analysis done!