17
A common connection: CMS (LHC@CERN) Computing Kajari Mazumdar Department of High Energy Physics Tata Institute of Fundamental Research Mumbai Accelerating International Collaboration and Science through Connective Computation University of Chicago Centre, Delhi March 10, 2015

A common connection: CMS (LHC@CERN) Computingmazumdar/talks/Uchicago.pdf · GLOBUS gLite GRIDBUS GridSphere IGE Middleware projects 3G Bridge Alchemi BioGrid Condor DCache DOE SciDAC

  • Upload
    others

  • View
    1

  • Download
    0

Embed Size (px)

Citation preview

Page 1: A common connection: CMS (LHC@CERN) Computingmazumdar/talks/Uchicago.pdf · GLOBUS gLite GRIDBUS GridSphere IGE Middleware projects 3G Bridge Alchemi BioGrid Condor DCache DOE SciDAC

A common connection: CMS (LHC@CERN) Computing

Kajari Mazumdar

Department of High Energy Physics

Tata Institute of Fundamental Research

Mumbai

Accelerating International Collaboration and Science through Connective Computation University of Chicago Centre, Delhi March 10, 2015

Page 2: A common connection: CMS (LHC@CERN) Computingmazumdar/talks/Uchicago.pdf · GLOBUS gLite GRIDBUS GridSphere IGE Middleware projects 3G Bridge Alchemi BioGrid Condor DCache DOE SciDAC

• Raison de etre of LHC: discover/rule out existence of Higgs boson particle (if it existed any time in nature?) • Yes, indeed! about 1 pico-second after the Big Bang.

• Today, after 13.7 billion years later, CERN-LHC recreates the condition of the very early universe. • Higgs boson discovered within 3 years of start of data taking at LHC. • Experiments study the aftermath of violent collisions of protons/ions -- using very complicated, mammoth detectors -- with ~ 1 million electronic channels taking data every 25 nano-sec. digital summary of information recorded as collision event. • 300 publications per experiment with data collected in Run1 (~4 years)

• LHC Computing Grid is the backbone of the success of LHC project.

Introduction

March 10, 2015 K. Mazumdar 2

Page 3: A common connection: CMS (LHC@CERN) Computingmazumdar/talks/Uchicago.pdf · GLOBUS gLite GRIDBUS GridSphere IGE Middleware projects 3G Bridge Alchemi BioGrid Condor DCache DOE SciDAC

Higgs boson discovery at CERN-LHC in 2012

March 10, 2015 K. Mazumdar 3

CERN-LHC 2009-2035: Physics exploitation Nobel prize in 2013, …..?

Page 4: A common connection: CMS (LHC@CERN) Computingmazumdar/talks/Uchicago.pdf · GLOBUS gLite GRIDBUS GridSphere IGE Middleware projects 3G Bridge Alchemi BioGrid Condor DCache DOE SciDAC

What happens in LHC experiment

March 10, 2015

K. Mazumdar 4

Selection only 1 in 1013 events. we do not know which one! need to check with all! Enormous computing resources needed.

Page 5: A common connection: CMS (LHC@CERN) Computingmazumdar/talks/Uchicago.pdf · GLOBUS gLite GRIDBUS GridSphere IGE Middleware projects 3G Bridge Alchemi BioGrid Condor DCache DOE SciDAC

In hard numbers: • 2 big experiments: ATLAS & CMS. • LHC collides 6-8 hundred million

protons/second for several years. • Digital information of each collision

: 1 – 10 MB • Capacity for Information storage: ~ 1000 collisions /sec. • Physicists must sift through ~ 20

Petabytes (1015 ) of data annually.

March 10, 2015 K. Mazumdar 5

Higgs search: ~ looking for a

particular person in a thousand

world populations of today.

Page 6: A common connection: CMS (LHC@CERN) Computingmazumdar/talks/Uchicago.pdf · GLOBUS gLite GRIDBUS GridSphere IGE Middleware projects 3G Bridge Alchemi BioGrid Condor DCache DOE SciDAC

Large Collaborative Research

• Distributed analysis/computing is essential and designed accordingly.

• Made possible by sharing resources across the globe:

Crosses organisational, national and international boundaries.

• High Energy Physics (HEP) is often compute as well as data intensive.

• Experiments performed by several thousand scientists over several years.

• CERN-LHC project is an excellent example of large scale collaboration,

dealing with huge amount of data.

• WWW was born in early 1989 to satisfy the needs of previous

generation of high energy experiments carried out by scientists across

the world : sharing information.

Becoming necessary in various other fields of research.

Fine example of application in societal benefit from fundamental research.

March 10, 2015 K. Mazumdar 6

Page 7: A common connection: CMS (LHC@CERN) Computingmazumdar/talks/Uchicago.pdf · GLOBUS gLite GRIDBUS GridSphere IGE Middleware projects 3G Bridge Alchemi BioGrid Condor DCache DOE SciDAC

1. Share more than information.

2. Provide resources and services to store/serve ~ 20 PB data/year

3. Provide access to all interesting physics events to ~ 4000

collaborators/expt.

Solution through WLCG : Efficient use of resources at many institutes.

• Minimize constraints due to user localisation and resource variety

• Decentralize control and costs of computing infrastructure

much faster delivery of physics!

From Web to Grid Computing

March 10, 2015 K. Mazumdar 7

World-wide LHC Computing GRID (conceptualized in ~2000)

natural evolution of internet technology in data analysis.

Page 8: A common connection: CMS (LHC@CERN) Computingmazumdar/talks/Uchicago.pdf · GLOBUS gLite GRIDBUS GridSphere IGE Middleware projects 3G Bridge Alchemi BioGrid Condor DCache DOE SciDAC

March 10, 2015 K. Mazumdar 8

LHC Computing Grid • Key for grid computing:

LAN speed is comparable to the speed of computing processors.

• Success due to high speed link across countries. India also pays dividends.

• Contributes to few percent level to the computing efforrts of CMS experiment

• Typical numbers in LHC for each experiment:

# grid WLCG users : ~1000/week, 1M jobs/week

# darasets available for analysis across grid: 50k

• Today ~ 200 sites across the world active 24X7 for LHC experiments.

• ~100 PB disk

• Increasing bandwidth fast evolution of LHC computing model in terms of

management and processing of data volume

From hierarchical structure to parallel/horizontal connections.

opportunistic usage of resources

Page 9: A common connection: CMS (LHC@CERN) Computingmazumdar/talks/Uchicago.pdf · GLOBUS gLite GRIDBUS GridSphere IGE Middleware projects 3G Bridge Alchemi BioGrid Condor DCache DOE SciDAC

Tier 0

Tier 1 National centres

Tier 2 Regional groups in

a continent/nation

Different Universities,

Institutes in a country

Individual scientist’s

PC,laptop, ..

Experimental

site

CERN

computer

centre,

Geneva

ASIA

(Taiwan)

India China Korea Taiwan

France Italy Germany

USA

TIFR

BARC Panjab

Univ.

Indiacms T2_IN_TIFR

Initial structure of LHC GRID connecting computers across globe

Delhi

Univ.

Pakistan

Online data

recording

Several

Petabytes/sec.

10 Gbps

1-2.5 Gbps

CERN

March 10, 2015 K. Mazumdar 9

Page 10: A common connection: CMS (LHC@CERN) Computingmazumdar/talks/Uchicago.pdf · GLOBUS gLite GRIDBUS GridSphere IGE Middleware projects 3G Bridge Alchemi BioGrid Condor DCache DOE SciDAC

Example: TIER-3 Center for CMS Experiment @ University of Delhi

Caters to the needs of the High Energy Physics Group of the Physics Department

of the University of Delhi

Allows one to fetch CMS data from Tier-2 centres, process data locally, generate

monte carlo etc. Kirti Ranjan

Page 11: A common connection: CMS (LHC@CERN) Computingmazumdar/talks/Uchicago.pdf · GLOBUS gLite GRIDBUS GridSphere IGE Middleware projects 3G Bridge Alchemi BioGrid Condor DCache DOE SciDAC

Tier-3@DU: Hardware Resources

Master Server Specifications: HP DL180 G6 Processor: 2xIntel Xeon Quad Core (64 bit) Memory : 32GB each Hard disk: 146 GB

Worker Node (Total 3 Server) Specifications:HP DL160 G6 Processor: 4xIntel Xeon Quad Core (64 bit) Memory : 16GB each Hard disk: 146 GB Specifications:HP DL160 G8 (New Node Added) Processor: 2xIntel Xeon Quad Core (64 bit) Memory : 16GB each Hard disk: 500 GB One KVM Switch, Console and 1GB Ethernet Switch StorageDisk Space : 12 TB Storage Area

24TB Storage (New Storage Added) Power BackUp: 2*6KV UPS

Cooling : AC: 2* 2 Tons each

A typical situation

Kirti Ranjan

Page 12: A common connection: CMS (LHC@CERN) Computingmazumdar/talks/Uchicago.pdf · GLOBUS gLite GRIDBUS GridSphere IGE Middleware projects 3G Bridge Alchemi BioGrid Condor DCache DOE SciDAC

March 10, 2015 K. Mazumdar 12

Tier 0: 1M jobs/day. Peak data-transfer rates: 10 gigabytes per second (2 full DVDs of data/s). Users don’t bother about the source of computing plug your computer in to the Internet, it will get the computer power you need to do the job!

The Grid infrastructure links together computing resources such as PCs, workstations, servers, storage elements, and provides the mechanism needed to access them. The pervasive nature of grid simply access the Grid through web browser. The Grid is a utility: you ask for computer power or storage capacity and you get it.

Middleware is the technical "glue" that allows different computers to "stick" together.

Behind the scene

We have now experts here offering us to make computing from India easy.

Page 13: A common connection: CMS (LHC@CERN) Computingmazumdar/talks/Uchicago.pdf · GLOBUS gLite GRIDBUS GridSphere IGE Middleware projects 3G Bridge Alchemi BioGrid Condor DCache DOE SciDAC

March 10, 2015 K. Mazumdar 13

WISDOM : search for a cure for malaria MammoGrid : grid for hospitals to share and analyse mammograms to improve breast cancer treatment. MathCell : grid-managed multi-purpose environment for further research in biology and bioinformatics P12S2: learn more about the spread of plant diseases

Examples for application of Grid technology

AstroGrid: astronomy BIRN: human disease CaBIG: cancer

neuGRID: Neuroscience outGRID: Neuroscience Virolab: infectious disease ...@home: volunteer computing

SETI@home: Extraterrestrial intelligence The Charity Engine World Community Grid: Many different projects all with humanitarian aims

Climateprediction.net: Climate research Compute Against Cancer: Cancer research FightAIDS@Home: HIV/AIDS research Folding@Home: Disease research GridRepublic: Many different research projects

LHC@home: High energy physics GIMPS: Mathematics

FusionGrid: fusion energy NEES : earthquakes

Page 14: A common connection: CMS (LHC@CERN) Computingmazumdar/talks/Uchicago.pdf · GLOBUS gLite GRIDBUS GridSphere IGE Middleware projects 3G Bridge Alchemi BioGrid Condor DCache DOE SciDAC

Summary

March 10, 2015 K. Mazumdar 14

• We, from India, are lucky to be part of LHC family.

• Grateful for the support received to place India in LHC Grid Map.

• Time for India to exert leverage from LHC experience.

• Grid computing is changing the way the world is doing science

( as well as business, entertainment, social science and more.)

• Efforts must start now to make it happen in India,

first the academic community.

• Experts are ready to help us in making the leap.

Computing aspect of your research can be made easier/bigger!

Page 15: A common connection: CMS (LHC@CERN) Computingmazumdar/talks/Uchicago.pdf · GLOBUS gLite GRIDBUS GridSphere IGE Middleware projects 3G Bridge Alchemi BioGrid Condor DCache DOE SciDAC

March 10, 2015 K. Mazumdar 15

Warm welcome to the participants.

Many thanks to University of Chicago Center, Delhi.

Page 16: A common connection: CMS (LHC@CERN) Computingmazumdar/talks/Uchicago.pdf · GLOBUS gLite GRIDBUS GridSphere IGE Middleware projects 3G Bridge Alchemi BioGrid Condor DCache DOE SciDAC

March 10, 2015 K. Mazumdar

ESnet GLOBUS

gLite GRIDBUS

GridSphere IGE

Middleware projects

3G Bridge Alchemi BioGrid Condor DCache DOE SciDAC EMI

ESnet GLOBUS gLite GRIDBUS GridSphere IGE

Page 17: A common connection: CMS (LHC@CERN) Computingmazumdar/talks/Uchicago.pdf · GLOBUS gLite GRIDBUS GridSphere IGE Middleware projects 3G Bridge Alchemi BioGrid Condor DCache DOE SciDAC

March 10, 2015 K. Mazumdar 17

LHC Grid Map

CMS and ALICE Tier2 GRID computing

centers in TIFR (Mumbai) and VECC (Kolkata).

~ 200 sites across the world active 24X7

~ 100 PB disk, 300,000 cores