Upload
others
View
1
Download
0
Embed Size (px)
Citation preview
A common connection: CMS (LHC@CERN) Computing
Kajari Mazumdar
Department of High Energy Physics
Tata Institute of Fundamental Research
Mumbai
Accelerating International Collaboration and Science through Connective Computation University of Chicago Centre, Delhi March 10, 2015
• Raison de etre of LHC: discover/rule out existence of Higgs boson particle (if it existed any time in nature?) • Yes, indeed! about 1 pico-second after the Big Bang.
• Today, after 13.7 billion years later, CERN-LHC recreates the condition of the very early universe. • Higgs boson discovered within 3 years of start of data taking at LHC. • Experiments study the aftermath of violent collisions of protons/ions -- using very complicated, mammoth detectors -- with ~ 1 million electronic channels taking data every 25 nano-sec. digital summary of information recorded as collision event. • 300 publications per experiment with data collected in Run1 (~4 years)
• LHC Computing Grid is the backbone of the success of LHC project.
Introduction
March 10, 2015 K. Mazumdar 2
Higgs boson discovery at CERN-LHC in 2012
March 10, 2015 K. Mazumdar 3
CERN-LHC 2009-2035: Physics exploitation Nobel prize in 2013, …..?
What happens in LHC experiment
March 10, 2015
K. Mazumdar 4
Selection only 1 in 1013 events. we do not know which one! need to check with all! Enormous computing resources needed.
In hard numbers: • 2 big experiments: ATLAS & CMS. • LHC collides 6-8 hundred million
protons/second for several years. • Digital information of each collision
: 1 – 10 MB • Capacity for Information storage: ~ 1000 collisions /sec. • Physicists must sift through ~ 20
Petabytes (1015 ) of data annually.
March 10, 2015 K. Mazumdar 5
Higgs search: ~ looking for a
particular person in a thousand
world populations of today.
Large Collaborative Research
• Distributed analysis/computing is essential and designed accordingly.
• Made possible by sharing resources across the globe:
Crosses organisational, national and international boundaries.
• High Energy Physics (HEP) is often compute as well as data intensive.
• Experiments performed by several thousand scientists over several years.
• CERN-LHC project is an excellent example of large scale collaboration,
dealing with huge amount of data.
• WWW was born in early 1989 to satisfy the needs of previous
generation of high energy experiments carried out by scientists across
the world : sharing information.
Becoming necessary in various other fields of research.
Fine example of application in societal benefit from fundamental research.
March 10, 2015 K. Mazumdar 6
1. Share more than information.
2. Provide resources and services to store/serve ~ 20 PB data/year
3. Provide access to all interesting physics events to ~ 4000
collaborators/expt.
Solution through WLCG : Efficient use of resources at many institutes.
• Minimize constraints due to user localisation and resource variety
• Decentralize control and costs of computing infrastructure
much faster delivery of physics!
From Web to Grid Computing
March 10, 2015 K. Mazumdar 7
World-wide LHC Computing GRID (conceptualized in ~2000)
natural evolution of internet technology in data analysis.
March 10, 2015 K. Mazumdar 8
LHC Computing Grid • Key for grid computing:
LAN speed is comparable to the speed of computing processors.
• Success due to high speed link across countries. India also pays dividends.
• Contributes to few percent level to the computing efforrts of CMS experiment
• Typical numbers in LHC for each experiment:
# grid WLCG users : ~1000/week, 1M jobs/week
# darasets available for analysis across grid: 50k
• Today ~ 200 sites across the world active 24X7 for LHC experiments.
• ~100 PB disk
• Increasing bandwidth fast evolution of LHC computing model in terms of
management and processing of data volume
From hierarchical structure to parallel/horizontal connections.
opportunistic usage of resources
Tier 0
Tier 1 National centres
Tier 2 Regional groups in
a continent/nation
Different Universities,
Institutes in a country
Individual scientist’s
PC,laptop, ..
Experimental
site
CERN
computer
centre,
Geneva
ASIA
(Taiwan)
India China Korea Taiwan
France Italy Germany
USA
TIFR
BARC Panjab
Univ.
Indiacms T2_IN_TIFR
Initial structure of LHC GRID connecting computers across globe
Delhi
Univ.
Pakistan
Online data
recording
Several
Petabytes/sec.
10 Gbps
1-2.5 Gbps
CERN
March 10, 2015 K. Mazumdar 9
Example: TIER-3 Center for CMS Experiment @ University of Delhi
Caters to the needs of the High Energy Physics Group of the Physics Department
of the University of Delhi
Allows one to fetch CMS data from Tier-2 centres, process data locally, generate
monte carlo etc. Kirti Ranjan
Tier-3@DU: Hardware Resources
Master Server Specifications: HP DL180 G6 Processor: 2xIntel Xeon Quad Core (64 bit) Memory : 32GB each Hard disk: 146 GB
Worker Node (Total 3 Server) Specifications:HP DL160 G6 Processor: 4xIntel Xeon Quad Core (64 bit) Memory : 16GB each Hard disk: 146 GB Specifications:HP DL160 G8 (New Node Added) Processor: 2xIntel Xeon Quad Core (64 bit) Memory : 16GB each Hard disk: 500 GB One KVM Switch, Console and 1GB Ethernet Switch StorageDisk Space : 12 TB Storage Area
24TB Storage (New Storage Added) Power BackUp: 2*6KV UPS
Cooling : AC: 2* 2 Tons each
A typical situation
Kirti Ranjan
March 10, 2015 K. Mazumdar 12
Tier 0: 1M jobs/day. Peak data-transfer rates: 10 gigabytes per second (2 full DVDs of data/s). Users don’t bother about the source of computing plug your computer in to the Internet, it will get the computer power you need to do the job!
The Grid infrastructure links together computing resources such as PCs, workstations, servers, storage elements, and provides the mechanism needed to access them. The pervasive nature of grid simply access the Grid through web browser. The Grid is a utility: you ask for computer power or storage capacity and you get it.
Middleware is the technical "glue" that allows different computers to "stick" together.
Behind the scene
We have now experts here offering us to make computing from India easy.
March 10, 2015 K. Mazumdar 13
WISDOM : search for a cure for malaria MammoGrid : grid for hospitals to share and analyse mammograms to improve breast cancer treatment. MathCell : grid-managed multi-purpose environment for further research in biology and bioinformatics P12S2: learn more about the spread of plant diseases
Examples for application of Grid technology
AstroGrid: astronomy BIRN: human disease CaBIG: cancer
neuGRID: Neuroscience outGRID: Neuroscience Virolab: infectious disease ...@home: volunteer computing
SETI@home: Extraterrestrial intelligence The Charity Engine World Community Grid: Many different projects all with humanitarian aims
Climateprediction.net: Climate research Compute Against Cancer: Cancer research FightAIDS@Home: HIV/AIDS research Folding@Home: Disease research GridRepublic: Many different research projects
LHC@home: High energy physics GIMPS: Mathematics
FusionGrid: fusion energy NEES : earthquakes
Summary
March 10, 2015 K. Mazumdar 14
• We, from India, are lucky to be part of LHC family.
• Grateful for the support received to place India in LHC Grid Map.
• Time for India to exert leverage from LHC experience.
• Grid computing is changing the way the world is doing science
( as well as business, entertainment, social science and more.)
• Efforts must start now to make it happen in India,
first the academic community.
• Experts are ready to help us in making the leap.
Computing aspect of your research can be made easier/bigger!
March 10, 2015 K. Mazumdar 15
Warm welcome to the participants.
Many thanks to University of Chicago Center, Delhi.
March 10, 2015 K. Mazumdar
ESnet GLOBUS
gLite GRIDBUS
GridSphere IGE
Middleware projects
3G Bridge Alchemi BioGrid Condor DCache DOE SciDAC EMI
ESnet GLOBUS gLite GRIDBUS GridSphere IGE
March 10, 2015 K. Mazumdar 17
LHC Grid Map
CMS and ALICE Tier2 GRID computing
centers in TIFR (Mumbai) and VECC (Kolkata).
~ 200 sites across the world active 24X7
~ 100 PB disk, 300,000 cores