44
Computing for high energy neutrino experiments Tommaso Chiarusi INFN Sez. Bologna & Phys. Dep. Univ. Bologna Workshop CCR-INFN GRID 2010 Monday, May 24, 2010

Computing Neutrino Telescope

Embed Size (px)

DESCRIPTION

Presentation to INFN CWS2010

Citation preview

Page 1: Computing Neutrino Telescope

Computing for high energy neutrino experiments

Tommaso Chiarusi INFN Sez. Bologna & Phys. Dep. Univ. Bologna

Workshop CCR-INFN GRID 2010

Monday, May 24, 2010

Page 2: Computing Neutrino Telescope

Talk overview⊙ The reasons for neutrino astrophysics... and why in the Mediterranean Sea [1]

⊡ the ANTARES experiment [2]

⊡ the NEMO project [3]

⊡ the KM3NeT Consortium [4]

So... computing for:

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 3: Computing Neutrino Telescope

Cherenkov light

ν μ

Upgoing μ

dense neutrino target

transparent media

ν μ μ

W

N X

Neutrino Telescopes perform indirect measurements

charge current interaction

PMTs lattice

incoming neutrino

(Markov, 1960 [5])

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 4: Computing Neutrino Telescope

with neutrinos, look far & inside possible thick sources

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 5: Computing Neutrino Telescope

⊙ very small neutrino cross-sections with matter σ ∼ 7.8 10-36(E0.4/GeV) cm2

enhanced astrophysical neutrino fluxes w.r.t. atmospheric background above 10 TeV,

so small fluxes , dF/dE ∼ 10-7.5 E-2 /(GeV s sr cm2 )

➠ Telescope Vol > 1 km3 for some events/year

⊙ big volumes ➠ save money with the cheapest materials

➠ use of sea water (or lake & polar ice...not discussed here)

BUT atmospheric background is less when staying deeper

➠ DEEP sea water ( below 2000 m usl)

these are the main Mediterranean nu-Tel. features

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 6: Computing Neutrino Telescope

ANTARES NEMO NESTORcurrent projects

Neutrino Telescopes “CLUB-MED”

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 7: Computing Neutrino Telescope

⊙ big volumes + water optical properties (absorption & scattering of blue-green photons ∼ 70-100 m) + good angular resolution (.2 o) for usefully pointing (that’s neutrino ASTRONOMY)

➠ Many detection elements (N. PMTs > 5000/km3)

Constraints for a “Mediterranean” computing model

⊙ signal-to-noise ratio extremely disfavored : muon rate (atmospheric’s dominate) : 100 Hz/km3

40K decays (constant) : 40 kHz/PMT(10”, 0.5 p.e. thld) Bioluminescence* (occasional) : up to some MHz/PMT(10”, 0.5 p.e. thld)

+ No “beam crossing” reference such as for experiments at Colliders

+ complex DAQ structures in extreme conditions (mandatory: minimal underwater maintaining)

➠ ALL DATA TO SHORE approach * should disappear below 2500 m u.s.l

Sea science studies

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 8: Computing Neutrino Telescope

Data Aggregation

•Data are received on-shore according to a given Detector-TriDAS farm topology

•Time slices management

Data Routing

•Packet switching: Time slices are transferred to the Trigger servers (via firmware or software)

Data Processing

•Calibration correction•System & Data monitoring• Data selection •Event building

Data Storage

•Transient Data (RAM - local storage)•Temporary storage (local storage)•Permanent storage (metadata in DB)

Monday, May 24, 2010

Page 9: Computing Neutrino Telescope

DAQ model defined with finite state machines.The global system can be seen as a collection of hierarchically organized subsystems that are dynamically instantiated or destroyed during state transitions(CORBA, ICE, CHSM)

•Metadata management •Configuration history •Run conditions management•Multiple programming languages access•Caching•Database access performance

Monday, May 24, 2010

Page 10: Computing Neutrino Telescope

optical data : the highest unidirectional throughput, from off- to on-shore it determines the TriDAS specifics ( ∼ 100 Mbps/PMT)

slow control : bidirectional, no filtering, low band occupying

clock : could be an external or main-stream embedded information

acoustic positioning : from off- to on-shore; low band consuming (∼10 Mbps/hydr.pair)

acoustic detection : (possibly) embedded in acoustic positioning data

environmental sensor : scheduled or on-demand; low band consuming (∼ 100 kbps/instrument)

Principal Information Channels

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 11: Computing Neutrino Telescope

The ANTARES “pioneer” experienceart

ist’s v

iew

Instrumented volume 0.05 km3

➠ high PMT density

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 12: Computing Neutrino Telescope

ARS1 estimed freq. (Hz):74934.5 ... x 2 = 149869ARS2 estimed freq. (Hz):74937.4 ... x 2 = 149875estimed ARS deadtime: 0.04 %

100 Mb/s

VxWorks OS

The continuous data stream is stored in the SDRAM subdivided into 10-100 ms “timeslices”

The single photon-electron hit on a PMT is sized 6 bytes - timestamp + total charge; the pulse wave-form is not used but for special acquisitions . @ 350 kHz/PMT , the LCM trhoughput is 50 Mb/s(on 100 Mb/s output: factor 2 safety)

Each PMT has a ARS pair in a token ring for minimizing the total sampling dead time (<0.04%). NOTE that the effective ANTARES dead time is bigger and it is the convolution of many other factors (see ahead).

LCM

The Clock system 20MHz, provides common clock signal to all the ARS chips, with accuracy of 100 ns. The ARSs, with local counters, push time resolution to ∼ 1 ns.

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 13: Computing Neutrino Telescope

4 eth x 100 Mb/s (4x 50 Mb/s max)

5 eth. x 1 Gb/s(5x 250 Mb/s max)

The Density Wavelenght Division Multiplexing (DWDM) uses multiple wavelengths to transmit different streams of data along a single fibre

string power and SC data

1 Strin

g se

ctor (

5 LCM

s)

on line trigger farm &monitoring PCs

GLOBAL SKETCH of ANTARES DAQ

40 km Electro-Optical

cable

4 ( max 15) Gb/soptical data

thput

✴ TCP/IP PROTOCOL

✴ 12 OPTICAL FIBRES (1/STRING)

✴ 15 Gb/s maximum throughput (@ 350 kHz/PMT) (normal conditions 4 Gb/s @ 80 kHz)

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 14: Computing Neutrino Telescope

BARREL SHIFT PARADIGM [6]

how is treated the 15 Gb/s throughput

DQueue DQueue DQueueDispatchersN.DQ: 15 servers

4-core

Software Routing with CONTROL-HOST [7] ✴ Barrel Shift delay: multiple of TimeSlices duration. It determines the number of concurrent transmitting LCMs. Limit: the FPGA buffer (64 MB).

✴ The number of DQueue server cannot be arbitrary. Too few DQueues ➠ too much throughput from LCMs.

Too many DQueues ➠ too much delay for LCMs

When high rates: VETO on LCM data sending ➠ Dead Time (10-20 %)

VxWorks OS on the LCMs cannot open a high number of sockets

Tommaso Chiarusi Workshop CCR-INFN Grid 2010

N.DF: 23 servers 4-core92 DataFilter processes

Monday, May 24, 2010

Page 15: Computing Neutrino Telescope

All the processes in the same state at any time

Concurrent Hierarchical State Machine sync-ingANTARES DAQ model

RUN definition: acquisition with a stated detector configuration

Oracle®

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 16: Computing Neutrino Telescope

ANTARES DATA Processing

DFilter machines: 23 Intel Xeon 4 core 2.6 GHz CPU Basic Trigger: the so called L1: - pair of coincidences in 20 ns on the same storey

- charge over threshold Other Triggers: complex 1D or 3D scan, Special Pointing (Galactic Center, TQ),

Magnetic Monopoles, “minimum bias”

data suppression: from “normal” 0.5 GB/s incoming to 2 GB/h stored (factor 10-3)

automatic run restart

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 17: Computing Neutrino Telescope

ANTARES data Storage, Bookkeeping and Access

Storage Request Broker (SRB)

✴Every night data are copied from local disks to storage’s

✴All data stored at CC Lyon [8]

✴Concurrent data delivery

✴Storage access via web/terminal within the SRB middleware [9]

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 18: Computing Neutrino Telescope

ANTARES data Storage, Bookkeeping and Access

Storage Request Broker (SRB)

✴Every night data are copied from local disks to storage’s

✴All data stored at CC Lyon [8]

✴Concurrent data delivery

✴Storage access via web/terminal within the SRB middleware [9]

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 19: Computing Neutrino Telescope

Views from the ANTARES Control Room

Institute Michel Pacha @ La Seyne sur MérTommaso Chiarusi Workshop CCR-INFN Grid 2010

ToulonLa Seyne sur mer

Monday, May 24, 2010

Page 20: Computing Neutrino Telescope

MONITORING PARADE

JavaC++

Run Control

Power Control

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 21: Computing Neutrino Telescope

Tommaso Chiarusi Workshop CCR-INFN Grid 2010

C++ / ROOT Slow Control

OMs status

On line event reco

Monday, May 24, 2010

Page 22: Computing Neutrino Telescope

The ANTARES (preliminary) upgoing neutrino sky map

Monday, May 24, 2010

Page 23: Computing Neutrino Telescope

1. NEMO Phase 1the Minitower @

the Test Site

2. NEMO Phase 2Full NEMO Tower @

Capopassero

3. NEMO KM3Project within KM3NeT

NEMOCapo Passero Site

100 km

20 km-2000 m

The NEMO mission is optimizing the telescope features to the KM3 scale

Tommaso Chiarusi Workshop CCR-INFN Grid 2010

Nemo Phase 1MiniTower

Nemo Phase 2Full Tower

Test Site

250

m

800

m

Monday, May 24, 2010

Page 24: Computing Neutrino Telescope

Hit samples Vs. pulse time on a 10” PMT with 0.5 p.e. thr. (NEMO Ph.1 data)

S.P.E. wave form(∼16 samples = 16 Bytes)

S.P.E. HIT SIZE:Hit PMT Info

+ Hit Time

+Hit Charge

+Hit Wave Form(samples)

28 Bytes

Tommaso Chiarusi Workshop CCR-INFN Grid 2010

Ethernet interface

FCM [10]

Monday, May 24, 2010

Page 25: Computing Neutrino Telescope

Expected Data Rates10” PMT

single rate(kHz)

Data Rateper PMT(Mbps)

Data Rate per Floor*

(Mbps)

Data Rateper D.U.**

(Gbps)

Data Rate per km3 ***

(Gbps)

40 8.8 35.2 0.5 50

80 16.8 67.2 1.1 110

150 32 128 2.0 200

375 64 256 5.0 500

** 16 Floor/D.U. * 4 PMT/ Floor ** 100 D.U.

(bare 40K)

(NEMO Ph. 1)

(present DAQ)

(expanded DAQ )

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 26: Computing Neutrino Telescope

Atmospheric Muon Signal (@ 3000 m depth) assuming - Rate µ atm: 100 Hz - Rate bkg : 300 kHz - N. PMTs: 8000 - Rec. time window: 6 µs

Post-Trigger data rate: ∼ 40 MBpsStored data /day ∼1 TB !!!!

Upgoing Neutrino Signal assuming - Rate ν : < 4 10-3 Hz - Rate bkg : 300 kHz - N. PMTs: 8000 - Rec. time window: 6 µs

Post-Trigger data rate: ≾ 2 kBpsStored data /day ≾ 150 MB !!!!

Approximative computations according to the KM3 document

The TriDAS reduces the data rate by filtering the data stream bunched in Time Slice (TS). A TS contains all data from all ( or part of) the detector occurred in a given time interval (~100 ÷200 ms).

WHAT IS SCALABLE?

The TriDAS principal elements:

- Trigger System Controller (TSC): monitors and serves the TriDAS - (1 GbE I/O) - Hit Managers (HM): receive optical data; prepare and distribute the TS to the TCPU - TriggerCPUs (TCPU): apply the trigger logics to the assigned TS and transfer the selected data to the EM;

- Event Manager (EM): receives the triggered data from TCPUs and build the Post-Trigger events data file

For 1 D.U. the throughput is to reduce from 256 MBps → 400 kBps (3 GB/day)

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 27: Computing Neutrino Telescope

TriDAS for NEMO Ph.1

Gbit switch Master CPU <70 MB/s

Hard Disk

Hard Disk

CPU

PC Floor 1

PC Floor 2

< 70 MB/s

Windows

Linux

PC Monitor

DM - RC

PC Floor 3

PC Floor 4

The maximum throughput of the MiniTower was ≤ 512 MbpsOne Machine, the Master CPU ( cpu: 2x 2.3 GHz, ram: 4 GB, HD: 2 x160 GB, OS: SL 3.3 ), playing all the roles (Hit Manager, Trigger, Event Manager) with multi-threading (posix threads).

Storaging via 1Gb eth. (TCP/IP) to a dedicated server

CABLEFloor 1

Floor 2

Floor 3

Floor 4

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 28: Computing Neutrino Telescope

TriDAS for a first prototype Detection Unit with 16 Floors; expected through-put ≤ 2 Gbps standard 1 GbEthernet Networking

10 servers HP cpu: 8-core 32 bit 2 GHz, ram: 6 GB, HD: 160+500/160 GB, OS: SL5Counting room @ Capo Passero, to be 1 Gb eth connected to LNS for off-line analysis

Floor 1Floor 2

Floor 3Floor 4

Floor 5Floor 6

Floor 7Floor 8

Floor 9Floor 10

Floor 11Floor 12

Floor 13Floor 14

Floor 15Floor 16

Shore interfaces to:

TCPU server

TCPU server

TCPU server

TCPU server

Tommaso Chiarusi Workshop CCR-INFN Grid 2010

CHSM(CORBA)

Monday, May 24, 2010

Page 29: Computing Neutrino Telescope

νKNCNPMTNTCPU = RCPU

0

0.5

1

1.5

2

2.5

3

3.5

4

4.5

5

0 50 100 150 200 250 300 350

N.

of

TC

PU

s

Single Rate on a PMT (kHz)

RCPU SHitNACC = Drate HM→TCPU

Drate HM→TCPU = 1 GbpsSHit= 224 b

RCPU = 2.3 GHzNACC=515

NPMT=64NC=NACC=515

Number of Available Clock Cycles per CPU per TS:

Estimed number of necessary TCPU:

If the req. NC>NACC(i.e. trigger algo. is slow)

ADD TCPU! 150

2.3

Overesized-a

ssumption

In principle we could divide by a factor 5 and live with only 1 TCPU !!! But alas, net-bottlenecks.

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 30: Computing Neutrino Telescope

MonitoringVia ControlHost: a Tag Controlled Data Dispatching [V. Maslenikov et al. CASPUR, http://afs.caspur.it/temp/ControlHost.pdf]

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 31: Computing Neutrino Telescope

Multi purpose on-line Visualizer (QT4+ROOT)

Monday, May 24, 2010

Page 32: Computing Neutrino Telescope

On(Off)-line Event Display

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 33: Computing Neutrino Telescope

GANGLIA publish on the web XXX-vs.-time plots

with A. Paolucci (CCL Bologna)

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 34: Computing Neutrino Telescope

NAGIOS allows to define “observables” for monitoring the good/critical status of the machines, and sends

alarms (mails, “beeps”, “red lights”...)Tommaso Chiarusi Workshop CCR-INFN Grid 2010

with A. Paolucci CCL Bologna

Monday, May 24, 2010

Page 35: Computing Neutrino Telescope

KM4

~80 km

Capo Passero ~25 km

Etna Volcano

Catania Gulf

Nemo Phase-1Test Site

25 km

NEMO Phase 1

NEMO Phase 2

NEMO Sites and shore stations

- 100 km electro-optical cable (>50 kW, 20 fibres) deployed in 2007- DC/DC power converter built by Alcatel tested and working; installation in 2009- On-shore laboratory (1000 m2) inside the harbour area of Portopalo completed

Monday, May 24, 2010

Page 36: Computing Neutrino Telescope

Single PMT OM (10” - 8” PMTs)

2 x

2 x

Monday, May 24, 2010

Page 37: Computing Neutrino Telescope

... in any case: the off- ➠ on-shore communication is stated to 2.5( max 5 )Gbps/D.U.

DWDM

Monday, May 24, 2010

Page 38: Computing Neutrino Telescope

Possible Network infrastructure for km3 TriDAS (90 D.U.)

[11]

Monday, May 24, 2010

Page 39: Computing Neutrino Telescope

0

10

20

30

40

50

60

0 50 100 150 200 250 300 350

N.

of

TC

PU

s

Single Rate on a PMT (kHz)

νKNCNPMTNTCPU = RCPU

RCPU SHitNACC = Drate HM→TCPU

Drate HM→TCPU = 10 GbpsSHit= 224 b

RCPU = 2.3 GHzNACC= 51

NPMT=8000NC=NACC=51

Number of Available Clock Cycles per CPU per TS:

Estimed number of necessary TCPU:

If the req. NC>NACC(i.e. trigger algo. is slow)

ADD TCPU!

Monday, May 24, 2010

Page 40: Computing Neutrino Telescope

Conclusions“all data to shore” is a challenging BUT feasible strategy for km3 underwater ν-telescope;

scalable TriDAS architecture suppliyng high data-stream (up to 500 Gbps) is possible and affordable with the present technology;

The NEMO Collaboration is developing a scaled TriDAS for a prototype Detection Unit (deployment scheduled for mid 2011 ).

ANTARES & NEMO pioneered the TriDAS for a u n d e rs e a n e u t r i n o t e l e s c o p e i n t h e Mediterranean. KM3NeT TDR will be delivered with the stated architecture.

Monday, May 24, 2010

Page 41: Computing Neutrino Telescope

Bibliography1. T. Chiarusi and M. Spurio, 2010, High-Energy Astrophysics with Neutrino Telescope, European Physics Journal C 65, nn. 3-4, 649

2. J. A. Aguilar, The Data Acquisition System of the ANTARES neutrino telescope, Nucl. Instrum. Meth. A 570 (2007) 107-116

3. T. Chiarusi, 2009, Scalable TriDAS for the NEMO Project, accepted for Nuclear Physics B (in press).

4. KM3NeT Collaboration – Conceptual Design Report. ( http://www.km3net.org/CDR/CDR-KM3NeT.pdf)

5. M.A. Markov, 1960, Proc. Int. Conf. on High Energy Physics, Univ. of Rocherster

6. M.F. Letheren, 1995 CERN School of Computing

7. V. Maslenikov et al. CASPUR (http://afs.caspur.it/temp/ControlHost.pdf)

8. Centre de Calcul IN2P3 (http://cc.in2p3.fr/)

9. http://www.sdsc.edu/srb/index.php/What_is_the_SRB

10. F. Ameli, et al., EEE Trans.Nucl.Sci.55,233 (2008).

11. http://www.cisco.com/en/US/products/ps9402/index.html

THANK YOU!Special thanks to A. Paolucci and all at CCL (Bologna), G. Terreni (Pisa), F.

Safai-Therani (Roma), D. Salomoni and S. Zani (CNAF)

Monday, May 24, 2010

Page 42: Computing Neutrino Telescope

One muon event (only muon hits shown, no background)theTelescope is made of

DETECTION UNITS

20 floors, with 4 10” PMTs each

Tommaso Chiarusi RICAP 2009 Monday, May 24, 2010

Page 43: Computing Neutrino Telescope

Sky for a neutrino Mediterranean Telescope

Galactic Center visibility: 67 %

no visibility

at least 25%

at least 75%

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010

Page 44: Computing Neutrino Telescope

MONITORING PARADE

Sea Currents

Rate monitor

Tommaso Chiarusi Workshop CCR-INFN Grid 2010Monday, May 24, 2010