Upload
page
View
51
Download
1
Embed Size (px)
DESCRIPTION
Triggering In High Energy Physics. Gordon Watts University of Washington, Seattle NSS 2003. N14-1. Thanks to interactions.org for the pictures!. Introduction The Collider Environment Trigger Design Hardware Based Triggers Software Based Triggers Farms Conclusions. Outline. - PowerPoint PPT Presentation
Citation preview
Triggering In High Energy Physics
Gordon WattsUniversity of Washington,
SeattleNSS 2003
Thanks to interactions.org for the pictures!
N14-1
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
2
Outline• Introduction• The Collider
Environment• Trigger Design
– Hardware Based Triggers
– Software Based Triggers
– Farms• Conclusions
I am bound to miss some features of the triggering systems in various detectors. I have attempted to look at techniques and then point out examples.
Thanks to the many people who helped me with this talk (too numerous to mention).
Focus mostly on ongoing discussions and presentations (RT’03)
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
3
Why Trigger?• Bandwidth
– Raw bandwidth of many detectors is GB/s• Can’t Archive that rate!• A HEP experiment can write TeraBytes of data.
• CPU– Reconstruction programs are becoming
significant• Farms are 1000’s of computers.
– GRID to the rescue?• Physics
– Is all the data going to be analyzed (L4)?– Analysis Speed
• Small samples can be repeatedly studied, faster
Don’t LoosePhysics!
Disk&
Tape
ReconstructionFarm
Eliminate BackgroundAs Early As Possible
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
4
Why Trigger? IIIt’s obvious
Still, stunning to consider
Level 1
Level 2
Level 3
Trigger Information
Full Readout
Detector
~50-70 Hz~50-70 Hz
~1 kHz~1 kHz
~10 kHz~10 kHz
~1 kHz~1 kHz
~7 MHz~7 MHz
Level
Accept Rate
TB/year $ Tape(raw only)
Raw 7 MHz 25,700,000
10,500 m
L1 10 kHz 32,700 13 m
L2 1 kHz 3,270 1.3 m
L3 50 Hz 185 75,000•Assume 50% duty cycle for accelerator
•15,768,000 working seconds in a year•Event size 250 kB•$ are for raw events only
•$0.40/GB Much worse at LHC!1999: $400k for L3!
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
6
Trigger DesignLarge Scale HEP Experiments
High rate – 40 MHz Raw Trigger RateDAQ/Trigger Integration
Trigger Flexibility – move decision out of hardware and into firmware or software as early as possible.
Farm NodeFarm NodeFarm Node
HardwareTrigger
Multilevel triggers, custom hardware, network topologies
Small HEP Experiments, Medical ApplicationsLarge Data, lower frequency, fewer sources
Stereotyping
Detector
Detector CPUSmaller CPU/cluster can handle directly with specialized board
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
7
HEP AcceleratorsAccelerator Time
between collisions
(ns)
Luminosity(1030 cm-1s-1)
Energy (GeV)
CESR (CLEO) 14 3000 9-12KEKB (Belle) 2 10,000 8 x 3.5PEP-II (BaBar) 4.2 3,000 9 x 3.1
Accelerator Time between collisions
(ns)
Luminosity(1030 cm-1s-1)
Energy (GeV)
HERA (H1, Zeus)
96 14 920 x 30TeV (DØ, CDF, BTeV)
396 200 2,000LHC (Atlas, CMS, LHCb)
25 10,000 14,000
e+e-
ep, pp, pp
Source: PDB’02
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
8
HEP DetectorsDetector Number of
ChannelsSilicon Part
of Trigger?
Input Trigger Rate
Largest (Non) Physics Background
CLEO III 400,000 Yes (L2) L1: 72 MHzL2/L3: 1 kHzTape: <100 Hz
electron pairs
Belle 150,000 Not Yet L1: 50 MHzL2: 500 HzTape: 100 Hz
BhaBha &
BaBar 150,000 No L1: 25 MHzL3: 2 kHzTape: 100 Hz
BhaBha &
H1, ZEUS 500,000 Yes (ZEUS)
L1: 10 MHzL2: 1 kHzL3: 100 HzTape: 2-4 Hz
Beam-gas
HERA-B 600,000 Yes (L2) L1: 10 MHzL2: 50 kHzL3: 500 HzL4: 50 HzTape: 2 Hz
Beam-wire scatteringInelastics
CDF, DØ (Run II)
1,000,000
Yes (L2) L1: 7 MHzL2: 10-50 kHzL3: .3-1 kHzTape: 50 Hz
QCD, pileup, multiple interactions
ATLASCMS
107 High Level Trigger
L1: 75 kHzHLT: 100 Hz
QCD, pileup, multiple interactions
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
9
High Rate Trigger/DAQData Rates are 100 MB/secBeam Crossing Rates are MHzMany times at edge of technology when first designed
But off-the-shelf by the time they are running!
Even GLAST uses themGamma-ray Large Area Space Telescope
1 of 25 towers
SiTracker
L1 Cal
L1 Tracking
L1 Decision Hardware
CalHardware
L2 CPU(in tower)
FullReadout
L3 CPU(Farm)
To Earth
Upon detection of Gamma-ray burst notify by internet within 5 seconds
ATLAS, BaBar, Belle, BTeV, CDF, CMS, DØ, LEP etc.
• Level 1: hardware based– Physics Objects, little associations– Coarse Detector Data– Deadtimeless
• Level 2– Hardware to preprocess data
• Some Muon processors, Silicon Triggers
– Software to combine• Matches, Jet finders, etc.
• Level 3: a commodity CPU farm– Complete event information
available
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
10
Trends In Big TriggerCommodity ElectronicsCPU and Networking
Delay the decision as long as possible •More Data Available•More sophisticated decision platform (bug fixing!)
DAQ typically sits before a Farm Trigger
DAQ is driving deeper into the Trigger System BTeV, ATLAS, CMS…
Commodity Electronics Enforces Common InterfaceCustom Hardware Plethora of data formats
Allows for economies of scale
Hardware SophisticationASICs, FPGAs
HLT Algorithms in Hardware
(ENET in a FPGA) HLT getting more sophisticated
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
11
The Age Of The Network
S
SSS
S
S
S
D D D
DD
D
DD
S
SSS
S
S
S
D D D
DD
D
DD
BTeV
Fast Networking Moves Data from a local environment to a Global One
Cheap DRAM allows for large buffering, and latency
Cheap Commodity Processors allow for powerful algorithms with large rejection
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
12
DØ & CDFTypical BIG Experiment
~1 million readout channelsEvent size is 250 KB12.5 MB/sec written to tape
Trigger – Typical Multi-level
Level 1
Level 2
Level 3
Trigger Information
Full Readout
Detector
50-100 50-100 HzHz
~1(.3) kHz~1(.3) kHz
~5(20) kHz~5(20) kHz
~1(.3) kHz~1(.3) kHz
~2.5 MHz~2.5 MHz
L1:Dead-timeless,pipelined, HardwareL2&L3: can cause deadtime, variable length algorithmsL2: Hardware and SoftwareL3: CPU Farms
Data for > 2.5 year
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
13
Triggering StrategySimilar but Different Approaches driven by PhysicsCDF: More emphasis on B Physics
Jet of particles w/displaced track.Hard to pick out of background
Too sophisticated to do well at L1Pumps the extra data into L2 (CDF)
DØ cuts harder at L1L1 Trigger better at object correlations
CAL, Tracks (connection), etc….
CAL, Tracks (connection), etc….
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
14
Tracking at the TeV– Track finding is difficult
• Must Implement hit searches and loops in hardware. Use FPGAs in the Track Trigger
– Fiber Tracker (Scintillating Fiber) – DØ– Open Cell Drift Chamber - CDF– Finds Tracks in bins of pT
• 1.5, 3, 5, and 10 GeV/c @ DØ– Done by equations:
• Layer 1 Fiber 5 & Layer 2 Fiber 6 & etc.• Inefficiencies are just more equations
– Firmware becomes part of the trigger• Versioning• Fast & Flexible - Can be reprogrammed
later.• Painful for the trigger simulation!
CFTLayers
ForwardPreshower
Track
– 80 sectors– 1-2 chips per
sector– 16K equations
per chip–One eqn per possible track.
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
15
Drift Chamber Tracking• DØ’s Fiber Tracker gives
instant position information– No drift time to account for.
• CDF divides hits into two classes– Drift time < 33 ns– Drift time 33-100ns
• This timing information is used as further input to their FPGAs for greater discrimination.– This is not done on every
wire.
Prompt Hit
Non Prompt Hit
•Babar also uses FPGAs & LUTs to do track finding
• 2ns beam crossing time!The Central Track Trigger System of D0 Experiment
The D0 Central Tracking TriggerNew Trigger Processor Algorithm for a Tracking Detector in a Solenoidal Magnetic FieldN36-57
N36-55N14-5
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
16
Trigger Manager• Simple matching
between detectors– track-cal– Usually first time
• Decision Logic is always programmable– Often part of Run
configuration– Human Readable Trigger
List• May manage more than
one Trigger Level.• Usually contains scalars
– Keeps track of trigger live-time for luminosity calculations.
Detector A Detector B Detector C
Trigger LogicFor A
Trigger LogicFor B
Trigger LogicFor C
BX Synchronization
Decision Logic(prescales too)
Readout Datato Next Level
FECNoti-
fication
7 BX 4 BX 9 BX
Per Trig Scalars
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
17
Level 2 Architecture
GlobalCAL
Preprocessor
SiPreprocessor
Common platform based on embedded DEC Alpha CPU and common communication lines for Run I
Hybrid of Commodity Processor & Custom Hardware
Custom HardwareControl CPU CPU
Similar to L1 Global
Upgrade CDF: Pulsar BoardDØ: L2 Boards
CDF Pulsar ProjectN29-4An Impact Parameter Trigger for the DØ ExperimentN29-5
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
18
High Level Trigger• Move the data from the ROCs• Build the Event• Process Event in Physics Software• Deliver accepted events
~100 Read Out Crates~100-250 MB/sec~100ms - 1 second/event
processing time
Both experiments use networking technology
ATM w/concentrators, SCRAMNet for routing
Gb Ethernet with central switch, ENet for routing
See the Online/DAQ Sessions
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
19
CDF• Data Flows when
Manager has Destination– Event Builder Machines
• One of first to use switch in DAQ.
• Dataflow over ATM– Traffic Shaping
• Backpressure provided by Manager.
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
20
DØ High Level TriggerNetwork Based DAQ
The Linux-based Single Board Computer for Front End Crates in the DZERO DAQ System
Dataflow in the DZERO Level 3 Trigger/DAQ SystemN36-71
N36-70
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
21
Fermi TeV Planned Upgrades
• Peak Luminosity matters for Trigger/DAQ
• Tevatron Trigger Upgrades are Under review– Decreased Tevatron
Luminosity Expectations
• Both experiments still believe they will be required.
•L1 CAL Sliding Window (ATLAS algorithm)•L1 CAL Track Match•Track Stereo info used at L1•L2 CPUs•L2 Si Trigger improvements
200
250
300
~2.8e32
Accelerator draft plan: Peak luminosities
Pea
k Lu
min
osity
(x10
30cm
-2se
c-1)
0
50
100
150
2003
Start of Fiscal Year
~1.6e32
2004
2005
2006
2007
2008
2009
2010
Today(4.5x1031)
All under Review•COT Readout Speedup•L1 Track Trigger•L2 Si•L2 CPUs•L3 Network DAQ Upgrade
13 racks of ’88 to 3 of present day, and better functionality (planned CAL upgrade in DZERO)
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
22
ATLAS & CMSO(10) Larger
Almost no custom Hardware after L1
L1: 75kHzHLT: 100 Hz
Both Experiments use full Network
DAQ after L1
Solve Data Rate Problem DifferentlyBig Switch (CMS)ROI (ATLAS)
Event size: 1MB
But Approach Is Similar
75GB/sec
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
23
CMS Trigger Architecture
CMSMust operate at 100 kHz
50 kHz @ startup CMS eISO Card
CMS Uses only Muon & Calorimeter• Tracking too Complex & Large• Muon TrackingBoth ASIC and FPGAPipeline is on-detector (3.2 s)
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
24
ATLAS Level 1Similar Design
2s Pipe Line
CAL
Clus
ter F
inde
r
CPU Farm (100-150 dual CPU, ~1ms/event)Offline Tracking @ L2 by CDF like XFT proposed
Level 2
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
25
Regions of Interest (ROI)• Bandwidth/Physics Compromise
– Farm usually reads out entire detector– Hardware often look at single detector
• ROI sits in the middle– Farm CPU requests bits of detector
• Uses previous trigger info to decide what regions of the detector it is interested in.
– Once event passes ROI trigger, complete readout is requested and triggering commences.
• Flexible, but not without problems.– Must keep very close watch on trigger
programming!– Farm decisions happen out-of-event-
order• Pipelines must be constructed
appropriately.• ATLAS, HERA-B, BTeV…
L2 Farm CPU
Basic Cal
L1 Info
Ele Conf
Full Trigger
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
26
CMS High Level TriggerNO L2 Trigger
•1 second buffering in Hardware (100Kx1meg)
•1000 HLT Farm Nodes
•700 ROC, 64 Event BuildersBuild Events in Two Stages
8 PlanesWill use Myrinet for first planeGB under evaluation
8x8 8x8 8x8 8x8 8x8 8x8
1 of 8 Planes
ROC
EVB
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
27
LHC Posters And Papers
Full Crate Test and Production of the CMS Regional Calorimeter Trigger SystemN29-3
N14-2 The First-Level and High-Level Muon Triggers of the CMS Experiment at CERN
The ATLAS Liquid Argon Calorimeters Readout System
ATLAS Level-1 Calorimeter Trigger: Subsystem Tests of a Jet/Energy-sum Processor Module
Test Beam Results from the ATLAS LVL1 Muon Barrel Trigger and RPC Readout Slice
Beam Test of the ATLAS End-cap Muon Level1 Trigger System
N29-2
N36-68
N29-1
N14-2
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
28
BTeV
23 million pixels – in Level 1!132 ns Bunch Spacing100 kb/eventTrigger full rate. 800 GB/sec!!
Bs Mixing, Rare Decays
Level 11. Hit clustering in Pixels –
FPGAs2. Cluster linking in inner
and outer layers – FPGAs3. Track finding in B field (pT>0.25) – Embedded CPU Farm4. Hard Scatter Vertex Finding – Embedded CPU Farm5. DCA test for displaced tracks due to B decays – Embedded
CPU Farm8 Planes of L1+L2/L3
(round robin)
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
29
BTeV Trigger Design
800 GB/secoptical links
Global L1
L1 Farms
L2/L3 Farms
Buffering for 300k interactions (300ms)
200MB/sec3.8 kHz
78 kHz
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
30
BTeVHLT Algorithm
Large # of High End Boxes (dual 4 GHz)Uses ROI – Similar to ATLASBut L2/L3 are interleaved on same box
Trigger/DAQ TDR Just Out Tremendous amount of
simulation work
Pre-prototype of the BTeV Trigger Level 1 Farm Processing Module.N36-66
N36-52 Hash Sorter - Firmware Implementation and an Application for the Fermilab BTeV Level 1 Trigger System
N36-65 Real-Time Embedded System Support for the BTeV Level 1 Muon Trigger
N36-61 Failure Analysis in a Highly Parallel Processor for L1 Triggering
N29-7 Data Flow Analysis of a Highly Parallel Processor for a Level 1 Pixel Trigger
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
32
OthersToo much out there to be covered in this talk!
• ZEUS has added a CPU farm to its custom Hardware L1/L2 trigger
• LHCb uses a 3D torus network topology to move events through its farm quickly (1200 CPUs)
• FPGAs come with Ethernet Controllers built-in
• The Programmable NIC as a switch
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
33
Other HEP Talks and Posters
N29-6 BaBar Level 1 Drift Chamber Trigger Upgrade
N14-4 The Trigger System of the COMPASS Experiment
N36-59 A First Level Vertex Trigger for the Inner Proportional Chamber of the H1 Detector
N36-60 The Trigger System for the New Silicon Vertex Belle Detector SVD 2.0
N36-62 Rapid 3-D Track Reconstruction with the BaBar Trigger Upgrade
N36-64 The ZEUS Global Tracking Trigger
N36-69 Electronics for Pretrigger on Hadrons with High Transverse Momentum for HERA-B Experiment
N36-56 A Pipeline Timing and Amplitude Digitizing Front-End and Trigger System
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
34
Air Shower Observatories
1.5km
• Looking for ultra high energy cosmic rays (beyond GZK).
• 1600 Auger detectors• 1.5 km between each• 4 communications
concentrators in the array• PLCs, ASICS, Power PC
Low Data Rate (1200bps)Low Power (Solar)
Central Cell Fires (20 Hz)Look at surrounding rings to decide
N14-6 PLD First Level Surface Detector Trigger in the Pierre Auger Observatory
N36-54 The Trigger System of the ARGO-YBJ Experiment
10 second pipeline
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
35
Medical ApplicationsDecay Detection
U F O V
Desintegrations
Time 1
* *
Time 2
Time 3
Truecoïncidence
Time 2Time 1 Time 3
Coincidence w indow 40 ns
Coincidence w indow 40 ns
Falsecoïncidence
No Accelerator Clock
Positron Emission Tomography
(PET)
Trigger is a Scintillator
Coincidence
BGO Scintillator
Reduced matching window increases
resolution
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
36
PET
BGO
Different strength signalsDifferent rise times
Signal Threshold doesn’t have good
enough timing resolution!
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
37
PET
Time
Ampl
itude
Fit signal to BGO Model
Fit and Error Matrix done in FPGA
Allows a significant reduction in window
width(9.2ns)
N36-58 Development of a High Count Rate Readout System Based on a Fast, Linear Transimpedance Amplifier for X-ray Imaging
N36-63 A New Front-End Electronics Design for Silicon Drift Detector
Review ofTriggering in
HEP
Gordon [email protected]
University ofWashington,
Seattle
NSS 2003Oct 21, 2003
38
Conclusions• In HEP the Network rules the day
– Enables experiments to get localized data off detector– And into CPU with Global View– Global Trigger Decisions have more discriminating
power.• Technology Continues to Make Hardware Look
more like Software– FPGA’s increase in complexity– Moves complex algorithms further up the trigger chain– Following commodity hardware – CPUs, Embedded
processors, etc. Where next?• Most adopting Ethernet or its near relatives
– A few important exceptions.
A Great Session with Lots of Good Talks and Posters!Enjoy!