Data acquisition and Trigger (with emphasis on LHC)

Similar documents
Data acquisition and Trigger (with emphasis on LHC)

Monika Wielers Rutherford Appleton Laboratory

Introduction to Trigger and Data Acquisition

Trigger and Data Acquisition Systems. Monika Wielers RAL. Lecture 3. Trigger. Trigger, Nov 2,

Overview of the ATLAS Trigger/DAQ System

Data acquisi*on and Trigger - Trigger -

LHC Experiments - Trigger, Data-taking and Computing

Trigger and Data Acquisition at the Large Hadron Collider

Triggers For LHC Physics

The Run-2 ATLAS. ATLAS Trigger System: Design, Performance and Plans

ATLAS Phase-II trigger upgrade

First-level trigger systems at LHC. Nick Ellis EP Division, CERN, Geneva

Operation and Performance of the ATLAS Level-1 Calorimeter and Level-1 Topological Triggers in Run 2 at the LHC

Trigger and data acquisition

Electronics, trigger and physics for LHC experiments

TRIGGER & DATA ACQUISITION. Nick Ellis PH Department, CERN, Geneva

Trigger and DAQ at the LHC. (Part II)

Track Triggers for ATLAS

First-level trigger systems at LHC

LHCb Trigger & DAQ Design technology and performance. Mika Vesterinen ECFA High Luminosity LHC Experiments Workshop 8/10/2016

The design and performance of the ATLAS jet trigger

DAQ & Electronics for the CW Beam at Jefferson Lab

arxiv: v2 [physics.ins-det] 13 Oct 2015

ATLAS Muon Trigger and Readout Considerations. Yasuyuki Horii Nagoya University on Behalf of the ATLAS Muon Collaboration

Real-time flavour tagging selection in ATLAS. Lidija Živković, Insttut of Physics, Belgrade

PoS(EPS-HEP2017)476. The CMS Tracker upgrade for HL-LHC. Sudha Ahuja on behalf of the CMS Collaboration

The LHCb trigger system

The LHC Situation. Contents. Chris Bee. First collisions: July 2005! Centre de Physique des Particules de Marseille, France,

CMS SLHC Tracker Upgrade: Selected Thoughts, Challenges and Strategies

Performance of the ATLAS Muon Trigger in Run I and Upgrades for Run II

The Compact Muon Solenoid Experiment. Conference Report. Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland

Level-1 Track Trigger R&D. Zijun Xu Peking University

The ATLAS Trigger in Run 2: Design, Menu, and Performance

Hardware Trigger Processor for the MDT System

GPU-accelerated track reconstruction in the ALICE High Level Trigger

Hardware Trigger Processor for the MDT System

What do the experiments want?

The detector read-out in ALICE during Run 3 and 4

Triggering at ATLAS. Vortrag von Johannes Haller, Uni HH Am ATLAS-D Meeting, September 2006

The Run-2 ATLAS Trigger System

LHCb Trigger System and selection for Bs->J/Ψ(ee)φ(KK)

Trigger Overview. Wesley Smith, U. Wisconsin CMS Trigger Project Manager. DOE/NSF Review April 12, 2000

Layout and prototyping of the new ATLAS Inner Tracker for the High Luminosity LHC

The Liquid Argon Jet Trigger of the H1 Experiment at HERA. 1 Abstract. 2 Introduction. 3 Jet Trigger Algorithm

EPJ C direct. The ATLAS trigger system. 1 Introduction. 2 The ATLAS experiment. electronic only. R. Hauser, on behalf of the ATLAS collaboration

The CMS Muon Trigger

Level-1 Calorimeter Trigger Calibration

Triggers: What, where, why, when and how

Field Programmable Gate Array (FPGA) for the Liquid Argon calorimeter back-end electronics in ATLAS

Signal Reconstruction of the ATLAS Hadronic Tile Calorimeter: implementation and performance

Streaming Readout for EIC Experiments

The Commissioning of the ATLAS Pixel Detector

9. TRIGGER AND DATA ACQUISITION

The CMS ECAL Laser Monitoring System

The LHCb Upgrade BEACH Simon Akar on behalf of the LHCb collaboration

The LHCb trigger system: performance and outlook

LHCb Preshower(PS) and Scintillating Pad Detector (SPD): commissioning, calibration, and monitoring

Upgrade of the ATLAS Thin Gap Chamber Electronics for HL-LHC. Yasuyuki Horii, Nagoya University, on Behalf of the ATLAS Muon Collaboration

Attilio Andreazza INFN and Università di Milano for the ATLAS Collaboration The ATLAS Pixel Detector Efficiency Resolution Detector properties

Calorimeter Monitoring at DØ

L1 Track Finding For a TiME Multiplexed Trigger

Opera&on of the Upgraded ATLAS Level- 1 Central Trigger System

Phase 1 upgrade of the CMS pixel detector

SLHC Trigger & DAQ. Wesley H. Smith. U. Wisconsin - Madison FNAL Forward Pixel SLHC Workshop October 9, 2006

The CMS electromagnetic calorimeter barrel upgrade for High-Luminosity LHC

Development of a Highly Selective First-Level Muon Trigger for ATLAS at HL-LHC Exploiting Precision Muon Drift-Tube Data

ATLAS and CMS Upgrades and the future physics program at the LHC D. Contardo, IPN Lyon

PoS(LHCP2018)031. ATLAS Forward Proton Detector

Micromegas calorimetry R&D

ATLAS ITk and new pixel sensors technologies

The Status of ATLAS. Xin Wu, University of Geneva On behalf of the ATLAS collaboration. X. Wu, HCP2009, Evian, 17/11/09 ATL-GEN-SLIDE

The upgrade of the LHCb trigger for Run III

Beam Condition Monitors and a Luminometer Based on Diamond Sensors

CTEQ Summer School. Wesley H. Smith U. Wisconsin - Madison July 19, 2011

Expected Performance of the ATLAS Inner Tracker at the High-Luminosity LHC

Introduc*on. Trigger Hands- On Advance Tutorial Session. A. Ave*syan, Tulika Bose (Boston University)

Installation, Commissioning and Performance of the CMS Electromagnetic Calorimeter (ECAL) Electronics

Firmware development and testing of the ATLAS IBL Read-Out Driver card

The trigger system of the muon spectrometer of the ALICE experiment at the LHC

The 1st Result of Global Commissioning of the ATALS Endcap Muon Trigger System in ATLAS Cavern

8.882 LHC Physics. Detectors: Muons. [Lecture 11, March 11, 2009] Experimental Methods and Measurements

Some Studies on ILC Calorimetry

Physics at the LHC and Beyond Quy Nhon, Aug 10-17, The LHCb Upgrades. Olaf Steinkamp. on behalf of the LHCb collaboration.

CMS Phase 2 Upgrade: Preliminary Plan and Cost Estimate

PoS(ICPAQGP2015)098. Common Readout System in ALICE. Mitra Jubin, Khan Shuaib Ahmad

Study of the ALICE Time of Flight Readout System - AFRO

Contents. Why waveform? Waveform digitizer : Domino Ring Sampler CEX Beam test autumn 04. Summary

HF Jet Trigger Upgrade R&:D Project

Calorimetry in particle physics experiments

The LHCb VELO Upgrade

Mitigating high energy anomalous signals in the CMS barrel Electromagnetic Calorimeter

CMS electron and _ photon performance at s = 13 TeV. Francesco Micheli on behalf of CMS Collaboration

Status of the LHCb Experiment

ATLAS LAr Electronics Optimization and Studies of High-Granularity Forward Calorimetry

CMS Tracker Upgrade for HL-LHC Sensors R&D. Hadi Behnamian, IPM On behalf of CMS Tracker Collaboration

PoS(Vertex 2007)034. Tracking in the trigger: from the CDF experience to CMS upgrade. Fabrizio Palla 1. Giuliano Parrini

US CMS Calorimeter. Regional Trigger System WBS 3.1.2

CMS Silicon Strip Tracker: Operation and Performance

Upgrade of the CMS Tracker for the High Luminosity LHC

Commissioning Status and Results of ATLAS Level1 Endcap Muon Trigger System. Yasuyuki Okumura. Nagoya TWEPP 2008

Transcription:

Lecture 2! Introduction! Data handling requirements for LHC! Design issues: Architectures! Front-end, event selection levels! Trigger! Upgrades! Conclusion Data acquisition and Trigger (with emphasis on LHC) Monika Wielers (RAL) DAQ and Trigger, Oct 6, 2014 1

DAQ challenges at LHC! Challenge 1! Physics Rejection power! Requirements for TDAQ driven by rejection power required for the search of rare events! Challenge 2! Accelerator Bunch crossing frequency! Highest luminosity needed for the production of rare events in wide mass range! Challenge 3! Detector Size and data volume! Unprecedented data volumes from huge and complex detectors DAQ and Trigger, Oct 6, 2014 2

Challenge 1: Physics! Cross sections for most processes at the LHC span 10 orders of magnitude! LHC is a factory for almost everything: t, b, W, Z! But: some signatures have small branching ratios (e.g. H γγ, BR 10-3 ) Process Production Rate 10 34 cm -2 s -1 inelastic ~1 GHz bbbar 5 MHz W lν 150 Hz Z lν 15 Hz ttbar 10 Hz Z 0.5 Hz H(125) SM 0.4 Hz! L=10 34 cm -2 s -1 : Collision rate: ~10 9 Hz. event selection: ~1/10 13 or 10-4 Hz! DAQ and Trigger, Oct 6, 2014 3

Challenge 1: Physics! Requirements for TDAQ driven by the search for rare events within the overwhelming amount of uninteresting collisions! Main physics aim! Measure Higgs properties! Searches for new particles beyond the Standard Model! Susy, extra-dimensions, new gauge bosons, black holes etc.! Plus many interesting Standard Model studies to be done! All of this must fit in ~300-1000 Hz of data written out to storage! Not as trivial, W lν: 150 Hz! Good physics can become your enemy! black DAQ and Trigger, Oct 6, 2014 4

Challenge 2: Accelerator! Unlike e + e - colliders, proton colliders are more messy due to proton remnants! In 2012 LHC already produced up to 30 overlapping p-p interactions on top of each collision (pile-up) è >1000 particles seen in the detector! no pile-up 20 pile-up events DAQ and Trigger, Oct 6, 2014 5

Challenge 3: Detector! Besides being huge: number of channels are O(10 6-10 8 ) at LHC, event sizes ~1.5 MB for pp collisions, 50 MB for pb-pb collisions in Alice! Need huge number of connections! Some detectors need > 25ns to readout their channels and integrate more than one bunch crossing's worth of information (e.g. ATLAS LArg readout takes ~400ns)! It's On-Line (cannot go back and recover events)! Need to monitor selection - need very good control over all conditions DAQ and Trigger, Oct 6, 2014 6

Let s build a Trigger and DAQ for this! What do we need? DAQ and Trigger, Oct 6, 2014 7

Let s build a Trigger and DAQ for this! What do we need?! Electronic readout of the sensors of the detectors ( front-end electronics )! A system to collect the selected data ( DAQ ) DAQ and Trigger, Oct 6, 2014 8

Let s build a Trigger and DAQ for this! What do we need?! Electronic readout of the sensors of the detectors ( front-end electronics )! A system to collect the selected data ( DAQ )! A system to keep all those things in sync ( clock ) DAQ and Trigger, Oct 6, 2014 9

Let s build a Trigger and DAQ for this! What do we need?! Electronic readout of the sensors of the detectors ( front-end electronics )! A system to collect the selected data ( DAQ )! A system to keep all those things in sync ( clock )! A trigger multi-level due to complexity DAQ and Trigger, Oct 6, 2014 10

Let s build a Trigger and DAQ for this! What do we need?! Electronic readout of the sensors of the detectors ( front-end electronics )! A system to collect the selected data ( DAQ )! A system to keep all those things in sync ( clock )! A trigger multi-level due to complexity! A Control System to configure, control and monitor the entire DAQ DAQ and Trigger, Oct 6, 2014 11

Let s look more at the trigger part DAQ and Trigger, Oct 6, 2014 12

Multi-level trigger system! Sometime impossible to take a proper decision in a single place! too long decision time! too far! too many inputs! Distribute the decision burden in a hierarchical structure! Usually τ N+1 >> τ N, f N+1 << f N! At the DAQ level, proper buffering must be provided for every trigger level! absorb latency! De-randomize DAQ and Trigger, Oct 6, 2014 13

LHC DAQ phase-space DAQ and Trigger, Oct 6, 2014 14

Hardware Trigger (L0, L1)! Custom electronics designed to make very fast decisions! Application-Specified Integrated Circuits (ASICs)! Field Programmable Gate Arrays (FPGAs)! Possible to change algorithms after installation! Must cope with input rate of 40 MHz! Reduce rate from 40 MHz to ~100 khz! Otherwise cannot process all events! Event buffering is expensive, too! Use pipeline for holding data during L1 processing! Digital/analog custom front-end pipelines! Parallel processing of different inputs as much as possible DAQ and Trigger, Oct 6, 2014 15

Trigger Latency This time determines the depth of the pipeline DAQ and Trigger, Oct 6, 2014 16

Software Trigger: Higher Level Trigger (HLT)! L1 selected a large rate (up to 100 khz) of events that might be interesting! These events are not kept yet (rate too high for storage), but sent to the HLT for additional filtering! Use network-based High Level Trigger computer farm(s)! commercially available HW organized in a farm CPU CPU CPU CPU CPU CPU CPU CPU CPU DAQ and Trigger, Oct 6, 2014 17

Higher Level Trigger! Massive commercial computer farm! ATLAS: L2 and L3 handled by separate computing farms in 2012! Roughly 17k CPUs that can be freely assigned to either! CMS: Single computing farm (roughly 13k CPUs in 2012)! Parallel processing, each CPU processes individual event! Resources are still limited! Offline: Full reconstruction takes seconds (minutes)! Online latency: ms - s (input rate dependent)! Need to reduce rate to O(few 100 Hz)! Note, output rate mainly driven by offline DAQ and Trigger, Oct 6, 2014 18

The ATLAS Trigger/DAQ System! Overall Trigger & DAQ architecture: 3 trigger levels! Level-1:! 2.5 µs latency! 75 khz output in 2012, 100 khz in 2015! DAQ/HLT! Analyse regions around particles identified at L1 or whole event! Average output rate in 2012: 400 Hz prompt, 200 Hz parked, ~1kHz in 2015! Processing time: few seconds! Average event size 1.5 MB in 2012, DAQ and Trigger, Oct 6, 2014 ~2 MB in 2015 19

The CMS Trigger/DAQ System! Overall Trigger & DAQ architechture: 2 trigger levels! Level-1:! 3.2 µs latency! 100 khz output! DAQ/HLT! Event building at full L1 rate! Average output rate in 2012: 350 Hz prompt, 300Hz parked, ~1 khz in 2015! Average event size 1 MB in 2012, 2 Mb in 2015! Average CPU time few 100 ms DAQ and Trigger, Oct 6, 2014 20

The LHCb Trigger/DAQ System! Overall Trigger & DAQ architechture: 3 trigger levels! Level-0:! 4 µs latency! 1 MHz output! DAQ/HLT! L1: look displaced high p T tracks, output 70 khz! L2: full event reconstruction! Average output rate in 2012: 5 khz, 2015: 12.5 khz! Average event size 35 kb in DAQ and Trigger, Oct 6, 2014 2012, 60 kb in 2015 21

The ALICE Trigger/DAQ System! Alice has different constraints! Low rate: max 8 khz pb+pb! Very large events: > 40MB! Slow detector (TPC ~ 100 µs)! Overall Trigger & DAQ architecture: 4 trigger levels! 3 hardware-based trigger, 1 software-based:! L0 L2: 1.2, 6.5, 88 µs latency! L3: further rejection and data compression DAQ and Trigger, Oct 6, 2014 22

L1 Trigger in ATLAS! Calorimeter and muons only! Simple algorithms on reduced data granularity! Selection based on particle type, multiplicities and thresholds! Reject the bulk of uninteresting collisions DAQ and Trigger, Oct 6, 2014 23

ATLAS L1 calorimeter trigger! Example: ATLAS e/γ trigger! Sum energy in calorimeter cells into EM and hadronic towers! Loop over grid and search in 4x4 towers for a local maximum 1x2 (2x1): cluster! Can do something similar for other particles: jets, tau or sum the energy of all towers: missing E T DAQ and Trigger, Oct 6, 2014 24

CMS L1 muon trigger DAQ and Trigger, Oct 6, 2014 25

Central/Global Trigger! Now we have the information on the particle candidates found by L1 in the detector! We know type, location and E T /p T threshold passed! Can also look at topological information! E.g. lepton opposite ETmiss, invariant mass of 2 leptons! Need to decide if this event is of any interest to us! This needs to be made quickly L1 calorimeter L1 muon Central / Global Trigger L1 minimum bias DAQ and Trigger, Oct 6, 2014 26

HLT Example: Muon! Muons in CMS:! Reconstruct and fit tracks using only the muon system! Continue if sufficient p T! Combine tracker hits with muon system to improve p T measurement! Keep the event if p T is large enough! Muons in ATLAS:! At Level 2, using detector information from the region around the L1 muon candidate, assign muon p T based on fast look up tables! Extrapolate to the collision point and find the associated track! Is the muon isolated in the tracker, calorimeters?! Refine selection at L3 using offline-based reconstruction, recompute p T! More on HLT in next lecture DAQ and Trigger, Oct 6, 2014 27

Upgrades DAQ and Trigger, Oct 6, 2014 28

Long Shutdown! LHC data acquisition system backbones installed >5 years ago! Very stable running in last 3 years, better than we were hoping for! Current shutdown is occasion to! Upgrade core systems and review architectures! Introduce new technologies, retire obsolete ones! Follow changes on the detector side! Prepare for challenges of Run2 (and Run3) DAQ and Trigger, Oct 6, 2014 29

Long Shutdown! LHC data acquisition system backbones installed >5 years ago! Very stable running in last 3 years, better than we were hoping for! Current shutdown is occasion to! upgrade core systems and review architectures! introduce new technologies, retire obsolete ones! follow changes on the detector side! prepare for challenges of Run2 (and Run3) DAQ and Trigger, Oct 6, 2014 30

Pileup Issues! CMS Simulation: 300 GeV H ZZ eeµµ at various luminosities DAQ and Trigger, Oct 6, 2014 31

Run 2 challenges! Increased pileup! more complex events increased computing needs, affects trigger efficiency and rejection power! larger data size bandwidth and storage DAQ and Trigger, Oct 6, 2014 32

Upgrades for Run 2 ATLAS CMS! Merge L2 and L3 into a single HLT farm! preserve Region of Interest, but diluted the farm separation and fragmentation! increased flexibly, computing power efficiency! No architectural changes, but all network technologies replaced! Myrinet Ethernet! Ethernet Infiniband! Filebased event distribution in the farm! achieve full decoupling between DAQ and HLT DAQ and Trigger, Oct 6, 2014 33

LS2 and beyond Alice! 500 Hz 50 khz of PbPb interactions! Importance: physics with low S/B! Implies need to store 500 PB/ month (1 HI period)! Data volume reduction! Online full reconstruction! discard raw data! Combined DAQ/HLT/offline farm! COTS, FPGA and GPGPU LHCb! 1 MHz 40 MHz readout and event building trigger-less! trigger support for staged computing power deployment (only soft L0)! Need full event reconstruction with track finding and fitting plus particle identification to extract interesting event! On detector zero suppression radhard FPGA! 4 TB/s event building! 100 kb/event! On the long term, all experiments looking forward to significant increase in L1 trigger rate and bandwidth. Alice and LHCb will pioneer this path during LS2 34

Summary! Challenge to design efficient trigger/daq for LHC! Very large collision rates (up to 40 MHz)! Very large data volumes (tens of MBytes per collision)! Very large rejection factors needed (>10 5 )! Showed data acquisition used in LHC experiments! Introduction to basic functionality of trigger! We ll look in detail at the trigger aspects in the next lecture! That one will be less technical and more physics-oriented! DAQ and Trigger, Oct 6, 2014 35

Backup DAQ and Trigger, Oct 6, 2014 36

Trigger/DAQ parameters No.Levels! Level-0,1,2 a Event Readout HLT Out Trigger Rate (Hz) Size (Byte) Bandw.(GB/s) MB/s (Event/s) 4 Pb-Pb 500 5x10 7 25 1250 (10 2 ) p-p 10 3 2x10 6 200 (10 2 ) 3 LV-1 10 5 1.5x10 6 4.5 300 (2x10 2 ) LV-2 3x10 3 2 LV-1 10 5 10 6 100 ~1000 (10 2 ) 2 LV-0 10 6 3.5x10 4 35 70 (2x10 3 ) DAQ and Trigger, Oct 6, 2014 37

TDAQ comparison DAQ and Trigger, Oct 6, 2014 38

Data handling requirements DAQ and Trigger, Oct 6, 2014 39