ideas for real time analysis for hl lhc using the cms daq
play

Ideas for Real-Time Analysis for HL-LHC using the CMS DAQ System - PowerPoint PPT Presentation

Ideas for Real-Time Analysis for HL-LHC using the CMS DAQ System Remigius K Mommsen Fermilab Disclaimer The idea of the L1 scou0ng originates from Emilio Meschi (CERN) This talk is based to a large extend on material presented by Hannes


  1. Ideas for Real-Time Analysis for HL-LHC 
 using the CMS DAQ System Remigius K Mommsen Fermilab

  2. Disclaimer The idea of the L1 scou0ng originates from Emilio Meschi (CERN) This talk is based to a large extend on material presented by Hannes Sakulin (CERN) 
 at CHEP 2019, Adelaide, Australia Any mistakes or misinterpreta0ons are mine 2

  3. All-new CMS for HL-LHC (2027 onwards) Barrel Calorimeters • new BE/FE electronics Muon Systems • ECAL: lower temperature • new DT/CSC BE/FE electronics • HCAL: partially new scintillator • GEM/RPC coverage in 1.5<| ƞ |<2.4 • Muon-tagging in 2.4<| ƞ |<3.0 Endcap Calorimeters • high granularity calorimeter • radiation tolerant scinitllator • 3D capability and timing MIP Timing Detector • 30-60 ps resolution L1 rate: 
 • coverage up to | ƞ | = 3.0 750 kHz HLT rate: 
 ~7.5 kHz Tracker Event size: 
 • radiation tolerant, high granularity, 7.5 MB low material budget • coverage up to | ƞ | = 3.8 300 TB/s 
 • track trigger at L1 @ 40 MHz 3

  4. CMS Trigger & DAQ — 2 Trigger Levels Only Phase 0 & 1 — 2008-24 Phase 2 — 2027- 40 MHz 40 MHz Digitizers LV1 Front end pipelines 100 kHz 750 kHz μ s 1.5 MB 
 7.5 MB 
 Event-builder nodes event size event size 0.15 TB/s 5.5 TB/s HLT sec 2 kHz 7.5 kHz Storage (pt5/tier 0) 4

  5. L1 Trigger for HL-LHC High resolu0on objects • Tracker track reconstruc0on in firmware • Vertex finding • Kalman filter muon 12 µs latency reconstruc0on • Displaced muons • High precision calorimetry • Par0cle flow reconstruc0on Topological algorithms including invariant/transverse mass cuts Machine learning algorithms Inter-bx algorithms 
 (limited to +/- 3 bx) 5

  6. What is Real-Time Analysis? Analyze events while the data is being taken • Par0al events with limited resolu0on • Full events with sub-op0mal calibra0ons • Much higher rate than possible with offline analysis • Stringent 0me constrains Store summary results for certain topologies at higher rate • E.g. low-mass di-jets, three-jet resonances, di-muons LHCb will does most analysis in “real-0me” • 2-step HLT selec0on • 2nd step is run aier calibra0ons have been done • Same physics quality as offline for most objects 6

  7. HLT Real-Lme Analysis Data Scou0ng at HLT used successfully in CMS since 2011 • Save HLT physics objects to disk Detectors • Perform offline analysis on these objects rather than on Digitizers offline reconstructed en00es LV1 Front end pipelines • No raw data is saved and no further reconstruc0on is performed for these events μ s • Typically 1-5 kHz of scou0ng data O(100 MB/s) Readout buffers Switching networks HLT Processor farms sec Tiny event 
 at higher rate 7

  8. L1 Trigger ScouLng Acquire L1 trigger data at full bunch crossing rate Detectors • No back pressure 40 MHz Level-1 
 Digitizers Trigger Scou0ng System • Drop data if system cannot 
 keep up with rate LV1 Front end pipelines μ s Analyze certain topologies at full rate • Real-0me analysis Readout buffers • Store 0ny event record Planned for HL-LHC Switching networks • Prototyping now • Tes0ng during run 3 HLT Processor farms sec 8

  9. Physics to Look at with L1 scouLng (non-exhausLve) Physics use case • Rare process • Difficult to select at Level-1 trigger - despite upgraded L1 trigger 
 (Available cuts give low efficiency at amributed rate budget) • Analysis is possible with resolu0on available at Level-1 • Scou0ng for new signal -> then point L1 trigger to it Several Physics channels iden0fied where L1 scou0ng could poten0ally make a difference 9

  10. Other uses for Level-1 Trigger scouLng Scou0ng provides invaluable diagnos0c and monitoring opportuni0es as well • BX-to-BX correla0ons available at all 0mes (cosmics, pre/post firing, etc.) • Real-0me heat maps to immediately spot problema0c channels • High-stat cross-check of algorithms (e.g. GT inputs/outputs) Per-bunch luminosity measurement using physics channels with high sta0s0cs Anomaly detec0on with deep-learning algorithms 10

  11. HL-LHC 40 MHz L1 ScouLng 
 Stageable Architecture

  12. 
 
 
 
 ScouLng system components Amached storage 
 long term Expect Xilinx Kintex Ultrascale+ based HW board to be commercially available Query- Infiniband HDR, 200 GbE op7cal based HPC Interconnect(s) Other 
 analysis 25 Gb/s GPU CPU Accel. from Input board • Zero suppression trigger distributed processing (MPI ?) • Pre-processing Input HW Feature Key-value 
 • Re-calibra0on 8x Features or 
 DB 
 store ? SoAware ZS 200 using ML DMA full events Short term medium Kintex Ultrascale+ Gbps storage 
 term PCIe 
 Kintex Ultrascale+ 2 min 
 NIC (mul7-bx possible) Same 
 Gen4 1-3 TB protocol 1 or 2 boards RAM ? 
 Distributed … NVRAM ? as in 
 (global) trigger I/O node … stream processing no back- 
 pressure 12

  13. Ingredients Trigger data captured directly from the Level-1 using spare outputs of the processing boards • Assuming same 16/25 Gbps serial op0cal links used for the Level-1 interconnects and using the same protocol Input hardware: PCIe boards with (modest) FPGA in 1U PC (I/O node) – (uGMT scou0ng uses KCU1500 [limited to 16 Gbps]) • Zero-suppression, local pre-processing (e.g. re-calibra0on using ML) in FPGA • DMA to host memory for short-term buffering (~2 min) • Baseline: eight op0cal inputs per board (PCIe Gen4 ~ 200Gbps over 16 lanes), one or two input boards per PC I/O nodes (CPU, GPU, other accelerators) use distributed algorithms to extract features while data are buffered in memory • 1-3 TB short-term buffer (e.g. NVRAM, could be cheaper with acceptable latency) • 200 Gbps low-latency interconnect (e.g. InfiniBand HDR or 200 GbE) • Interes0ng features and/or full “events” (mul0-bx possible) streamed over interconnect to global processing “farm” Distributed global stream processing and storage into “feature DB” • Organizes features in “searchable” data structures • Search-engine-like system op0mized for numerical data, medium term storage (e.g. key-value store) Analysis by query, analysis results to permanent storage 13

  14. L1 Trigger System 14

  15. 
 
 L1 Trigger System Scou0ng System I/O nodes Amached • Local processing storage 
 • Transient storage long term Trigger Query- Primi0ves Infiniband HDR, 200 GbE based HPC Interconnect(s) analysis Tracker Feature Calo DB 
 medium Muon term Distributed Global (global) stream Decision processing 14

  16. 
 
 L1 Trigger System Scou0ng System I/O nodes Amached • Local processing storage 
 • Transient storage long term Trigger Query- Primi0ves Infiniband HDR, 200 GbE based HPC Interconnect(s) analysis Tracker Feature Calo DB 
 medium Muon term Distributed Global Stage1: (global) 9 nodes 
 stream @ 200 Gbps Decision processing 14

  17. 
 
 L1 Trigger System Scou0ng System I/O nodes Amached • Local processing storage 
 • Transient storage long term Trigger Query- Primi0ves Infiniband HDR, 200 GbE based HPC Interconnect(s) analysis Tracker Feature Calo Stage 2: DB 
 add 28 nodes 
 medium Muon @ 200 Gbps term Distributed Global Stage1: (global) 9 nodes 
 stream @ 200 Gbps Decision processing 14

  18. 
 
 L1 Trigger System Scou0ng System I/O nodes Amached • Local processing storage 
 • Transient storage long term Trigger Query- Primi0ves Infiniband HDR, 200 GbE based HPC Interconnect(s) Stage 3: analysis Tracker add 98 nodes 
 @ 200 Gbps Feature Calo Stage 2: DB 
 add 28 nodes 
 medium Muon @ 200 Gbps term Distributed Global Stage1: (global) 9 nodes 
 stream @ 200 Gbps Decision processing 14

  19. 
 
 L1 Trigger System Scou0ng System I/O nodes Amached • Local processing storage 
 • Transient storage long term Stage 4: Trigger add 100's nodes 
 Query- Primi0ves @ 200 Gbps Infiniband HDR, 200 GbE based HPC Interconnect(s) Stage 3: analysis Tracker add 98 nodes 
 @ 200 Gbps Feature Calo Stage 2: DB 
 add 28 nodes 
 medium Muon @ 200 Gbps term Distributed Global Stage1: (global) 9 nodes 
 stream @ 200 Gbps Decision processing 14

  20. GMT scouLng prototype in Run 2

  21. Global Muon Trigger ScouLng in Run 2 When: Oct / Nov 2018 Types of runs: • 1 week of pp run • Large part of HI run Capture @ 40 MHz • Up to 8 final muon candidates • Up to 8 intermediate muon 
 candidates from barrel region • GMT adds bunch and orbit counters 
 40 MHz Scouting 
 Prototype System 16

  22. Global Muon Trigger (GMT) ScouLng Prototype 2x QSFP = 
 8x 10 Gbps op7cal PCIe Gen3x8 (2x) KCU 1500 
 Xilinx Kintex Ultrascale 115 10 Gb/s from GMT SoAware ZS 
 Infini 10 40 BZIP (1/2) DMA KCU1500 10/40 8x RAMdisk 
 (1/8) band Gbps Gbps RAM 
 RAID 
 Lustre Gbps mount firmware ZS 
 PCIe 
 disk 8 TB switch NIC NIC NIC (1/20) Gen3 8 GB/s max 800 MB/s max 100 MB/s max 50 MB/s Dell R720 Dell R720 1.1 TB/24 hour beam day 
 Controller PC 
 aier compression in pp @2E34 Firmware update & monitoring 17

Recommend


More recommend