Presentation is loading. Please wait.

Presentation is loading. Please wait.

1 The PHENIX Experiment in the RHIC Run 7 Martin L. Purschke, Brookhaven National Laboratory for the PHENIX Collaboration RHIC from space Long Island,

Similar presentations


Presentation on theme: "1 The PHENIX Experiment in the RHIC Run 7 Martin L. Purschke, Brookhaven National Laboratory for the PHENIX Collaboration RHIC from space Long Island,"— Presentation transcript:

1 1 The PHENIX Experiment in the RHIC Run 7 Martin L. Purschke, Brookhaven National Laboratory for the PHENIX Collaboration RHIC from space Long Island, NY (with some references to Run 6...) ‏

2 2 Our best run ever! Or, much simpler: (Ok, the DOE continuing budget resolution cut down our running time to 13 cryo-weeks and we would have gotten > 1Pb of data else, but even so we outdid previous runs.) ‏

3 3 RHIC/PHENIX at a glance RHIC: 2 independent rings, one beam clockwise, the other counterclockwise sqrt(S NN )= 500GeV * Z/A ~200 GeV for Heavy Ions ~500 GeV for proton-proton (polarized) ‏ PHENIX: 4 spectrometer arms 15 Detector subsystems 500,000 detector channels Lots of readout electronics Uncompressed Event size typically 280 -220 - 110 KB for AuAu, CuCu, pp Data rate ~5KHz (Au+Au) ‏ Front-end data rate 0.5 - 1.1 GB/s Data Logging rate ~400MB/s, 700 MB/s max

4 4 TOF-W RXNP HBD MPC-N...and 4 new Detector Systems in Run 7

5 5 Building up to record speed Over the previous runs we have been adding improvements Had lighter systems, d+Au, p-p, Cu-Cu in the last runs, less of a challenge than 200GeV Au+Au Distributed data compression (run 4) ‏ Multi-Event buffering (run 5) ‏ Mostly consolidating the achievements/tuning/etc in run 6, also lots of improvements in operations (increased uptime) ‏ 10G Network upgrade in run 7, added Lvl2 filtering Ingredients: With increased luminosity, we saw the previously demonstrated 600++MB/s data rate in earnest for the first time.

6 6 Data Compression LZO algorithm New buffer with the compressed one as payload Add new buffer hdr buffer LZO Unpack Original uncompressed buffer restored This is what a file then looks like On readback: This is what a file normally looks like All this is handled completely in the I/O layer, the higher-level routines just receive a buffer as before. Found that the raw data are still gzip-compressible after zero-suppression and other data reduction techniques Introduced a compressed raw data format that supports a late-stage compression

7 7 Distributed Compression ATP SEB Gigabit Crossbar Switch To HPSS Event Builder The compression is handled in the “Assembly and Trigger Processors” (ATP’s) and can so be distributed over many CPU’s -- that was the breakthrough Buffer Box The Event builder has to cope with the uncompressed data flow, e.g. 600MB/s … 1200MB/s The buffer boxes and storage system see the compressed data stream, 350MB/s … 650MB/s Buffer Box

8 8 Multi-Event Buffering: DAQ Evolution PHENIX is a rare-event experiment, after all -- you don’t want to go down this path Without MEB

9 9 MEB: trigger delays by analog Memory trigger electronics needs to buy some time to make its decision done by storing the signal charge in an analog memory (AMU) ‏ Memory keeps the state of some 40us worth of bunch crossings Trigger decision arrives. FEM goes back a given number of analog memory cells and digitizes the contents of that memory location time Multi-Event buffering means to start the AMU sampling again while the current sample is still being digitized. Trigger busy released much earlier deadtime greatly reduced

10 10 The Multi-Event Buffering Effect

11 11 ~600 MB/s This shows the aggregated data rate from the DAQ to disk in a RHIC fill We are very proud of this performance... Decay of RHIC Luminosity Length of a DAQ run It's not the best, it's one where I was there... the best RHIC fill best went up to 650MB/s

12 12 Event statistics 5.7 Billion Events in ~650TB of data Run 6 pp – 6.8 Billion @200GeV 1 Billion @62GeV

13 13 Online Filtering and Reconstruction We ran Level-2 triggers in the ATP’s in so-called filter mode lvl2 triggers don't reject but fish out interesting events for priority reconstruction Filtered data were sent to IN2P3 in France where resources were available AND where the people most interested in the filtered dataset are ~10% of min bias data were sent to Vanderbilt University where Computing resources were available to reconstruct the data set, find problems in reconstruction, new detectors' software, make early DST's available, gear up for “real” production Valuable tool to get a reading how you are doing, as well as preliminary physics signals to check calibrations etc Used to refine our GRID file transfer procedures to “new” remote sites (not that much data volume transferred during this run, ~70TB – Run6 - 300TB)

14 14 Summary Very successful run, 650TB of data on tape despite short run due to DOE budget woes Can do > 600MB/s 4 new detector systems which still needed some “shakedown” Reached 5KHz event rate in Au-Au with larger event size successful filtering effort for priority reconstruction First iterations of min bias data production at a remote site (Vanderbilt University) ‏

15 15 Where we are w.r.t. others ATLAS CMS LHCb ALICE ~25~40 ~100 ~300 All in MB/s all approximate ~100 ~150 600 ~1250 400-600MB/s are not so Sci-Fi these days


Download ppt "1 The PHENIX Experiment in the RHIC Run 7 Martin L. Purschke, Brookhaven National Laboratory for the PHENIX Collaboration RHIC from space Long Island,"

Similar presentations


Ads by Google