Computing Strategy of the AMS-02 Experiment

Slides:



Advertisements
Similar presentations
Computing Strategy of the AMS-02 Experiment B. Shan 1 On behalf of AMS Collaboration 1 Beihang University, China CHEP 2015, Okinawa.
Advertisements

CMS ECAL Laser Monitoring System Toyoko J. Orimoto, California Institute of Technology, on behalf of the CMS ECAL Group 10th ICATPP Conference on Astroparticle,
CMS ECAL Laser Monitoring System Toyoko J. Orimoto, California Institute of Technology, on behalf of the CMS ECAL Group High-resolution, high-granularity.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
March 13thXXXXth RENCONTRES DE MORIOND 1 The Alpha Magnetic Spectrometer on the International Space Station Carmen Palomares CIEMAT (Madrid) On behalf.
Performance of AMS-02 on the International Space Station DESY Theory Workshop, Hamburg Melanie Heil Supported by the Carl-Zeiss Foundation.
Reconstruction and Analysis on Demand: A Success Story Christopher D. Jones Cornell University, USA.
SG Det IA 2007 oct29 Alain Blondel ECFA study groups BENE (in CARE) Int'l Scoping Study (ISS) Steering group (EU): A.Blondel (Geneva) P.Soler (Glasgow)
CMS Full Simulation for Run-2 M. Hildrith, V. Ivanchenko, D. Lange CHEP'15 1.
IFIN-HH LHCB GRID Activities Eduard Pauna Radu Stoica.
Exploiting the Grid to Simulate and Design the LHCb Experiment K Harrison 1, N Brook 2, G Patrick 3, E van Herwijnen 4, on behalf of the LHCb Grid Group.
WP2: Detector development Summary G. Pugliese INFN - Politecnico of Bari.
The Time-of-Flight system of the PAMELA experiment: in-flight performances. Rita Carbone INFN and University of Napoli RICAP ’07, Rome,
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
OO Software and Data Handling in AMS Computing in High Energy and Nuclear Physics Beijing, September 3-7, 2001 Vitali Choutko, Alexei Klimentov MIT, ETHZ.
Codeigniter is an open source web application. It occupies a very small amount of space in the memory and is most useful for developers who aim to develop.
Jiawen Zhang, IHEP, 2008, April 10, frascati Status of BEPCII/BESIII and Physics preparation Jiawen Zhang 2008/4/7—10 , PHIPSI08, Frascati.
C OMPUTER O RGANIZATION AND D ESIGN The Hardware/Software Interface 5 th Edition Chapter 1 Computer Abstractions and Technology Sections 1.5 – 1.11.
CHEP'07 September D0 data reprocessing on OSG Authors Andrew Baranovski (Fermilab) for B. Abbot, M. Diesburg, G. Garzoglio, T. Kurca, P. Mhashilkar.
GRAPES-3 ROOT Framework Pravata K Mohanty Tata Institute of Fundamental Research On behalf of the GRAPES-3 collaboration Workshop on Astroparticle Physics,
March 13thXXXXth RENCONTRES DE MORIOND 1 The Alpha Magnetic Spectrometer on the International Space Station Carmen Palomares CIEMAT (Madrid) On behalf.
We have developed a GUI-based user interface for Chandra data processing automation, data quality evaluation, and control of the system. This system, known.
Real data reconstruction A. De Caro (University and INFN of Salerno) CERN Building 29, December 9th, 2009ALICE TOF General meeting.
ATLAS Liquid Argon Calorimeter Monitoring & Data Quality Jessica Levêque Centre de Physique des Particules de Marseille ATLAS Liquid Argon Calorimeter.
LOGO PROOF system for parallel MPD event processing Gertsenberger K. V. Joint Institute for Nuclear Research, Dubna.
Tracker data quality monitoring based on event display M.S. Mennea – G. Zito University & INFN Bari - Italy.
Detector Simulation on Modern Processors Vectorization of Physics Models Philippe Canal, Soon Yung Jun (FNAL) John Apostolakis, Mihaly Novak, Sandro Wenzel.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
S. Guatelli, A. Mantero, J. Moscicki, M. G. Pia Geant4 medical simulations in a distributed computing environment 4th Workshop on Geant4 Bio-medical Developments.
LOGO Development of the distributed computing system for the MPD at the NICA collider, analytical estimations Mathematical Modeling and Computational Physics.
LHCb production experience with Geant4 LCG Applications Area Meeting October F.Ranjard/ CERN.
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
CBM ECAL simulation status Prokudin Mikhail ITEP.
The CMS CERN Analysis Facility (CAF) Peter Kreuzer (RWTH Aachen) - Stephen Gowdy (CERN), Jose Afonso Sanches (UERJ Brazil) on behalf.
Computing Strategy of the AMS-02 Experiment V. Choutko 1, B. Shan 2 A. Egorov 1, A. Eline 1 1 MIT, USA 2 Beihang University, China CHEP 2015, Okinawa.
Parallelization Strategies Laxmikant Kale. Overview OpenMP Strategies Need for adaptive strategies –Object migration based dynamic load balancing –Minimal.
ScECAL Beam FNAL Short summary & Introduction to analysis S. Uozumi Nov ScECAL meeting.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
Motivation: dynamic apps Rocket center applications: –exhibit irregular structure, dynamic behavior, and need adaptive control strategies. Geometries are.
Summary of User Requirements for Calibration and Alignment Database Magali Gruwé CERN PH/AIP ALICE Offline Week Alignment and Calibration Workshop February.
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Predrag Buncic CERN ALICE Status Report LHCC Referee Meeting 01/12/2015.
Susanna Guatelli Geant4 in a Distributed Computing Environment S. Guatelli 1, P. Mendez Lorenzo 2, J. Moscicki 2, M.G. Pia 1 1. INFN Genova, Italy, 2.
POCC activities AMS02 meeting at KSC October 11 th 2010.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Feb. 3, 2007IFC meeting1 Beam test report Ph. Bruel on behalf of the beam test working group Gamma-ray Large Area Space Telescope.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
ALICE Physics Data Challenge ’05 and LCG Service Challenge 3 Latchezar Betev / ALICE Geneva, 6 April 2005 LCG Storage Management Workshop.
Recent Evolution of the Offline Computing Model of the NOA Experiment Talk #200 Craig Group & Alec Habig CHEP 2015 Okinawa, Japan.
Virtualisation: status and plans Dag Toppe Larsen
Minerva Test Beam 2 Status Geoff Savage for the Test Beam 2 Team March 23, /23/2015AEM - Minerva Test Beam 2 Status1.
1 GlueX Software Oct. 21, 2004 D. Lawrence, JLab.
ALICE Offline Week – 22 Oct Visualization of embedding Matevz Tadel, CERN Adam Kisiel, Ohio State University.
Monthly video-conference, 18/12/2003 P.Hristov1 Preparation for physics data challenge'04 P.Hristov Alice monthly off-line video-conference December 18,
Evolution at CERN E. Da Riva1 CFD team supports CERN development 19 May 2011.
Use of FPGA for dataflow Filippo Costa ALICE O2 CERN
WP18, High-speed data recording Krzysztof Wrona, European XFEL
Evolution of Monitoring System for AMS Science Operation Center
Kilohertz Decision Making on Petabytes
Calibrating ALICE.
Cosmic-Rays Astrophysics with AMS-02
US CMS Testbed.
Measurements of Cosmic-Ray Lithium and Beryllium Isotopes
Computing Infrastructure for DAQ, DM and SC
Preparation of the CLAS12 First Experiment Status and Time-Line
Simulation in a Distributed Computing Environment
Hybrid Programming with OpenMP and MPI
Presentation transcript:

Computing Strategy of the AMS-02 Experiment V. Choutko1, B. Shan2 A. Egorov1, A. Eline1 1MIT, USA 2Beihang University, China CHEP 2015, Okinawa, April 13, 2015

Outline AMS Experiment AMS Data/Computing Model Parallelization of Reconstruction Software Parallelization of Simulation Software Production Supporting Software Summary

Z, P are measured independently by the Tracker, RICH, TOF and ECAL AMS: A TeV precision, multipurpose particle physics spectrometer in space. TRD Identify e+, e- Particles and nuclei are defined by their charge (Z) and energy (E ~ P) TOF Z, E TRD TOF Tracker RICH ECAL 1 2 7-8 3-4 9 5-6 Magnet ±Z Silicon Tracker Z, P RICH Z, E ECAL E of e+, e-, γ Z, P are measured independently by the Tracker, RICH, TOF and ECAL 3 3 3

AMS Experiment The Alpha Magnetic Spectrometer (AMS) is a high energy physics experiment on board the ISS, featured: TOF: Four Layers of Scintillators (120 ps); Tracker: Nine Layers of Silicon Detectors (10 μ); Gaseous TRD Detector: h/e Rejection O(102); Pb/Sc EM Calorimeter: h/e Rejection O(104); RICH Detector: Velocity Measurement O(10-4); Geometrical Acceptance: 0.5 m2sr; Number of ReadOut Channels: ≈200K TRD TOF Tracker RICH ECAL 1 2 7-8 3-4 9 5-6

AMS Experiment The Alpha Magnetic Spectrometer (AMS) is a high energy physics experiment on board the ISS, featured: TOF: Four Layers of Scintillators (120 ps); Tracker: Nine Layers of Silicon Detectors (10 μ); Gaseous TRD Detector: h/e Rejection O(102); Pb/Sc EM Calorimeter: h/e Rejection O(104); RICH Detector: Velocity Measurement O(10-4); Geometrical Acceptance: 0.5 m2sr; Number of ReadOut Channels: ≈200K

AMS Data Model Average 500 cosmic ray events per second Data collection rate ~1.2MB/s Transferred via relay satellite, Marshall Space Flight Center, to CERN Recording original electron signals Organized to one RUN every ¼ orbit (~23 minutes) First Production Runs 24/7 on freshly arrived data Initial data validation and indexing Produces Data Summary Files and Event Tags for fast events selection Requires ~ 80 CPU cores to cope with data rate Usually be available within 2 hours after flight data arrived used to produce various calibrations for the second production as well as quick performance evaluation. Second Production To use all the available calibrations, alignments, ancillary data from the ISS, and slow control data (temperatures, pressures, voltages) to produce physics analysis ready set of reconstructed data.

Remote Computing Facilities AMS Computing Model DATABASE ORACLE(PDBR) Frames Web Server Preproduction Production Server Production platform RAW AMS Cluster (PBS) CERN lxbatch (LSF) Remote Computing Facilities (Producer) ROOT(DST) Remote Centers Validator Uploader EOS, CASTOR Data Meta-data / Control CERN Facilities

Highlights in AMS Offline Software Parallelization of Reconstruction Parallelization of Simulation Production Supporting Software

Parallelization of Reconstruction Motivations To shorten the (elapsed) time of data reconstruction To increase productivity by using multi-core/hyper-thread enabled processors To ease the production jobs management; Tool: openmp No explicit thread programming Nearly no code change except few pragmas Possible to change thread number during execution

Sequential Procession

Parallel Procession

Parallel Reconstruction Performance (Xeon?)

Parallel Reconstruction Performance (Intel MIC)

Parallel Reconstruction Performance (Intel X5482)

Parallel Reconstruction Performance (Intel Old Xeon HT)

Parallel Reconstruction Performance (Intel i7 HT)

Parallel Reconstruction Performance (Efficiency)

Parallel Reconstruction Performance (Xeon Nehalem HT)

Parallel Reconstruction Performance (Elapsed time change)

Parallelization of Simulation Software Motivations To decrease required memory amount per simulation job To allow faster Monte-Carlo parameter/software tuning Minimal software change: Geant 4.10.1 MT model + openmp (icc 15) Thread ID, thread number, and core number are taken from Geant 4.10.1, instead of from openmp barrier, master, parallel from openmp do not work, so barriers are implemented by emulation All other openmp pragmas work

Parallelization of Simulation Software Memory optimization -- G4AllocatorPool modification Garbage Collection with preserving of n chunks  (1 by default) with modes: Fully automatic At the end of event Fast unsafe by name of G4Allocator at the end of event Decreased the memory consumption by factor 2 and more for long jobs (W.R.T. original G4AllocatorPool) Decreased the memory consumption by factor ?? W.R.T. sequential simulation job

Parallel Simulation Performance (Memory consumption)

Parallel Simulation Performance (Xeon?)

Parallel Simulation Performance (Intel MIC)

Production Supporting Software Fully automated production cycle for: Production in local (CERN) and remote computing centers Data reconstruction and Monte-Carlo simulation Local production Production server/monitor Remote computing centers Light-weight production platform

Local production management Centralized job management by production server/monitor All production database reading is through server Load balance between different kinds of computing resources CORBA-based communication between producers and server Dynamic thread number adjusting Graphical production monitoring

Light-weight production platform Fully-automated production cycle Job acquiring, submission, monitoring, validation, transferring, and (optional) scratching Easy to deploy Based on perl/python/sqlite3 Installation-free Customizable Batch system, storage area, transferring method, etc. Running at: JUROPA, CNAF, IN2P3, NLAA

Latest Full Data Reconstruction Second half of Feb. 2015 Performed at CERN, JUROPA, CNAF, ACAD. SINICA 41 months’ flight data Reconstruction completed essentially in two weeks 100 times data collection speed 10-5 failing rate

Latest Full Data Reconstruction (I will add one plot here to show the completing status of the pass6 reconstruction)

Summary Parallelization of the AMS reconstruction software allows to increase the productivity of the reconstruction program up to 45% using the HT enable processors as well as to short per run reconstruction time and calibration jobs execution time by factor of 10. Parallelization of the AMS simulation software allows to decrease the total memory consumption up to ??, which makes it much easier to run long simulation jobs for light nuclei. With production supporting software, data reconstruction rate reached 100 times of data collection.