“Running Monte Carlo for the Fermi Telescope using the SLAC farm”

Slides:



Advertisements
Similar presentations
PanDA Integration with the SLAC Pipeline Torre Wenaus, BNL BigPanDA Workshop October 21, 2013.
Advertisements

9 v 2012Kavli IPMU1 Computational Astrophysics at the Kavli Institute for Particle Astrophysics and Cosmology at Stanford Roger Blandford.
Operations Testing in the ISOC Service Challenges are a successor and extension to the successful Data Challenge model Broader set of objectives Continue.
Batch Production and Monte Carlo + CDB work status Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
ProActive Task Manager Component for SEGL Parameter Sweeping Natalia Currle-Linde and Wasseim Alzouabi High Performance Computing Center Stuttgart (HLRS),
Experiences From The Fermi Data Archive Dr. Thomas Stephens Wyle IS/Fermi Science Support Center.
David Adams ATLAS DIAL Distributed Interactive Analysis of Large datasets David Adams BNL March 25, 2003 CHEP 2003 Data Analysis Environment and Visualization.
GLAST LAT ProjectISOC CDR, 4 August 2004 Document: LAT-PR-04500Section 3.11 GLAST Large Area Telescope: Instrument Science Operations Center CDR Section.
GLAST Large Area Telescope - OperationsDOE Review, 15 June GLAST Large Area Telescope Operations Review Rob Cameron Instrument Science Operations.
GLAST LAT ProjectISOC Peer Review - March 2, 2004 Document: LAT-PR Section 2.1 Requirements 1 Gamma-ray Large Area Space Telescope GLAST Large.
GLAST LAT Project ISOC Peer Review - March 2, 2004 Document: LAT-PR Section 4.3 Pipeline, Data Storage and Networking Issues1 Gamma-ray Large.
GLAST LAT ProjectOnline Peer Review – July 21, Integration and Test J. Panetta 1 Gamma-ray Large Area Space Telescope GLAST Large Area Telescope:
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
GLAST LAT Project ISOC Peer Review - March 2, 2004 Section Gamma-ray Large Area Space Telescope GLAST Large Area Telescope ISOC Peer Review 7.2.
MCTS Guide to Microsoft Windows Server 2008 Network Infrastructure Configuration Chapter 8 Introduction to Printers in a Windows Server 2008 Network.
Jun Peng Stanford University – Department of Civil and Environmental Engineering Nov 17, 2000 DISSERTATION PROPOSAL A Software Framework for Collaborative.
GLAST Large Area Telescope - OperationsDOE Review, 15 June GLAST Large Area Telescope Operations Review Rob Cameron Instrument Science Operations.
MCTS Guide to Microsoft Windows Server 2008 Network Infrastructure Configuration Chapter 11 Managing and Monitoring a Windows Server 2008 Network.
1 GLAST is now Fermi Gamma-ray Space Telescope First Light Introduction Dave Thompson NASA GSFC on behalf of the Fermi Mission Team see
Slide 1 of 9 Presenting 24x7 Scheduler The art of computer automation Press PageDown key or click to advance.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
GRID job tracking and monitoring Dmitry Rogozin Laboratory of Particle Physics, JINR 07/08/ /09/2006.
Grid Information Systems. Two grid information problems Two problems  Monitoring  Discovery We can use similar techniques for both.
KARMA with ProActive Parallel Suite 12/01/2009 Air France, Sophia Antipolis Solutions and Services for Accelerating your Applications.
Fermi Large Area Telescope (LAT) Integration and Test (I&T) Data Experience and Lessons Learned LSST Camera Workshop Brookhaven, March 2012 Tony Johnson.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
GLAST LAT ProjectDOE/NASA Baseline-Preliminary Design Review, January 8, 2002 K.Young 1 LAT Data Processing Facility Automatically process Level 0 data.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
COMP3019 Coursework: Introduction to GridSAM Steve Crouch School of Electronics and Computer Science.
8th November 2002Tim Adye1 BaBar Grid Tim Adye Particle Physics Department Rutherford Appleton Laboratory PP Grid Team Coseners House 8 th November 2002.
Wenjing Wu Computer Center, Institute of High Energy Physics Chinese Academy of Sciences, Beijing BOINC workshop 2013.
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
Overview The GLAST Data Handling Pipeline (the Pipeline) provides a uniform interface to the diverse data handling needs of the GLAST collaboration. Its.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
 Apache Airavata Architecture Overview Shameera Rathnayaka Graduate Assistant Science Gateways Group Indiana University 07/27/2015.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
GLAST Science Support CenterJuly, 2003 LAT Ground Software Workshop Status of the D1 (Event) and D2 (Spacecraft Data) Database Prototypes for DC1 Robert.
1 GLAST The Gamma-ray Large Area Space Telescope Status of the Mission F.Longo see
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
GLAST Ground Software CHEP 2007, Victoria, Daniel Flath 1 The GLAST Data Handling Pipeline “Pipeline II” Daniel Flath for the GLAST Collaboration Stanford.
Mission Status - S. Ritz 1 GLAST The Gamma-ray Large Area Space Telescope Mission Status DC2 Kickoff 1 March 2006 S. Ritz Gamma-ray Large Area Space Telescope.
GLAST Large Area Telescope – ISOC Overview and StatusDOE Annual Review, 13 June GLAST Large Area Telescope Instrument Science Operations Center.
GLAST Large Area Telescope LAT Flight Software System Checkout TRR Test Environment Sergio Maldonado FSW Test Team Lead Stanford Linear Accelerator Center.
Monitoring GLAST Operations: Building tracking displays for the LAT Instrument Science Operations Center By Christina Ketchum.
Simulation Production System Science Advisory Committee Meeting UW-Madison March 1 st -2 nd 2007 Juan Carlos Díaz Vélez.
Fermi Fermi (previously GLAST) Gamma-Ray Space Telescope Processing Pipeline and Data CatalogGamma-Ray Space Telescope Processing Pipeline and Data Catalog.
Web Based Data Quality Monitoring for the Fermi Gamma-Ray Space Telescope Tony Johnson
A Web Based Job Submission System for a Physics Computing Cluster David Jones IOP Particle Physics 2004 Birmingham 1.
GLAST LAT ProjectNovember 18, 2004 I&T Two Tower IRR 1 GLAST Large Area Telescope: Integration and Test Two Tower Integration Readiness Review SVAC Elliott.
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
Source catalog generation Aim: Build the LAT source catalog (1, 3, 5 years) Jean Ballet, CEA SaclaySLAC, 23 May 2005 Four main functions: Find unknown.
Workload Management Workpackage
GLAST Large Area Telescope Instrument Science Operations Center
Glast Collaboration Data Server and Data Catalog
Towards a CTA high-level science analysis framework
Existing Perl/Oracle Pipeline
GLAST Large Area Telescope
LAT Data Server Serve what?
GLAST Large Area Telescope
GLAST Large Area Telescope
Module 01 ETICS Overview ETICS Online Tutorials
GLAST Large Area Telescope
Gridifying the LHCb Monte Carlo production system
LO2 – Understand Computer Software
LAT Instrument Operations Center
Level 1 Processing Pipeline
GLAST Large Area Telescope Instrument Science Operations Center
Status and plans for bookkeeping system and production tools
Presentation transcript:

“Running Monte Carlo for the Fermi Telescope using the SLAC farm” Fermi Gamma-ray Space Telescope Fermi Large Area Telescope “Running Monte Carlo for the Fermi Telescope using the SLAC farm” Tony Johnson Stanford Linear Accelerator Center tonyj@slac.stanford.edu http://glast-ground.slac.stanford.edu/DataPortal

Contents Fermi Data Handling Overview Fermi Processing Pipeline Overview Using Pipeline for MC processing Future plans for Fermi pipeline

Launched 11 June 2008 – LAT activated 25 June

In Orbit: Single Events in the LAT The green crosses show the detected positions of the charged particles, the blue lines show the reconstructed track trajectories, and the yellow line shows the candidate gamma-ray estimated direction.  The red crosses show the detected energy depositions in the calorimeter.   0.25 CPU sec/event to reconstruct: downlink 500 Hz Each photon event independent of others

Fermi One Year All Sky Map

Fermi MISSION ELEMENTS LAT Instrument Science Operations Center (SLAC) • m sec Large Area Telescope & GBM GPS • • Telemetry 1 kbps - Fermi Spacecraft • TDRSS SN S & Ku DELTA 7920H • • - S - • GN • LAT Instrument Science Operations Center (SLAC) White Sands Schedules Mission Operations Center (MOC) Fermi Science Support Center HEASARC GRB Coordinates Network Schedules GBM Instrument Operations Center Alerts Data, Command Loads

Data Processing Flow at SLAC Downlink from Goddard Space Flight Center ~8/day 15 GB total daily Half-pipe Automatic launched as data arrives at SLAC Decode & repackage incoming data Split science data from telemetry data Level 1 Processing Full event reconstruction: factor ~x50 expansion on raw data! 750 GB/day To turn around data in required timeframe we typically use ~800 cores Data Quality Monitoring Transfer science summary files to Goddard Science Support Ctr - 200 MB/day Processing requirements ASP (Automated Science Processing) GRB and Flare detection Spectral analysis

ISOC Control Room All of the data processing and data quality monitoring can be done from the web No need for anyone in the control room, monitoring load shared globally

Monitoring Pipeline + Data Quality

Data Access

Pipeline Introduction Pipeline design goals Automated submission and monitoring of batch jobs Very high reliability Ability to define graph of jobs to be run Ability to parallelize processing tasks Ability to perform simple computations as part of job graph E.g. Compute how many parallel streams to create as a function of the number of events to be processed Ability to “Roll Back” jobs (whether successful or not) Capability to automatically compute sub-graph of jobs to rerun Maintain full history of all data processing Data catalog to keep track of all data products Web interface for monitoring jobs and submitting new tasks Plus command line client, and programmatic API

Pipeline and Data Catalog Components SLAC Oracle Command Client read-only Data Catalog Client Catalog Web Interface Client Data Portal Client Client SLAC SLAC Job Control Batch Farm Client Pipeline Server Line Mode Client IN2P3 Job Control Batch Farm Client Client INFN / Grid? Job Control Batch Farm Pipeline Web Interface Oracle Jython Engine read-only

Pipeline Task specification (XML)

Level 1 Task Specification Reconstruction Digitization

Pipeline Implementation Connection Pool Back-End Components Middle-Ware Tomcat Server Web Application HTTP Web Browser Front-End User Interfaces Oracle RDBMS Stored Procedures JMX RMI* Connection Pool Pipeline Server Interface MBean JMX Line-mode Client SQL TNS Data SQL JMX JConsole Monitor Scheduler RMI Job Control Service Thread Pool Batch Farms Apache/James Email POP3

Pipeline Web Interface

Pipeline Performance and Reliability x108

Pipeline for MC production Why use the pipeline for MC production? Automated submission of thousands of jobs Complete bookkeeping and access to log files from web Ability to easily rerun failed jobs (automatically or manually) Data catalog + download manager simplifies access to data products Useful facilities not directly pipeline related “xrootd” alternative file system which scales to large number of readers/writers much better than NFS GPLTOOLS: Set of python utilities for staging input/output from scratch disk to/from NFS/xrootd Fermi has used pipeline for all MC production 250 CPU-years of MC production in last two years

Pipeline for MC production

Pipeline + Data Catalog Plans We have already set up a second “non-Fermi” version of pipeline server and data catalog Currently starting to be used for EXO and AGIS EXO will use pipeline for data processing starting next summer AGIS will use pipeline for MC data processing We are working with Fermi Italian collaborators to create a job submission daemon to work with the Grid We would like to extend pipeline to be able to submit jobs to the “Cloud” Improved Data Catalog web interface with more “AJAX” features.