J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April 2009 1 ATLAS Grid Activities Preparing for Data Analysis Jim Shank.

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

SLUO LHC Workshop, SLACJuly 16-17, Analysis Model, Resources, and Commissioning J. Cochran, ISU Caveat: for the purpose of estimating the needed.
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
The PanDA Distributed Production and Analysis System Torre Wenaus Brookhaven National Laboratory, USA ISGC 2008 Taipei, Taiwan April 9, 2008 Torre Wenaus.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
Stefano Belforte INFN Trieste 1 CMS SC4 etc. July 5, 2006 CMS Service Challenge 4 and beyond.
December 17th 2008RAL PPD Computing Christmas Lectures 11 ATLAS Distributed Computing Stephen Burke RAL.
US ATLAS Western Tier 2 Status and Plan Wei Yang ATLAS Physics Analysis Retreat SLAC March 5, 2007.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
Integration Program Update Rob Gardner US ATLAS Tier 3 Workshop OSG All LIGO.
Alexei Klimentov : ATLAS Computing CHEP March Prague Reprocessing LHC beam and cosmic ray data with the ATLAS distributed Production System.
ATLAS Metrics for CCRC’08 Database Milestones WLCG CCRC'08 Post-Mortem Workshop CERN, Geneva, Switzerland June 12-13, 2008 Alexandre Vaniachine.
ATLAS Database Operations Invited talk at the XXI International Symposium on Nuclear Electronics & Computing Varna, Bulgaria, September 2007 Alexandre.
F. Fassi, S. Cabrera, R. Vives, S. González de la Hoz, Á. Fernández, J. Sánchez, L. March, J. Salt, A. Lamas IFIC-CSIC-UV, Valencia, Spain Third EELA conference,
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
ATLAS: Heavier than Heaven? Roger Jones Lancaster University GridPP19 Ambleside 28 August 2007.
Nurcan Ozturk University of Texas at Arlington US ATLAS Transparent Distributed Facility Workshop University of North Carolina - March 4, 2008 A Distributed.
PROOF Farm preparation for Atlas FDR-1 Wensheng Deng, Tadashi Maeno, Sergey Panitkin, Robert Petkus, Ofer Rind, Torre Wenaus, Shuwei Ye BNL.
ATLAS Bulk Pre-stageing Tests Graeme Stewart University of Glasgow.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
PROOF tests at BNL Sergey Panitkin, Robert Petkus, Ofer Rind BNL May 28, 2008 Ann Arbor, MI.
The ATLAS Computing Model and USATLAS Tier-2/Tier-3 Meeting Shawn McKee University of Michigan Joint Techs, FNAL July 16 th, 2007.
Pavel Nevski DDM Workshop BNL, September 27, 2006 JOB DEFINITION as a part of Production.
U.S. ATLAS Facility Planning U.S. ATLAS Tier-2 & Tier-3 Meeting at SLAC 30 November 2007.
Victoria, Sept WLCG Collaboration Workshop1 ATLAS Dress Rehersals Kors Bos NIKHEF, Amsterdam.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
T3g software services Outline of the T3g Components R. Yoshida (ANL)
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
The ATLAS Computing & Analysis Model Roger Jones Lancaster University ATLAS UK 06 IPPP, 20/9/2006.
Distributed Analysis Tutorial Dietrich Liko. Overview  Three grid flavors in ATLAS EGEE OSG Nordugrid  Distributed Analysis Activities GANGA/LCG PANDA/OSG.
Finding Data in ATLAS. May 22, 2009Jack Cranshaw (ANL)2 Starting Point Questions What is the latest reprocessing of cosmics? Are there are any AOD produced.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
WLCG November Plan for shutdown and 2009 data-taking Kors Bos.
Acronyms GAS - Grid Acronym Soup, LCG - LHC Computing Project EGEE - Enabling Grids for E-sciencE.
Data Distribution Performance Hironori Ito Brookhaven National Laboratory.
ATLAS Computing Model Ghita Rahal CC-IN2P3 Tutorial Atlas CC, Lyon
ATLAS Computing: Experience from first data processing and analysis Workshop TYL’10.
Panda Monitoring, Job Information, Performance Collection Kaushik De (UT Arlington), Torre Wenaus (BNL) OSG All Hands Consortium Meeting March 3, 2008.
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
LHCb Computing 2015 Q3 Report Stefan Roiser LHCC Referees Meeting 1 December 2015.
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
Computing Operations Roadmap
University of Texas At Arlington Louisiana Tech University
Tier3(g,w) meeting at ANL ASC
Database Replication and Monitoring
Virtualization and Clouds ATLAS position
Data Challenge with the Grid in ATLAS
ATLAS activities in the IT cloud in April 2008
A full demonstration based on a “real” analysis scenario
Extended OSG client for WLCG
Bernd Panzer-Steindel, CERN/IT
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
PanDA in a Federated Environment
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
Readiness of ATLAS Computing - A personal view
Data Federation with Xrootd Wei Yang US ATLAS Computing Facility meeting Southern Methodist University, Oct 11-12, 2011.
Off-line & GRID Computing
ATLAS Sites Jamboree, CERN January, 2017
Simulation use cases for T2 in ALICE
Grid Computing 6th FCPPL Workshop
ATLAS DC2 & Continuous production
The ATLAS Computing Model
The LHCb Computing Data Challenge DC06
Presentation transcript:

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April ATLAS Grid Activities Preparing for Data Analysis Jim Shank

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April 2009 Overview lATLAS Monte Carlo produciton in 2008 lData (cosmic and single beam) in 2008 lProduction and Distributed Analysis (PandDA) system lSome features of the ATLAS Computing Model nAnalysis model for the US lDistributed Analysis Worldwide: Ganga/PanDA and Hammercloud + other readiness tests lTier 3 centers in the US 2

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April 2009 Beam Splash Event 3

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April 2009 First ATLAS Beam Events, 10 Sept Data Exports to T1s Throughput in MB/s Effect of concurrent data access from centralized transfers and user activity (overload of disk server) Number of errors 4 CERN Storage system overload. DDM worked. Subsequently we limited user access to the storage system.

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April 2009 December 2008 Reprocessing 5

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April 2009 PanDA production (Monte Carlo Simulation/Reconstruction) Grouped by Cloud = Tier 1 center + all it’s associated Tier 2 centers

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April 2009 U.S. Production in More than our share—indicates others not delivering their expected levels

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April DDM : Data Replication Datasets subscription intervals Data replication to Tier-2s US Tier2s ATLAS Beam and Cosmics data replication from CERN to Tier-1s and calibration Tier-2s. Sep-Nov 2008 BNL&AGLT2

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April DDM : Data replication between Tier-1s Functional Test. Tier-1-Tier-1s data replication status. FZK experienced problems with dCache. Data export is affected Tier-1-Tier-1s and prestaging data replication status. Data reprocessing. All Tier-1s Operational. Red : data transfer completion on 95% (data staging at CNAF)

Torre Wenaus, BNL 10 PanDA Overview Launched 8/05 by US ATLAS to achieve scalable data-driven WMS Designed for analysis as well as production Insulates users from distributed computing complexity Low entry threshold US ATLAS production since late ‘05 US analysis since Spring ’06 ATLAS-wide production since early ‘08 ATLAS-wide analysis still rolling out OSG WMS program since 9/06 Launched 8/05 by US ATLAS to achieve scalable data-driven WMS Designed for analysis as well as production Insulates users from distributed computing complexity Low entry threshold US ATLAS production since late ‘05 US analysis since Spring ’06 ATLAS-wide production since early ‘08 ATLAS-wide analysis still rolling out OSG WMS program since 9/06 Workload management system for Production ANd Distributed Analysis

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April Panda/pathena Users 4 million jobs in last 6 months 473 users in last 6 months 352 users in last 3 months 90 users in last month 271 users with >1000 jobs 96 users with >10000 jobs

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April 2009 ATLAS ANALYSIS 12

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April 2009 ATLAS Data Types lStill evolving… 13

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April 2009 ATLAS Analysis Data Flow 14

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April US US T2 sites Analysis Readiness Tests

Nurcan Ozturk 18 Ideas for a Stress Test (1)   Initiated by Jim Cochran (US ATLAS Analysis Support Group Chair).   Below is a summary of plans from Akira Shibata (March 10th).   Goal: Stress testing of the analysis queues in the Tier2 sites with analysis jobs as realistic as possible both in volume and quality. We would like to make sure that the Tier2 sites are ready to accept real data and analysis queues to analyze them.   Time scale: sometime near the end of May   Outline of this exercise:  To make this exercise more useful and interesting we will generate and simulate (Atlfast-II) a large amount of mixed sample at Tier2’s.  We are currently trying to define the jobs for this exercise and we expect this to be finalized after the BNL jamboree this week.  The mixed sample is a blind mix of all Standard Model processes, which we call "data" in this exercise.  For the one day stress test, we will invite people with existing analysis to try and analyze the data using Tier2 resources only.  We will compile a list of people who have the ability to participate.

Nurcan Ozturk 19 Ideas for a Stress Test (2)   Estimate of data volume: A very rough estimate of the data volume is 100M-1B events. Assuming 100kB/event (realistic considering no truth info and no trigger info), this sets an upper limit of 100TB in total (split among 5 Tier2’s). This is probably an upper-limit from the current availability of USER/GROUP disk on Tier2 (which is in addition to MC/DATA/PROD and CALIB disk).   Estimate of computing capability: There are "plenty" of machines assigned to analysis though the current load of analysis queue is rather low. The computing nodes are usually shared between production and analysis and typically configured with upper limit and priority. For example MWT2 has 1200 cores and setup to run analysis jobs with priority with an upper limit of 400 cores. If production jobs are not coming in, the number of running analysis jobs can exceed this limit.   Site configuration: Site configuration varies among the Tier2 sites. We will compile a table showing configuration of each analysis queue; direct reading versus local copying, xrootd versus dcache, etc. We will compare the performance of queues based on their configuration.

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April 2009 Four Types of Tier 3 Systems lT3gs nT3 with Grid Services Details in next slides lT3g nT3 with Grid Connectivity details in next slides lT3w nTier 3 Workstation  unclustered workstations...OSG, DQ2 client, root, etc lT3af nTier 3 system built into lab or university analysis facility 20

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April 2009 Conclusions lMonte Carlo Simulation/Reconstruction working well world wide with PanDA submission system lData reprocessing with PanDA working, but need further tests of file staging from tape. lAnalysis Model still evolving nIn the U.S., big emphasis on getting T3’s up and running nAnalysis stress test coming in May-June lReady for collision data in late

J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April 2009 Backup 24

Torre Wenaus, BNL 25 PanDA Operation T. Maeno Data management Data management ATLAS production Analysis

Torre Wenaus, BNL 26 PanDA Production Dataflow/Workflow

Torre Wenaus, BNL 27 Analysis with PanDA: pathena Tadashi Maeno Running the ATLAS software: Locally: athena PanDA: pathena --inDS --outDS Running the ATLAS software: Locally: athena PanDA: pathena --inDS --outDS Outputs can be sent to xrootd/PROOF farm, directly accessible for PROOF analysis