ATLAS Distributed Analysis Dietrich Liko. Thanks to … pathena/PANDA: T. Maneo, T. Wenaus, K. De DQ2 end user tools: T. Maneo GANGA Core: U. Edege, J.

Slides:



Advertisements
Similar presentations
Metadata Progress GridPP18 20 March 2007 Mike Kenyon.
Advertisements

ATLAS Tier-3 in Geneva Szymon Gadomski, Uni GE at CSCS, November 2009 S. Gadomski, ”ATLAS T3 in Geneva", CSCS meeting, Nov 091 the Geneva ATLAS Tier-3.
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
Analysis demos from the experiments. Analysis demo session Introduction –General information and overview CMS demo (CRAB) –Georgia Karapostoli (Athens.
GANGA PANDA Dietrich Liko. Motivation Access to OSG resources by GANGA Collaboration with US colleagues Possibly an alternative way of submitting jobs.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
Ganga Developments Karl Harrison (University of Cambridge) 18th GridPP Meeting University of Glasgow, 20th-21st March 2007
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
Wahid Bhimji University of Edinburgh J. Cranshaw, P. van Gemmeren, D. Malon, R. D. Schaffer, and I. Vukotic On behalf of the ATLAS collaboration CHEP 2012.
December 17th 2008RAL PPD Computing Christmas Lectures 11 ATLAS Distributed Computing Stephen Burke RAL.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
David Adams ATLAS ATLAS Distributed Analysis David Adams BNL March 18, 2004 ATLAS Software Workshop Grid session.
BaBar Grid Computing Eleonora Luppi INFN and University of Ferrara - Italy.
The first year of LHC physics analysis using the GRID: Prospects from ATLAS Davide Costanzo University of Sheffield
ATLAS DIAL: Distributed Interactive Analysis of Large Datasets David Adams – BNL September 16, 2005 DOSAR meeting.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
F. Fassi, S. Cabrera, R. Vives, S. González de la Hoz, Á. Fernández, J. Sánchez, L. March, J. Salt, A. Lamas IFIC-CSIC-UV, Valencia, Spain Third EELA conference,
David Adams ATLAS ADA, ARDA and PPDG David Adams BNL June 28, 2004 PPDG Collaboration Meeting Williams Bay, Wisconsin.
Enabling Grids for E-sciencE System Analysis Working Group and Experiment Dashboard Julia Andreeva CERN Grid Operations Workshop – June, Stockholm.
ATLAS Data Challenges US ATLAS Physics & Computing ANL October 30th 2001 Gilbert Poulard CERN EP-ATC.
David Adams ATLAS DIAL/ADA JDL and catalogs David Adams BNL December 4, 2003 ATLAS software workshop Production session CERN.
Andrei Gheata, Mihaela Gheata, Andreas Morsch ALICE offline week, 5-9 July 2010.
Nurcan Ozturk University of Texas at Arlington US ATLAS Transparent Distributed Facility Workshop University of North Carolina - March 4, 2008 A Distributed.
ATLAS is a general-purpose particle physics experiment which will study topics including the origin of mass, the processes that allowed an excess of matter.
David Adams ATLAS ATLAS Distributed Analysis David Adams BNL September 30, 2004 CHEP2004 Track 5: Distributed Computing Systems and Experiences.
A PanDA Backend for the Ganga Analysis Interface J. Elmsheuser 1, D. Liko 2, T. Maeno 3, P. Nilsson 4, D.C. Vanderster 5, T. Wenaus 3, R. Walker 1 1: Ludwig-Maximilians-Universität.
PROOF Farm preparation for Atlas FDR-1 Wensheng Deng, Tadashi Maeno, Sergey Panitkin, Robert Petkus, Ofer Rind, Torre Wenaus, Shuwei Ye BNL.
D. Adams, D. Liko, K...Harrison, C. L. Tan ATLAS ATLAS Distributed Analysis: Current roadmap David Adams – DIAL/PPDG/BNL Dietrich Liko – ARDA/EGEE/CERN.
Post-DC2/Rome Production Kaushik De, Mark Sosebee University of Texas at Arlington U.S. Grid Phone Meeting July 13, 2005.
Alex Read, Dept. of Physics Grid Activities in Norway R-ECFA, Oslo, 15 May, 2009.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
David Adams ATLAS ATLAS Distributed Analysis: Overview David Adams BNL December 8, 2004 Distributed Analysis working group ATLAS software workshop.
ANALYSIS TOOLS FOR THE LHC EXPERIMENTS Dietrich Liko / CERN IT.
Using Ganga for physics analysis Karl Harrison (University of Cambridge) ATLAS Distributed Analysis Tutorial Milano, 5-6 February 2007
David Adams ATLAS ATLAS-ARDA strategy and priorities David Adams BNL October 21, 2004 ARDA Workshop.
ATLAS-specific functionality in Ganga - Requirements for distributed analysis - ATLAS considerations - DIAL submission from Ganga - Graphical interfaces.
MND review. Main directions of work  Development and support of the Experiment Dashboard Applications - Data management monitoring - Job processing monitoring.
GangaTasks Automating of job submission wit Athena/AthenaMC in Ganga sw week, johannes ebke, tariq mahmoud.
GangaTasks Automating of job submission wit Athena/AthenaMC in Ganga sw week, johannes ebke, tariq mahmoud.
TAGS in the Analysis Model Jack Cranshaw, Argonne National Lab September 10, 2009.
INFSO-RI Enabling Grids for E-sciencE Using of GANGA interface for Athena applications A. Zalite / PNPI.
The ATLAS Strategy for Distributed Analysis on several Grid Infrastructures D. Liko, IT/PSS for the ATLAS Distributed Analysis Community.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
CLRC Grid Team Glenn Patrick LHCb GRID Plans Glenn Patrick LHCb has formed a GRID technical working group to co-ordinate practical Grid.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
ATLAS Distributed Analysis Dietrich Liko IT/GD. Overview  Some problems trying to analyze Rome data on the grid Basics Metadata Data  Activities AMI.
Distributed Analysis Tutorial Dietrich Liko. Overview  Three grid flavors in ATLAS EGEE OSG Nordugrid  Distributed Analysis Activities GANGA/LCG PANDA/OSG.
David Adams ATLAS ATLAS Distributed Analysis (ADA) David Adams BNL December 5, 2003 ATLAS software workshop CERN.
Testing Infrastructure Wahid Bhimji Sam Skipsey Intro: what to test Existing testing frameworks A proposal.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Response of the ATLAS Spanish Tier2 for.
ATLAS Distributed Analysis DISTRIBUTED ANALYSIS JOBS WITH THE ATLAS PRODUCTION SYSTEM S. González D. Liko
Status of gLite-3.0 deployment and uptake Ian Bird CERN IT LCG-LHCC Referees Meeting 29 th January 2007.
David Adams ATLAS ADA: ATLAS Distributed Analysis David Adams BNL December 15, 2003 PPDG Collaboration Meeting LBL.
Analysis vs Storage 4 the LHC Experiments (… ATLAS biased view ) Alessandro Di Girolamo CERN / INFN.
WLCG November Plan for shutdown and 2009 data-taking Kors Bos.
ATLAS TIER3 in Valencia Santiago González de la Hoz IFIC – Instituto de Física Corpuscular (Valencia)
ATLAS Physics Analysis Framework James R. Catmore Lancaster University.
Acronyms GAS - Grid Acronym Soup, LCG - LHC Computing Project EGEE - Enabling Grids for E-sciencE.
Joe Foster 1 Two questions about datasets: –How do you find datasets with the processes, cuts, conditions you need for your analysis? –How do.
EXPERIENCE WITH ATLAS DISTRIBUTED ANALYSIS TOOLS S. González de la Hoz L. March IFIC, Instituto.
Seven things you should know about Ganga K. Harrison (University of Cambridge) Distributed Analysis Tutorial ATLAS Software & Computing Workshop, CERN,
Analysis Facility Infrastructure: ATLAS in Valencia S. González de la Hoz (On behalf of Tier3 Team) IFIC – Institut de Física Corpuscular de València First.
BaBar & Grid Eleonora Luppi for the BaBarGrid Group TB GRID Bologna 15 febbraio 2005.
ATLAS Distributed Analysis S. González de la Hoz 1, D. Liko 2, L. March 1 1 IFIC – Valencia 2 CERN.
Eleonora Luppi INFN and University of Ferrara - Italy
A full demonstration based on a “real” analysis scenario
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Readiness of ATLAS Computing - A personal view
Introduction to Distributed Analysis
Presentation transcript:

ATLAS Distributed Analysis Dietrich Liko

Thanks to … pathena/PANDA: T. Maneo, T. Wenaus, K. De DQ2 end user tools: T. Maneo GANGA Core: U. Edege, J. Mosckiki, K. Harrison, A. Murao, A. Soroko GANGA GUI C.L. Tan Athena AOD J. Elmshäuser User Production F. Brochu TAG Navigator C. Nicholson, M. Keynon EGEE/gLite: H.C. Lee, D. Liko Nordugrid B. Samset, K. Pajchet OSG/Panda D.Liko + PANDA team Dashbord Monitoring B. Gaidioz, S. Reddy, J. Wu AMI Integration F. Fassi, C.L. Tan + AMI team

ATLAS Analysis Model

Different scenarios Inside the Framework ◦ Athena ◦ AOD & ESD ◦ Grid based (GANGA, PANDA etc) Outside the framework ◦ ROOT ◦ DPD (ntuple) ◦ ROOT & PROOF Development aims to minimize the difference ◦ Migrate programs between ROOT and Athena ◦ AOD access from ROOT

Some recent activities pathena/PANDA ◦ See Tadashi’s talk GANGA ◦ EGEE ◦ Nordugrid ◦ PANDA (based on pathena) GANGA statistic and use cases

Migration to DQ2 0.3 The new release GANGA is available as of today. It supports DQ2 0.3 All users have to update their GANGA versions, as the changes of DQ2 are not backward compatible

EGEE – new features Support for very large sandboxes Support for group areas (EventViews) Support for LFC bulk operations AOD to ESD backnavigation Support for xrootd in Lyon and GridKa dCache readahead buffer Automatic patches for ROOT access to DPM and dCache More backends for AthenaMC

EGEE – gLite WMS At CERN we have started to use the gLite WMS ◦ It will be available also at sites in some time Bulk submission Better sandbox handling Will be aware of job priority settings at the site when available

Nordugrid upgraded to the latest ARC middleware, 0.6.0, distributed with made sure Athena-related code works with DQ2 0.3 done some work on ensuring proper use of credentials and backend commands, especially for systems without AFS made several internal changes for better job monitoring etc. Next Steps ◦ get DQ2 user dataset registration to work, using the new DQ2 tools ◦ connect NG backend jobs with arda dashboard

PANDA New interface to PANDA ◦ Newly developed ◦ Based on pathena Now it is possible to send a job to all ATLAS grids ◦ Not all options are supported yet … Can be the basis of further integration between the projects Will be available in the next release (4.3.6)

GANGA Users ATLAS: 435 users, 60 to 90 Users per week

GANGA ATLAS Jobs Data collected by ARDA dashboard

GANGA Jobs at T1 sites Data collected by ARDA dashboard

WMS Site efficiency at IN2P3

Job Distribution over the Tiers Data collected by ARDA dashboard

Where are the GANGA users?

Some examples of GANGA Usage Spain ◦ IFIC Valencia Italy ◦ INFN-Naples UK ◦ Lancaster & Manchester Germany ◦ Munich

IFIC Valencia Top Physics ◦ AANT Ntuples + ROOT Exotic Physics ◦ AOD + Athena ◦ HPTV Ntuples + ROOT Marcos Use GANGA ◦ AOD analysis ◦ Ntuple production Investigate StoRM with Lustre as disk based SE

INFN Naples

LMU Munich Production of Signal MC sample with AthenaMC additional events: ◦ event generation: 5 jobs or use already validated evgen samples where only a small fraction has officially been simulated/reconstructed ◦ simulation: 1000 jobs ◦ reconstruction: 50 jobs From this exercise a prototype for automatic job submission has evolved which will eventually will be part of Ganga Statistics from the dashboard: 11/4-11/6: 79% Grid eff. * 77% Application eff. = 53% overall eff. Distributed Analysis as part of the CSC homework: ◦ process signal and background MC samples at well known and maintained sites: GridKa, Lyon and LRZ ◦ This involves: SUSY signal, ttbar (5200), Di-boson backgrounds, etc. - Using SUSYView

UK Use Cases 1. GANGA FOR PHYSICS VALIDATION Require rapid validation of new releases as soon as they are deployed (usually reconstruction. EVGEN and DIGITS prepared for a variety of bench-mark channels some months back; AthenaMC used to reconstruct the digits using the new release/cache. AODs analyzed via Athena or sometimes copied directly to a local disk with DQ2 if there are a small number of files.Typical statistics: ~ events 2. GANGA FOR SMALL PRIVATE PRODUCTIONS Require small non-official samples for people developing new reconstruction code. Typical example: Eva Bouhova and Rob Henderson developing the V0 finder - this work has required several samples not included in the official production. Events generated, simulated and digitized using AthenaMC. The files are then copied to the local site with DQ2 so that they can be scanned using the reconstruction code under development Typical statistics: ~5000 events 3. GANGA FOR CSC PHYSICS ANALYSIS Running physics analysis code over CSC datasets with the aim of producing CSC notes GANGA used to submit analysis jobs on CSC AOD datasets; resulting ntuples merged and copied to local disks for analysis in ROOT. Typical statistics: events 4. GANGA FOR ANALYSIS TOOLS DEVELOPMENT 4. Using GANGA to test new tools for vertexing, tagging etc. GANGA used to submit analysis jobs on CSC AOD datasets, running new analysis tools; resulting ntuples merged and copied to local disks for analysis in ROOT. Typical statistics: events

Summary Distributed Analysis Use Cases in ATLAS are growing ◦ I was showing GANGA, you will hear soon about PANDA While we have two projects, we try to integrate them close and increase the collaboration Datamanagment and data distribution are the crucial aspects 2 nd week of July: GANGA Developer Days at CERN ◦ All developers will be locked in a room …