Oxana Smirnova LCG/ATLAS/Lund August 27, 2002, EDG Retreat ATLAS-EDG Task Force status report.

Slides:



Advertisements
Similar presentations
DataTAG WP4 Meeting CNAF Jan 14, 2003 Interfacing AliEn and EDG 1/13 Stefano Bagnasco, INFN Torino Interfacing AliEn to EDG Stefano Bagnasco, INFN Torino.
Advertisements

EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks MyProxy and EGEE Ludek Matyska and Daniel.
16 Sep 2002F Harris GridPP Imperial College1 WP8 Status and Plans F Harris (Oxford/CERN)
Stephen Burke - WP8 Status - 9/5/2002 Partner Logo WP8 Status Stephen Burke, PPARC/RAL.
Stephen Burke - WP8 Status - 14/2/2002 Partner Logo WP8 Status Stephen Burke, PPARC/RAL.
Andrew McNab - Manchester HEP - 17 September 2002 Putting Existing Farms on the Testbed Manchester DZero/Atlas and BaBar farms are available via the Testbed.
EU 2nd Year Review – Jan – Title – n° 1 WP1 Speaker name (Speaker function and WP ) Presentation address e.g.
Réunion DataGrid France, Lyon, fév CMS test of EDG Testbed Production MC CMS Objectifs Résultats Conclusions et perspectives C. Charlot / LLR-École.
Job Submission The European DataGrid Project Team
The DataGrid Project NIKHEF, Wetenschappelijke Jaarvergadering, 19 December 2002
Oxana Smirnova LCG/ATLAS/Lund November 11, 2002, Uppsala 4th NorduGrid Workshop ATLAS Data Challenges on EDG.
Current Monte Carlo calculation activities in ATLAS (ATLAS Data Challenges) Oxana Smirnova LCG/ATLAS, Lund University SWEGRID Seminar (April 9, 2003, Uppsala)
K.Harrison CERN, 23rd October 2002 HOW TO COMMISSION A NEW CENTRE FOR LHCb PRODUCTION - Overview of LHCb distributed production system - Configuration.
The B A B AR G RID demonstrator Tim Adye, Roger Barlow, Alessandra Forti, Andrew McNab, David Smith What is BaBar? The BaBar detector is a High Energy.
C. Loomis – Testbed: Status… – Sep. 5, 2002 – 1 Testbed: Status & Plans Charles Loomis (CNRS) Sept. 5, th Project Conference (Budapest)
The ATLAS Production System. The Architecture ATLAS Production Database Eowyn Lexor Lexor-CondorG Oracle SQL queries Dulcinea NorduGrid Panda OSGLCG The.
JetWeb on the Grid Ben Waugh (UCL), GridPP6, What is JetWeb? How can JetWeb use the Grid? Progress report The Future Conclusions.
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
Don Quijote Data Management for the ATLAS Automatic Production System Miguel Branco – CERN ATC
WP9 – Earth Observation Applications – n° 1 Experiences with Testbed1, plans and objectives for Testbed 2 Testbed retreat th August 2002
3 Sept 2001F HARRIS CHEP, Beijing 1 Moving the LHCb Monte Carlo production system to the GRID D.Galli,U.Marconi,V.Vagnoni INFN Bologna N Brook Bristol.
GridPP CM, ICL 16 September 2002 Roger Jones. RWL Jones, Lancaster University EDG Integration  EDG decision to put short-term focus of effort on making.
8th November 2002Tim Adye1 BaBar Grid Tim Adye Particle Physics Department Rutherford Appleton Laboratory PP Grid Team Coseners House 8 th November 2002.
Nick Brook Current status Future Collaboration Plans Future UK plans.
Belle MC Production on Grid 2 nd Open Meeting of the SuperKEKB Collaboration Soft/Comp session 17 March, 2009 Hideyuki Nakazawa National Central University.
LHCb and DataGRID - the workplan for 2001 Eric van Herwijnen Wednesday, 28 march 2001.
WP8 Status – Stephen Burke – 30th January 2003 WP8 Status Stephen Burke (RAL) (with thanks to Frank Harris)
- Distributed Analysis (07may02 - USA Grid SW BNL) Distributed Processing Craig E. Tull HCG/NERSC/LBNL (US) ATLAS Grid Software.
The huge amount of resources available in the Grids, and the necessity to have the most up-to-date experimental software deployed in all the sites within.
5 Sep 2002F Harris Plenary Budapest1 WP8 Report F Harris (Oxford/CERN)
CMS Stress Test Report Marco Verlato (INFN-Padova) INFN-GRID Testbed Meeting 17 Gennaio 2003.
Oxana Smirnova LCG/ATLAS/Lund September 19, 2002, RHUL ATLAS Software Workshop ATLAS-EDG Task Force report.
Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.
Author - Title- Date - n° 1 Partner Logo WP5 Summary Paris John Gordon WP5 6th March 2002.
Quick Introduction to NorduGrid Oxana Smirnova 4 th Nordic LHC Workshop November 23, 2001, Stockholm.
First attempt for validating/testing Testbed 1 Globus and middleware services WP6 Meeting, December 2001 Flavia Donno, Marco Serra for IT and WPs.
David Adams ATLAS DIAL/ADA JDL and catalogs David Adams BNL December 4, 2003 ATLAS software workshop Production session CERN.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
Stephen Burke – Data Management - 3/9/02 Partner Logo Data Management Stephen Burke, PPARC/RAL Jeff Templon, NIKHEF.
AliEn AliEn at OSC The ALICE distributed computing environment by Bjørn S. Nilsen The Ohio State University.
2-Sep-02Steve Traylen, RAL WP6 Test Bed Report1 RAL and UK WP6 Test Bed Report Steve Traylen, WP6
Andrew McNab - Manchester HEP - 17 September 2002 UK Testbed Deployment Aim of this talk is to the answer the questions: –“How much of the Testbed has.
The ATLAS Cloud Model Simone Campana. LCG sites and ATLAS sites LCG counts almost 200 sites. –Almost all of them support the ATLAS VO. –The ATLAS production.
2 Sep 2002F Harris EDG/WP6 meeeting at Budapest LHC experiments use of EDG Testbed F Harris (Oxford/CERN)
Oxana Smirnova LCG/ATLAS/Lund September 3, 2002, Budapest 5 th EU DataGrid Conference ATLAS-EDG Task Force status report.
Data Management The European DataGrid Project Team
ATLAS-specific functionality in Ganga - Requirements for distributed analysis - ATLAS considerations - DIAL submission from Ganga - Graphical interfaces.
Data Management The European DataGrid Project Team
Testing the HEPCAL use cases J.J. Blaising, F. Harris, Andrea Sciabà GAG Meeting April,
Site Certification Process (Round Table) Fabio Hernandez IN2P3 Computing Center - Lyon October
Pavel Nevski DDM Workshop BNL, September 27, 2006 JOB DEFINITION as a part of Production.
15-Feb-02Steve Traylen, RAL WP6 Test Bed Report1 RAL/UK WP6 Test Bed Report Steve Traylen, WP6 PPGRID/RAL, UK
Application examples Oxana Smirnova (Lund, EPF) 3 rd NorduGrid Workshop, May23, 2002.
Overview of ATLAS Data Challenge Oxana Smirnova LCG/ATLAS, Lund University GAG monthly, February 28, 2003, CERN Strongly based on slides of Gilbert Poulard.
The DataGrid Project NIKHEF, Wetenschappelijke Jaarvergadering, 19 December 2002
David Adams ATLAS ATLAS Distributed Analysis (ADA) David Adams BNL December 5, 2003 ATLAS software workshop CERN.
Stephen Burke – Sysman meeting - 22/4/2002 Partner Logo The Testbed – A User View Stephen Burke, PPARC/RAL.
ATLAS Distributed Analysis DISTRIBUTED ANALYSIS JOBS WITH THE ATLAS PRODUCTION SYSTEM S. González D. Liko
WMS baseline issues in Atlas Miguel Branco Alessandro De Salvo Outline  The Atlas Production System  WMS baseline issues in Atlas.
Testbed: Status & Plans
The EDG Testbed Deployment Details
Work Package 9 – EO Applications
U.S. ATLAS Grid Production Experience
Moving the LHCb Monte Carlo production system to the GRID
INFN-GRID Workshop Bari, October, 26, 2004
Patricia Méndez Lorenzo ALICE Offline Week CERN, 13th July 2007
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
CRC exercises Not happy with the way the document for testbed architecture is progressing More a collection of contributions from the mware groups rather.
Stephen Burke, PPARC/RAL Jeff Templon, NIKHEF
ATLAS DC2 & Continuous production
Presentation transcript:

Oxana Smirnova LCG/ATLAS/Lund August 27, 2002, EDG Retreat ATLAS-EDG Task Force status report

ATLASEDG Jean-Jacques BlaisingLaura PeriniIngo Augustin Frederic BrochuGilbert PoulardStephen Burke Alessandro De SalvoAlois PutzerFrank Harris Michael GardnerDi QingBob Jones Luc GoossensDavid RebattoPeter Kunszt Marcus HardtZhongliang RenEmanuele Leonardi Roger JonesSilvia ResconiMario Reale Christos KanellopoulosOxana SmirnovaMarkus Schulz Guido NegriStan ThompsonJeffrey Templon Fairouz Ohlsson-MalekLuca Vaccarossa Steve O'Neale Members and sympathizers

 ATLAS is eager to use Grid tools for the Data Challenges ATLAS Data Challenges are already on the Grid (NorduGrid, iVDGL) The DC1/phase2 (to start in October) is expected to be done mostly using the Grid tools  By September 16 (ATLAS SW Week), we should try to evaluate the usability of EDG for the DC tasks  The task: to process 5 input partitions of the Dataset 2000 at the EDG Testbed + one non-EDG site (Karlsruhe) The short-term use case

Task description  Input: set of generated events as ROOT files (each input partition ca 1.8 GB, events); master copies are stored in CERN CASTOR  Processing: ATLAS detector simulation using a pre- installed software release Each input partition is processed by 20 jobs (5000 events each) Full simulation is applied only to filtered events, ca 450 per job A full event simulation takes ca 150 seconds per event on a 1GHz processor)  Output: simulated events are stored in ZEBRA files (ca 1 GB each output partition); an HBOOK histogram file and a log-file (stdout+stderr) are also produced  Total: 9 GB of input, 2000 CPU-hours of processing, 100 GB of output.

Execution of jobs  It is expected that we make full use of the Resource Broker functionality Data-driven job steering Best available resources otherwise  A job consists of the standard DC1 shell-script, very much the way it is done in a non-Grid world  A Job Definition Language is used to wrap up the job, specifying: The executable file (script) Input data Files to be retrieved manually by the user Optionally, other attributes (maxCPU, Rank etc)  Storage and registration of output files is a part of the job

What’s already there  ATLAS RPMs are distributed with the EDG tools to provide the ATLAS runtime environment  Validation of the ATLAS runtime environment by submitting a short (100 input events) DC1 job was done at several sites: CERN NIKHEF RAL CNAF  A fruitful cooperation between ATLAS users and EDG experts  The task force attributes: Mailing list Web-page

What’s almost there  Input file replication: apparently, not a trivial procedure, requiring: Launching a job to retrieve files from CASTOR to an SE (rfcp is not available on UI, GridFTP is not installed on CASTOR) Executing a full bunch of GDMP directives  Define the GDMP_CONFIG_FILE depending on user’s VO  Register files in the local catalogue  Check mutual subscription of SE’s  Publish the catalogue  Initiate replication  Theoretically, it works; practically, there’s always something broken along the chain (e.g., RC can not be updated or can not be found altogether)

What’s not there  Submission of long jobs A job exceeding approx. 20 minutes is being “lost” by the system: the returned status is wrong, result retrieval using the EDG tools is impossible A temporary solution: the production testbed has a “regular” RB for frequent submission of short jobs, and an “ATLAS” RB for long jobs (sees only CERN and Karlsruhe sites) This solution have not been thoroughly tested yet, and anyway it’s not feasible in a long run Bottom line: so far, no full-scale ATLAS job ran on EDG machines

Other problems  Installation of applications software should not be combined with the system installation  Assuming EDG will be able to handle ATLAS jobs and researches from 40-odd countries would like to use it – the procedure of accepting new CAs has to be streamlined  Documentation: so far, not particularly user-friendly and very technical. Things seem to be improving though  Information system: despite being LDAP-based, lacks hierarchy; very difficult to browse/search and retrieve relevant info. Reportedly, to be upgraded soon.  Information about file collections: no easy way to find out what is stored in any given SE, even in the local one (RC browser?)  A small suggestion: store JobIDs of each user’s jobs in her home area by default, so that even if LB goes down, this information does not disappear

Summary  ATLAS users are ready and waiting  As soon as one real job runs, the rest 99 are expected to follow smoothly Given the production Testbed’s CPU power, it will take just one day  Hopefully, this will happen before September 19  Next meeting of the Task Force: 9/11