Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.

Slides:



Advertisements
Similar presentations
B A B AR and the GRID Roger Barlow for Fergus Wilson GridPP 13 5 th July 2005, Durham.
Advertisements

4/2/2002HEP Globus Testing Request - Jae Yu x Participating in Globus Test-bed Activity for DØGrid UTA HEP group is playing a leading role in establishing.
Amber Boehnlein, FNAL D0 Computing Model and Plans Amber Boehnlein D0 Financial Committee November 18, 2002.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Title US-CMS User Facilities Vivian O’Dell US CMS Physics Meeting May 18, 2001.
The D0 Monte Carlo Challenge Gregory E. Graham University of Maryland (for the D0 Collaboration) February 8, 2000 CHEP 2000.
High Energy Physics At OSCER A User Perspective OU Supercomputing Symposium 2003 Joel Snow, Langston U.
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
1 Data Management D0 Monte Carlo needs The NIKHEF D0 farm The data we produce The SAM data base The network Conclusions Kors Bos, NIKHEF, Amsterdam Fermilab,
The SAMGrid Data Handling System Outline:  What Is SAMGrid?  Use Cases for SAMGrid in Run II Experiments  Current Operational Load  Stress Testing.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Snapshot of the D0 Computing and Operations Planning Process Amber Boehnlein For the D0 Computing Planning Board.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
Distribution After Release Tool Natalia Ratnikova.
A Design for KCAF for CDF Experiment Kihyeon Cho (CHEP, Kyungpook National University) and Jysoo Lee (KISTI, Supercomputing Center) The International Workshop.
Jan. 17, 2002DØRAM Proposal DØRACE Meeting, Jae Yu 1 Proposal for a DØ Remote Analysis Model (DØRAM) IntroductionIntroduction Remote Analysis Station ArchitectureRemote.
DØ RAC Working Group Report Progress Definition of an RAC Services provided by an RAC Requirements of RAC Pilot RAC program Open Issues DØRACE Meeting.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
DØ RACE Introduction Current Status DØRAM Architecture Regional Analysis Centers Conclusions DØ Internal Computing Review May 9 – 10, 2002 Jae Yu.
International Workshop on HEP Data Grid Nov 9, 2002, KNU Data Storage, Network, Handling, and Clustering in CDF Korea group Intae Yu*, Junghyun Kim, Ilsung.
CDF Offline Production Farms Stephen Wolbers for the CDF Production Farms Group May 30, 2001.
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
7April 2000F Harris LHCb Software Workshop 1 LHCb planning on EU GRID activities (for discussion) F Harris.
Status of UTA IAC + RAC Jae Yu 3 rd DØSAR Workshop Apr. 7 – 9, 2004 Louisiana Tech. University.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Spending Plans and Schedule Jae Yu July 26, 2002.
D0RACE: Testbed Session Lee Lueking D0 Remote Analysis Workshop February 12, 2002.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
1 DØ Grid PP Plans – SAM, Grid, Ceiling Wax and Things Iain Bertram Lancaster University Monday 5 November 2001.
DØSAR a Regional Grid within DØ Jae Yu Univ. of Texas, Arlington THEGrid Workshop July 8 – 9, 2004 Univ. of Texas at Arlington.
Computing plans from UKDØ. Iain Bertram 8 November 2000.
Online Software 8-July-98 Commissioning Working Group DØ Workshop S. Fuess Objective: Define for you, the customers of the Online system, the products.
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
From DØ To ATLAS Jae Yu ATLAS Grid Test-Bed Workshop Apr. 4-6, 2002, UTA Introduction DØ-Grid & DØRACE DØ Progress UTA DØGrid Activities Conclusions.
GridPP11 Liverpool Sept04 SAMGrid GridPP11 Liverpool Sept 2004 Gavin Davies Imperial College London.
GRID activities in Wuppertal D0RACE Workshop Fermilab 02/14/2002 Christian Schmitt Wuppertal University Taking advantage of GRID software now.
DØRACE Workshop Jae Yu Feb.11, 2002 Fermilab Introduction DØRACE Progress Workshop Goals Arrangements.
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
Feb. 14, 2002DØRAM Proposal DØ IB Meeting, Jae Yu 1 Proposal for a DØ Remote Analysis Model (DØRAM) Introduction Partial Workshop Results DØRAM Architecture.
CD FY09 Tactical Plan Status FY09 Tactical Plan Status Report for Neutrino Program (MINOS, MINERvA, General) Margaret Votava April 21, 2009 Tactical plan.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
Frank Wuerthwein, UCSD Update on D0 and CDF computing models and experience Frank Wuerthwein UCSD For CDF and DO collaborations October 2 nd, 2003 Many.
Remote Institute Tasks Frank Filthaut 11 February 2002  Monte Carlo production  Algorithm development  Alignment, calibration  Data analysis  Data.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
MC Production in Canada Pierre Savard University of Toronto and TRIUMF IFC Meeting October 2003.
Feb. 13, 2002DØRAM Proposal DØCPB Meeting, Jae Yu 1 Proposal for a DØ Remote Analysis Model (DØRAM) IntroductionIntroduction Partial Workshop ResultsPartial.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
McFarm Improvements and Re-processing Integration D. Meyer for The UTA Team DØ SAR Workshop Oklahoma University 9/26 - 9/27/2003
ATLAS TIER3 in Valencia Santiago González de la Hoz IFIC – Instituto de Física Corpuscular (Valencia)
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
CDF SAM Deployment Status Doug Benjamin Duke University (for the CDF Data Handling Group)
Apr. 25, 2002Why DØRAC? DØRAC FTFM, Jae Yu 1 What do we want DØ Regional Analysis Centers (DØRAC) do? Why do we need a DØRAC? What do we want a DØRAC do?
July 10, 2002DØRACE Status Report, Jae Yu DØ IB Meeting, OU Workshop 1 DØ RACE Status Report Introduction Software Distribution (DØRACE Setup) Hardware.
Belle II Physics Analysis Center at TIFR
DØ MC and Data Processing on the Grid
DØ Internal Computing Review
DØ RAC Working Group Report
Proposal for a DØ Remote Analysis Model (DØRAM)
LHCb thinking on Regional Centres and Related activities (GRIDs)
Development of LHCb Computing Model F Harris
The LHCb Computing Data Challenge DC06
Presentation transcript:

Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University

24 May 2002Iain Bertram - Draft 1 2 Overview  Definitions /What is Remote Analysis  MC Production /Status /Future  Remote Analysis  Remote Code Development  DØ-Grid  Conclusions

24 May 2002Iain Bertram - Draft 1 3 Definitions  What is Remote? /We have defined it as follows: Local: d0mino, FNAL Farms, Online Remote: clued0, club, non-FNAL. /76 Institutions people on Masthead Most of these work remoteley

24 May 2002Iain Bertram - Draft Collaborators 78 Institutions 18 Countries  What is Remote? /We have defined it as follows: Local: d0mino, FNAL Farms, Online Remote: clued0, club, non-FNAL. /76 Institutions people on Masthead Most of these work remoteley

24 May 2002Iain Bertram - Draft 1 5 Goals  Remote Production Facilities will provide the bulk of processing power and storage for the DØ collaboration.  Production Tasks are: /Monte Carlo (MC) Production /Reconstruction and Secondary reprocessing of the data. /CPU intensive user Analysis Jobs. /Data storage  MC Production  Run IIB Data Rate – 50 Hz /Aim MC rate of 25 Hz

24 May 2002Iain Bertram - Draft 1 6 Monte Carlo Production - Current  All Production Monte Carlo Running Off-Site / Full Chain, generation, simulation, digitization, reconstruction, standard analysis are run. / A fully functional SAM station, capable of storing and retrieving data from any location. / MC_Runjob allows fully parameterized running of MC using macros / Redhat Linux 6.2 or 7.1 / Tarballs created from Official Releases / In process of implementing integrated request system using SAM (what do you mean by this?)

24 May 2002Iain Bertram - Draft 1 7 Monte Carlo Production Rates Upgrade to RH 7.1 Monte Carlo Production: Total 17.2 M Events Full Geant Simulation

24 May 2002Iain Bertram - Draft 1 8 MC Requirements RunIIB  Requirements for 25 Hz Monte Carlo Rate /25% Event full Geant - Rest Parameterized MC /Time Per Geant Event on 500 MHz PIII 216 seconds /Digitize, Reconstruct and Analyze each event 6 times /Time per Geant event: 550 seconds /Meet 25 Hz Goal requires 5k 500 MHz CPU's

24 May 2002Iain Bertram - Draft 1 9 MC Production Cost requirements  Same Cost Assumptions as Fermi Farms /Dual CPU Node: $2,500 /One I/O Node per 100 node's /No remote mass storage costs included /Three Year Purchase Profile

24 May 2002Iain Bertram - Draft 1 10 Software Requirements  Operating Systems: / Current RH 7.1 / Future: DØ Fermi Redhat Linux or a similar Redhat release / DØ will support official builds for the Fermi Redhat Releases / Remote production sites could be shared facilities / Will not be able to upgrade operating systems purely to meet DØ software requirements.

24 May 2002Iain Bertram - Draft 1 11 Remote Analysis  LHC and Grid Devlopments lead to many new large Computing Resources.  Need to access data software, and database information remote from FNAL  Integrate Software with new GRID Tools Propose to set up Regional Analysis Centers A series of locations that offer centralized access points for data analysis for remote users.

24 May 2002Iain Bertram - Draft 1 12 Progressive

24 May 2002Iain Bertram - Draft 1 13 DØRACE Deployment Map (US and EU only) Processing Center Analysis Site w/ SAM Analysis Site w/o SAM No DØRACE

24 May 2002Iain Bertram - Draft 1 14 Remote Analysis Centers (This picture needs to be changed. I’ve put in a new slide after this one) Central Analysis Center (CAC) Regional Analysis Centers RAC Institutional Analysis Centers Desktop Analysis Stations DAS Provide Various Services IAC ….…. Normal Interaction Communication Path Occasional Interaction Communication Path DAS

24 May 2002Iain Bertram - Draft 1 15 Central Analysis Center (CAC) Desktop Analysis Stations DAS …. DAS …. RAC …. Regional Analysis Centers Provide various services IAC …. Institutional Analysis Centers IAC …. IAC Normal Interaction Communication Path Occasional Interaction Communication Path Remote Analysis Centers

24 May 2002Iain Bertram - Draft 1 16 RAC Requirements  RAC will provide / Significant CPU Power / Access to complete DØ Thumbnails and some DSTs / Access to DØ Codes / Storage Space  An institute with large concentrated and available computing resources / 100s of CPUs / 10s of TBs of disk cache / 100s Mbytes of network bandwidth / Possibly equipped with Mass Storage

24 May 2002Iain Bertram - Draft 1 17 Costs for RAC  Assume Shared Facility  Costs Based on Lancaster Farm /Year 00: 200 CPU, 2TB disks, 30 TB expandable Mass Storage /Cost 600 k$. /Will need larger facility.....  Each RAC ~ 1M$ (For run IIa only) /Need to put better numbers in here,,, /Required disk storage is about 50TB for Run IIa alone. /The compute resource should be taken from the overall C&S plan document.

24 May 2002Iain Bertram - Draft 1 18 Remote Analysis Requirements  Off Site Data Reconstruction / Access the appropriate DØ calibration, and luminosity databases. It is assumed that these will be accessed via a server interface and not require a local copy of the database. / Transfer of relevant tier of data / Well synchronized reconstruction executable  Generic User Analysis Jobs / The user will be required to meet several mandatory conditions: / The software must be built using Dørte to ensure that the package can run without the full DØ release. / Any input/output data must be stored in SAM. (depending on the “user”. If the user is representing a physics group, this is correct but if the user is doing the analysis for his own analysis, this should not be the case.) / That the jobs can be fully described and submitted to a SAM queue.  Grid Enhancements / It is assumed that Remote Production Jobs will make full use of the current DØGrid project. Use of the Grid is not specific to remote analysis, but rather is a coherent part of the overall computing plan. (Remote analysis also is a coherent part of the computing plan. This statement sort of implicates that the remote analysis is not really in the overall plan…)

24 May 2002Iain Bertram - Draft 1 19 CLUB  Analysis Backend to Cluedo /For analyzing 1TB data samples /Similar to MC farm's /Essentially a batch engine. /Fermilab to provide seed and infrastructure. /This cluster is a RAC for IAC’s around Fermilab and for Fermilab physicist. Should be included in the pilot RAC program.  Assume 200 CPU system (100 nodes 250k$) /15 TB data storage (10 servers) (50k$) /Looks like 300k$ per CLUB type installation

24 May 2002Iain Bertram - Draft 1 20 Open Issues  Remote Mass Storage Costs.  Network Requirements /Depend on development of grid (Still there are basic requirements to just simply transfer data from CAC to RAC. ~14MB/sec aggregated at FNAL for Run IIa.) /Data to processors or program to data  Available Resources: Incomplete List Not Easy to Plan /Expect minimum 600 processors.

24 May 2002Iain Bertram - Draft 1 21 DØRAC Implementation Timescale  Implement First RAC by Oct. 1, 2002 /CLUBs cluster at FNAL and Karlsruhe, Germany /Cluster associated IAC’s /Transfer TMB (10kB/evt) data set constantly from CAC to the RACs  Workshop on RAC in Nov., 2002  Implement the next set of RAC by Apr. 1, 2003  Implement and test DØGridware as they become available  The DØGrid should work by the end of Run IIa (2004), retaining the DØRAM architecture  The next generation DØGrid, a truly gridfied network without

24 May 2002Iain Bertram - Draft 1 22 Pilot DØRAC Program  RAC Pilot sites: /Karlsruhe, Germany  Already agreed to do this /CLUBS, Fermilab  Need to verify  What should we accomplish? /Transfer TMB files as they get produced ¥A File server (both hardware and software) at CAC for this job •Request driven or constant push? ¥Network monitoring tools •To observe network occupation and stability –From CAC to RAC –From RAC to IAC /Allow IAC users to access the TMB ¥Observe •Use of the data set •Accessing pattern •Performance of the accessing system •SAM system performance for locating

24 May 2002Iain Bertram - Draft 1 23 ¥The user account assignment? ¥Resource (CPU and Disk space) need? /What are the needed Grid software functionality? ¥To interface with the users ¥To locate the input and necessary resources ¥To gauge the resources ¥To package the job requests