D0RACE: Testbed Session Lee Lueking D0 Remote Analysis Workshop February 12, 2002.

Slides:



Advertisements
Similar presentations
GridPP July 2003Stefan StonjekSlide 1 SAM middleware components Stefan Stonjek University of Oxford 7 th GridPP Meeting 02 nd July 2003 Oxford.
Advertisements

4/2/2002HEP Globus Testing Request - Jae Yu x Participating in Globus Test-bed Activity for DØGrid UTA HEP group is playing a leading role in establishing.
CMS Applications Towards Requirements for Data Processing and Analysis on the Open Science Grid Greg Graham FNAL CD/CMS for OSG Deployment 16-Dec-2004.
February 12, 2002 DØRACE 1 Grid activities in NIKHEF Willem van Leeuwen.
EU-GRID Work Program Massimo Sgaravatto – INFN Padova Cristina Vistoli – INFN Cnaf as INFN members of the EU-GRID technical team.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
GridPP meeting Feb 03 R. Hughes-Jones Manchester WP7 Networking Richard Hughes-Jones.
The LHC Computing Grid Project Tomi Kauppi Timo Larjo.
Workload Management Massimo Sgaravatto INFN Padova.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
The D0 Monte Carlo Challenge Gregory E. Graham University of Maryland (for the D0 Collaboration) February 8, 2000 CHEP 2000.
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
High Energy Physics At OSCER A User Perspective OU Supercomputing Symposium 2003 Joel Snow, Langston U.
HEP Experiment Integration within GriPhyN/PPDG/iVDGL Rick Cavanaugh University of Florida DataTAG/WP4 Meeting 23 May, 2002.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London.
The Data Grid: Towards an Architecture for the Distributed Management and Analysis of Large Scientific Dataset Caitlin Minteer & Kelly Clynes.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
LCG and HEPiX Ian Bird LCG Project - CERN HEPiX - FNAL 25-Oct-2002.
WG Goals and Workplan We have a charter, we have a group of interested people…what are our plans? goalsOur goals should reflect what we have listed in.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
A Proposal of Application Failure Detection and Recovery in the Grid Marian Bubak 1,2, Tomasz Szepieniec 2, Marcin Radecki 2 1 Institute of Computer Science,
Jan. 17, 2002DØRAM Proposal DØRACE Meeting, Jae Yu 1 Proposal for a DØ Remote Analysis Model (DØRAM) IntroductionIntroduction Remote Analysis Station ArchitectureRemote.
SAM and D0 Grid Computing Igor Terekhov, FNAL/CD.
DataGrid is a project funded by the European Commission under contract IST rd EU Review – 19-20/02/2004 WP1 activity, achievements and plans.
Grid Workload Management Massimo Sgaravatto INFN Padova.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
DataTAG Research and Technological Development for a Transatlantic Grid Abstract Several major international Grid development projects are underway at.
7April 2000F Harris LHCb Software Workshop 1 LHCb planning on EU GRID activities (for discussion) F Harris.
26SEP03 2 nd SAR Workshop Oklahoma University Dick Greenwood Louisiana Tech University LaTech IAC Site Report.
1 Overview of IEPM-BW - Bandwidth Testing of Bulk Data Transfer Tools Connie Logg & Les Cottrell – SLAC/Stanford University Presented at the Internet 2.
16 September GridPP 5 th Collaboration Meeting D0&CDF SAM and The Grid Act I: Grid, Sam and Run II Rick St. Denis – Glasgow University Act II: Sam4CDF.
GridPP Presentation to AstroGrid 13 December 2001 Steve Lloyd Queen Mary University of London.
1 DØ Grid PP Plans – SAM, Grid, Ceiling Wax and Things Iain Bertram Lancaster University Monday 5 November 2001.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
HEPiX FNAL ‘02 25 th Oct 2002 Alan Silverman HEPiX Large Cluster SIG Report Alan Silverman 25 th October 2002 HEPiX 2002, FNAL.
 Load balancing is the process of distributing a workload evenly throughout a group or cluster of computers to maximize throughput.  This means that.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Grid Deployment Board – 10 February GD LCG Workshop Goals Give overview where we are Stimulate cooperation between the centres Improve the communication.
High Energy Physics and Grids at UF (Dec. 13, 2002)Paul Avery1 University of Florida High Energy Physics.
Connect. Communicate. Collaborate Click to edit Master title style PERT OPERATIONS.
HEPiX 2 nd Nov 2000 Alan Silverman Proposal to form a Large Cluster SIG Alan Silverman 2 nd Nov 2000 HEPiX – Jefferson Lab.
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
Feb. 14, 2002DØRAM Proposal DØ IB Meeting, Jae Yu 1 Proposal for a DØ Remote Analysis Model (DØRAM) Introduction Partial Workshop Results DØRAM Architecture.
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 24 January 2005.
Feb. 13, 2002DØRAM Proposal DØCPB Meeting, Jae Yu 1 Proposal for a DØ Remote Analysis Model (DØRAM) IntroductionIntroduction Partial Workshop ResultsPartial.
D0 File Replication PPDG SLAC File replication workshop 9/20/00 Vicky White.
DØRACE Workshop Agenda Feb. 11, 2002 (Morning – 1West) 8:30 Registration at High Rise next to 1 West Conference room 9:00 Welcome (John W.) - Session Chair:
DØRACE Workshop Agenda Feb. 11, 2002 (Morning – 1West) 8:30 Registration at High Rise next to 1 West Conference room 9:00 Welcome Womersley 9:10-9:30 DØRACE.
Activities and Perspectives at Armenian Grid site The 6th International Conference "Distributed Computing and Grid- technologies in Science and Education"
14 June 2001LHCb workshop at Bologna1 LHCb and Datagrid - Status and Planning F Harris(Oxford)
10-Feb-00 CERN HepCCC Grid Initiative ATLAS meeting – 16 February 2000 Les Robertson CERN/IT.
Grid Computing at NIKHEF Shipping High-Energy Physics data, be it simulated or measured, required strong national and trans-Atlantic.
Workload Management Workpackage
Clouds , Grids and Clusters
Distributed Data Access and Resource Management in the D0 SAM System
WP1 activity, achievements and plans
WP7 objectives, achievements and plans
US CMS Testbed.
Remote SAM Initiative (RSI) – Proposed Work Plan
Lee Lueking D0RACE February 28, 2002
Grid activities in NIKHEF
SAM Offsite Coordination
Status of Grids for HEP and HENP
STATEL an easy way to transfer data
Presentation transcript:

D0RACE: Testbed Session Lee Lueking D0 Remote Analysis Workshop February 12, 2002

Lee Lueking - D0 RACE2 Overview The network reliability and performance is of great importance to the D0 Data Model. D0 would like to be involved in the DataGrid/WP7 and DataTag studies to monitor and improve the network performance. DataGrid/WP7 covers the network within Europe whereas DataTag concentrates on the Europe-US intercontinental link. An initiative to enhance the performance of the connecting network in the US could be a result of this initiative. The network reliability and performance is of great importance to the D0 Data Model. D0 would like to be involved in the DataGrid/WP7 and DataTag studies to monitor and improve the network performance. DataGrid/WP7 covers the network within Europe whereas DataTag concentrates on the Europe-US intercontinental link. An initiative to enhance the performance of the connecting network in the US could be a result of this initiative. Until recently SAM used ftp and bbftp for the transport of files between the data storage location and the cache area of the computer used to process the files. Recently tests have started to make use of GridFTP together with the Grid Security Infrastructure GSI. This is particularly interesting because it will require a marriage between this security layer and the Kerberos security in operation at Fermilab. Until recently SAM used ftp and bbftp for the transport of files between the data storage location and the cache area of the computer used to process the files. Recently tests have started to make use of GridFTP together with the Grid Security Infrastructure GSI. This is particularly interesting because it will require a marriage between this security layer and the Kerberos security in operation at Fermilab. SAM can be used to select files and run production or analysis jobs on them. Recently an initiative has started to use Condor as a workload scheduler within SAM in order to make maximal use of the compute resources available at Fermilab and the participating institutes. SAM can be used to select files and run production or analysis jobs on them. Recently an initiative has started to use Condor as a workload scheduler within SAM in order to make maximal use of the compute resources available at Fermilab and the participating institutes. D0 has a request system for Monte Carlo generation of specific data channels. At this moment these requests are send to the (mostly external) Monte Carlo production sites by and submitted through human intervention. The goal is to evolve to an automatic system where the user requests are submitted to the full D0 Monte Carlo Compute Fabric consisting of all cpu resources within the collaboration. Most likely this services will be integrated within SAM. D0 has a request system for Monte Carlo generation of specific data channels. At this moment these requests are send to the (mostly external) Monte Carlo production sites by and submitted through human intervention. The goal is to evolve to an automatic system where the user requests are submitted to the full D0 Monte Carlo Compute Fabric consisting of all cpu resources within the collaboration. Most likely this services will be integrated within SAM. At this moment only limited use is made of the data storage capacity of institutes other than Fermilab. One of the difficulties has been the multitude of storage systems. At this moment Monte Carlo data is stored at SARA in Amsterdam and at the Computer Centre CCIN2P3 in Lyon but an effort will be made to integrate maximally all available storage locations within the collaboration for Monte Carlo generated data as well as analysis data. At this moment only limited use is made of the data storage capacity of institutes other than Fermilab. One of the difficulties has been the multitude of storage systems. At this moment Monte Carlo data is stored at SARA in Amsterdam and at the Computer Centre CCIN2P3 in Lyon but an effort will be made to integrate maximally all available storage locations within the collaboration for Monte Carlo generated data as well as analysis data. Planning Planning Most of the above projects have at least been discussed or have even been started at some level as part of the D0 PPDG and wider D0 Grid efforts. We estimate that the International Grid Testbed initiative will boost most if not all of these activities. The network monitoring and GridFTP tests have started on the European side but could be taken to a similar level on the US site still this year. Most of the above projects have at least been discussed or have even been started at some level as part of the D0 PPDG and wider D0 Grid efforts. We estimate that the International Grid Testbed initiative will boost most if not all of these activities. The network monitoring and GridFTP tests have started on the European side but could be taken to a similar level on the US site still this year. First contacts have been made with the Condor team and some initial tests have been done but more work is needed. The use of Condor as a workload scheduler within SAM will still take several months and the use of Condor to make use of grid cpu resources in the participating institutes in Europe will have about the same timescale. First contacts have been made with the Condor team and some initial tests have been done but more work is needed. The use of Condor as a workload scheduler within SAM will still take several months and the use of Condor to make use of grid cpu resources in the participating institutes in Europe will have about the same timescale. The distributed data storage can only proceed with the speed new storage locations become available but the present situation can still be largely improved. Within half a year it should be possible to store all externally produced data locally at the producing institutes. One year from now it should be possible to store any D0 data at a location where the grid (or SAM in the D0 case) decides the data can be stored best. The distributed data storage can only proceed with the speed new storage locations become available but the present situation can still be largely improved. Within half a year it should be possible to store all externally produced data locally at the producing institutes. One year from now it should be possible to store any D0 data at a location where the grid (or SAM in the D0 case) decides the data can be stored best. The network performance has to be increased such that the user will not notice the difference in use of files that are stored locally or not. Distributed compute resources within the collaboration should become more integrated just as the data storage systems. A workload scheduler should be able to make optimal use of all these resources. The network performance has to be increased such that the user will not notice the difference in use of files that are stored locally or not. Distributed compute resources within the collaboration should become more integrated just as the data storage systems. A workload scheduler should be able to make optimal use of all these resources. Management Management D0 is preparing a more detailed proposal for an International D0 Grid Testbed and the management will be described in there. It will have a small managerial board with people from the participating institutes in the US and Europe and it will have a technical board, which will address architectural issues and practical problems. The managerial board will have its representatives in the International DataGird Coordination meeting as well in the other appropriate bodies such as the PPDG. D0 is preparing a more detailed proposal for an International D0 Grid Testbed and the management will be described in there. It will have a small managerial board with people from the participating institutes in the US and Europe and it will have a technical board, which will address architectural issues and practical problems. The managerial board will have its representatives in the International DataGird Coordination meeting as well in the other appropriate bodies such as the PPDG.

February 12, 2002Lee Lueking - D0 RACE3 Who is interested Who is interested  Oklahoma U., IN2P3, Wuppertal, NIKHEF, UTA, Lancaster, Imperial, Prague, Micigan). Standard network testing procedure: Standard network testing procedure:  Netperf, iperf(Horst Severini, Shawn McKee ) performance.  Transmission rates from station logs as function of time.  Throughput numbers.  Measurements of error rates, packet and higher level.

February 12, 2002Lee Lueking - D0 RACE4 What are we trying to achieve? What are we trying to achieve?  Find bottlenecks, network understanding and debugging at various sites.  Understand scalability issues, operation of multiple sites.  End-to-end test  Transfers not only from FNAL but among all sites, or configured locations.  Break into specific tests, isolate components.Unit testing.  How to optimize caching  How can we do real work? Run reco or reco analyze at multiple sites. Simultainously.

February 12, 2002Lee Lueking - D0 RACE5 By March 15 By March 15 Iperf test  Single file, from cache, from enstore, from central-analysis cache  Project package running. Iain will set up Begin more complex tests and monitoring Begin more complex tests and monitoring  Touch base biweekly, at least in , maybe in d0grid meetings or other.

February 12, 2002Lee Lueking - D0 RACE6 Linux clusters: Clued0/sam combination. Roger Linux clusters: Clued0/sam combination. Roger How do we judge when a release is useful? Gordon How do we judge when a release is useful? Gordon Remote tasks: software shifts, moderating FAQ pages, web master. Remote tasks: software shifts, moderating FAQ pages, web master. Suggestion: Maintain a standard, operational, sam reference station for people to look at and see how things are configured. Suggestion: Maintain a standard, operational, sam reference station for people to look at and see how things are configured. Pass the torch. Encourage and help each other get things setup and running. Pass the torch. Encourage and help each other get things setup and running. Documentation needs to be kept up to date. Meena Documentation needs to be kept up to date. Meena

February 12, 2002Lee Lueking - D0 RACE7 All of these projects are working towards the common goal of providing transparent access to the massively distributed computing infrastructure that is needed to meet the challenges of modern experiments … (From the EU DataTAG proposal) All of these projects are working towards the common goal of providing transparent access to the massively distributed computing infrastructure that is needed to meet the challenges of modern experiments … (From the EU DataTAG proposal)

February 12, 2002Lee Lueking - D0 RACE8 Grid Projects Timeline Q3 00 Q4 00 Q4 01 Q3 01 Q2 01 Q1 01 Q1 02 GriPhyN: $11.9M+$1.6M PPDG:$9.5M iVDGL:$13.65M EU DataGrid: $9.3M EU DataTAG:4M Euros GridPP: