U.S. ATLAS Grid Testbed Status & Plans Kaushik De University of Texas at Arlington PCAP Review LBNL, November 2002.

Slides:



Advertisements
Similar presentations
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Advertisements

Workload Management Workpackage Massimo Sgaravatto INFN Padova.
Experience with ATLAS Data Challenge Production on the U.S. Grid Testbed Kaushik De University of Texas at Arlington CHEP03 March 27, 2003.
Workload Management Massimo Sgaravatto INFN Padova.
Magda – Manager for grid-based data Wensheng Deng Physics Applications Software group Brookhaven National Laboratory.
Pacman in a Nutshell Saul Youssef Boston University.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
Grappa: Grid access portal for physics applications Shava Smallen Extreme! Computing Laboratory Department of Physics Indiana University.
XCAT Science Portal Status & Future Work July 15, 2002 Shava Smallen Extreme! Computing Laboratory Indiana University.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
ATLAS Data Challenge Production and U.S. Participation Kaushik De University of Texas at Arlington BNL Physics & Computing Meeting August 29, 2003.
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
HEP Experiment Integration within GriPhyN/PPDG/iVDGL Rick Cavanaugh University of Florida DataTAG/WP4 Meeting 23 May, 2002.
Slide 1 Experiences with NMI R2 Grids Software at Michigan Shawn McKee April 8, 2003 Internet2 Spring Meeting.
Pacman issues affect Core software Infrastructure Testbed Preparing demos Preparing interoperability tests This is a critical time for establishing good.
3 Sept 2001F HARRIS CHEP, Beijing 1 Moving the LHCb Monte Carlo production system to the GRID D.Galli,U.Marconi,V.Vagnoni INFN Bologna N Brook Bristol.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
Central Reconstruction System on the RHIC Linux Farm in Brookhaven Laboratory HEPIX - BNL October 19, 2004 Tomasz Wlodek - BNL.
Distributed Facilities for U.S. ATLAS Rob Gardner Indiana University PCAP Review of U.S. ATLAS Physics and Computing Project Argonne National Laboratory.
US-ATLAS Grid Efforts John Huth Harvard University Agency Review of LHC Computing Lawrence Berkeley Laboratory January 14-17, 2003.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
1 Use of SRMs in Earth System Grid Arie Shoshani Alex Sim Lawrence Berkeley National Laboratory.
13-May-2003Barcelona EDG Conference L.Perini 1 ATLAS Grid Planning ATLAS has used in “production mode” different Grids with simulation jobs –NorduGrid,
K. De UTA Grid Workshop April 2002 U.S. ATLAS Grid Testbed Workshop at UTA Introduction and Goals Kaushik De University of Texas at Arlington.
Grid Status - PPDG / Magda / pacman Torre Wenaus BNL U.S. ATLAS Physics and Computing Advisory Panel Review Argonne National Laboratory Oct 30, 2001.
Tier 1 Facility Status and Current Activities Rich Baker Brookhaven National Laboratory NSF/DOE Review of ATLAS Computing June 20, 2002.
PPDG and ATLAS Particle Physics Data Grid Ed May - ANL ATLAS Software Week LBNL May 12, 2000.
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
ANL/BNL Virtual Data Technologies in ATLAS Alexandre Vaniachine Pavel Nevski US-ATLAS Core/GRID software workshop Brookhaven National Laboratory May 6-7,
Grid Workload Management Massimo Sgaravatto INFN Padova.
U.S. ATLAS Grid Testbed Status and Plans Kaushik De University of Texas at Arlington DoE/NSF Mid-term Review NSF Headquarters, June 2002.
ATLAS Data Challenge Production Experience Kaushik De University of Texas at Arlington Oklahoma D0 SARS Meeting September 26, 2003.
MAGDA Roger Jones UCL 16 th December RWL Jones, Lancaster University MAGDA  Main authors: Wensheng Deng, Torre Wenaus Wensheng DengTorre WenausWensheng.
NOVA Networked Object-based EnVironment for Analysis P. Nevski, A. Vaniachine, T. Wenaus NOVA is a project to develop distributed object oriented physics.
US ATLAS Grid Projects Rob Gardner Indiana University Mid Year Review of US ATLAS Computing NSF Headquarters, Arlington VA June 20, 2002
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
Production Tools in ATLAS RWL Jones GridPP EB 24 th June 2003.
Virtual Batch Queues A Service Oriented View of “The Fabric” Rich Baker Brookhaven National Laboratory April 4, 2002.
ATLAS Data Challenges US ATLAS Physics & Computing ANL October 30th 2001 Gilbert Poulard CERN EP-ATC.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
Atlas Grid Status - part 1 Jennifer Schopf ANL U.S. ATLAS Physics and Computing Advisory Panel Review Argonne National Laboratory Oct 30, 2001.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
February 28, 2003Eric Hjort PDSF Status and Overview Eric Hjort, LBNL STAR Collaboration Meeting February 28, 2003.
Grid Production Experience in the ATLAS Experiment Horst Severini University of Oklahoma Kaushik De University of Texas at Arlington D0-SAR Workshop, LaTech.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
NOVA A Networked Object-Based EnVironment for Analysis “Framework Components for Distributed Computing” Pavel Nevski, Sasha Vanyashin, Torre Wenaus US.
August 30, 2002Jerry Gieraltowski Launching ATLAS Jobs to either the US-ATLAS or EDG Grids using GRAPPA Goal: Use GRAPPA to launch a job to one or more.
December 26, 2015 RHIC/USATLAS Grid Computing Facility Overview Dantong Yu Brookhaven National Lab.
U.S. ATLAS Computing Facilities Bruce G. Gibbard GDB Meeting 16 March 2005.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
K. Harrison CERN, 22nd September 2004 GANGA: ADA USER INTERFACE - Ganga release status - Job-Options Editor - Python support for AJDL - Job Builder - Python.
Magda Distributed Data Manager Prototype Torre Wenaus BNL September 2001.
Pavel Nevski DDM Workshop BNL, September 27, 2006 JOB DEFINITION as a part of Production.
Overview of ATLAS Data Challenge Oxana Smirnova LCG/ATLAS, Lund University GAG monthly, February 28, 2003, CERN Strongly based on slides of Gilbert Poulard.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Grid Status - PPDG / Magda / pacman Torre Wenaus BNL DOE/NSF Review of US LHC Software and Computing Fermilab Nov 29, 2001.
Gennaro Tortone, Sergio Fantinel – Bologna, LCG-EDT Monitoring Service DataTAG WP4 Monitoring Group DataTAG WP4 meeting Bologna –
David Adams ATLAS ATLAS Distributed Analysis (ADA) David Adams BNL December 5, 2003 ATLAS software workshop CERN.
Planning Session. ATLAS(-CMS) End-to-End Demo Kaushik De is the Demo Czar Need to put team together Atlfast production jobs –Atlfast may be unstable over.
10 March Andrey Grid Tools Working Prototype of Distributed Computing Infrastructure for Physics Analysis SUNY.
Magda Distributed Data Manager Torre Wenaus BNL October 2001.
Testbed Monitoring Kaushik De Univ. of Texas at Arlington
U.S. ATLAS Grid Production Experience
US CMS Testbed.
U.S. ATLAS Testbed Status Report
ATLAS DC2 & Continuous production
Presentation transcript:

U.S. ATLAS Grid Testbed Status & Plans Kaushik De University of Texas at Arlington PCAP Review LBNL, November 2002

November 14, 2002 Kaushik De PCAP Review 2 Outline  Testbed introduction  Recent accomplishments  Software distribution  Application toolkit  Monte Carlo production  Monitoring  Integration issues  Since summer 2002, U.S. Grid Testbed has concentrated on rapid software distribution, grid-based Monte Carlo production, and SC2002 demonstrations  This talk will focus primarily on the grid- based production experience

November 14, 2002 Kaushik De PCAP Review 3 Testbed Goals  Demonstrate success of grid computing model for High Energy Physics  in data production  in data access  in data analysis  Develop, deploy and test grid middleware and applications  integrate middleware with applications  simplify deployment - robust, rapid & scalable  inter-operate with other testbeds & grid organizations (iVDGL, DataTag…)  provide single point-of-service for grid users  Evolve into fully functioning scalable distributed tiered grid

November 14, 2002 Kaushik De PCAP Review 4 Lawrence Berkeley National Laboratory Brookhaven National Laboratory Indiana University Boston University Argonne National Laboratory U Michigan University of Texas at Arlington Oklahoma University US -ATLAS testbed launched February 2001 Two new sites joining soon - UNM, SMU Grid Testbed Sites

November 14, 2002 Kaushik De PCAP Review 5 Testbed Fabric  >8 production gatekeepers at ANL, BNL, LBNL, BU, IU, UM, OU, UTA   Large clusters at BNL, LBNL, IU, UTA, BU  Heterogeneous system - Condor, LSF, PBS  Currently > 100 nodes available  Could double capacity quickly, if needed  + Multiple R&D gatekeepers  - iVDGL GIIS  - ATLAS hierarchical GIIS  - EDG testing  - glue schema  - GRAT development  few sites - Grappa portal  bnl - VO server  few sites - iVDGL testbed

November 14, 2002 Kaushik De PCAP Review 6 Testbed Tools  Many tools developed by the U.S. ATLAS testbed group during past 2 years   GridView - simple tool to monitor status of testbed Kaushik De, Patrick McGuigan   Gripe - unified user accounts Rob Gardner   Magda - MAnager for Grid DAta Torre Wenaus, Wensheng Deng (see Gardner & Wenaus talks)   Pacman - package management and distribution tool Saul Youssef   Grappa - web portal using active notebook technology Shava Smallen, Dan Engh   GRAT - GRid Application Toolkit   Gridsearcher - MDS browser Jennifer Schopf   GridExpert - Knowledge Database Mark Sosebee   VO Toolkit - Site AA Rich Baker   This talk will not describe all tools - only some examples of tools used for production

November 14, 2002 Kaushik De PCAP Review 7 Recent Accomplishments  May 2002  Globus 2.0 beta RPM developed at BNL  Athena-atlfast grid package developed at UTA  Software installation - Pacman developed at BU  GRAT toolkit for job submission on grid, developed at UTA & OU  June 2002  Tested interoperability - successfully ran ATLAS MC jobs on CMS & D0 grid sites  ANL demonstrated that U.S. testbed package can run successfully at EDG sites  July 2002  New production software released & deployed  2 week Athena-atlfast MC production run using GRAT & GRAPPA  Generated 10 million events, thousand files catalogued in Magda, all sites participating

November 14, 2002 Kaushik De PCAP Review 8 Accomplishments contd.  August/September 2002  New production software package deployed  3 week dc1 production run using GRAT  Generated 200,000 events, using ~ 30,000 CPU hours, 2000 files, 100 GB storage  Primary files stored at BNL HPSS  Replicas stored at grid testbed sites  Deployed VO server at BNL Tier I facility  October/November 2002  Prepare demos for SC2002  Deploy new VDT on testbed  Deploy iVDGL packages Worldgrid, Sciencegrid  Interoperability tests with EDG  Prepare software for DC1 Phase 2

November 14, 2002 Kaushik De PCAP Review 9 Software Distribution  Goals:  Easy installation  Uniform software versions  Pacman perfect for this task  Deployment in 3 steps  VDT Package (root)  Pacman, Globus 2.0b, Condor, Pippy...  ATLAS Package (non-root)  Magda, Grappa, JDK, MySQL, Perl…  User Package  GRAT, Grappa  Experience  Repeated process 3 times during summer 2002  Worked well!  Deployed successfully at all testbed sites

November 14, 2002 Kaushik De PCAP Review 10 my_environment.pacman Pacman  Pacman - package manager developed at Boston University by Saul Youssef  Pacman lets you define how a mixed tarball/rpm/gpt/native environment is  Fetched  Installed  Setup  Updated  Experts can figure this out once and export to the rest of the world via trusted caches.  % pacman –get my_environment  Package examples: VDT, atlas, atlas_user, worldgrid, sciencegrid, dc1package  Pacmanization of CMS underway

November 14, 2002 Kaushik De PCAP Review 11 Installation Web Page

November 14, 2002 Kaushik De PCAP Review 12 Installation Status  Installation information also available through Pippy using Globus MDS

November 14, 2002 Kaushik De PCAP Review 13 User Toolkits

November 14, 2002 Kaushik De PCAP Review 14 Testbed Production  Goals:  Demonstrate distributed ATLAS data production, access and analysis using grid middleware and tools developed by the testbed group  Plans:  Atlfast production to test middleware and tools, and produce physics data for summer students, based on athena-atlfast fast MC  2 weeks to regenerate data, repeatable  deploy new tools and middleware quickly  move away from farm paradigm to grid model  very aggressive schedule - people limited!  DC1 production to test fabric capabilities, produce and access data, meet ATLAS deliverables, using atlsim, atrig and athena  not repeatable - but real test of grid software  increase U.S. participation - involve grid testbed  CPU intensive - ~14 hours per job/file

November 14, 2002 Kaushik De PCAP Review 15 Atlfast Production  Goal: demonstrate grid capabilities  July 1-15: Phase 0, 10^7 events  Globus 2.0 beta  GRAT and Grappa - using athena-atlfast 3.0.1, common disk model  Magda - data management, 1000 files  5 physics processes - http cached  Use Pythia, standard ATLAS physics group jobOptions  atlfast default  minbias  dijets 3 GeV < P t < 100 GeV  all jets  W + jets  Z (ll) + jets  processes repeat

November 14, 2002 Kaushik De PCAP Review 16 GRAT  GRid Applications Toolkit  Based on Globus, Magda & MySQL  Shell & Python scripts  Modular design  parts can be run independently  Rapid development platform  Test grid middleware, test grid performance  Useful as diagnostic tool  Includes > 50 command-line tools for  Athena-atlfast job submission (binary or afs)  Athena-atlfast production  dc1 production and data management  Modules can be easily enhanced or replaced by Condor-G, EDG resource broker, Chimera, replica catalogue… (mostly in progress)

November 14, 2002 Kaushik De PCAP Review 17 Database use in GRAT  MySQL databases play a central role in GRAT production scripts  Production database (Magda)  dc1jobstatus and dc1jobinfo tables  track job status, updated periodically by scripts  Data management (Magda)  input/output file transfers  file registration/catalogue  Virtual Data Catalogue (ANL)  job definition  job parameters, random numbers  Metadata catalogue (AMI)  post-production summary information  data provenance

November 14, 2002 Kaushik De PCAP Review 18 Atlfast Tools  Production toolkit examples:  run-atlfast-production  at [sites/all, # events, physics/all]  production-filename  generate base+date+events+physics+site  production-getvars  query gatekeeper for env variables  production-queues  return list of queues (static)  production-listqueues  MDS queried list of queues  production-primestore  primary magda cache location  production-replicastore  return replica location  checknodes  return available nodes (globus)  stats/stats2  magda file statistics

November 14, 2002 Kaushik De PCAP Review 19 Athena-atlfast Production Architecture Storage selection module Replica storage selection module Job scheduler module Software stage module Move files/cleanup module Execute Athena Job Query env module Filename module JobOptions module Physics Processes from Web GRAT User Node Gatekeeper Queue Node ATLAS_SCRATCH

November 14, 2002 Kaushik De PCAP Review 20 Production Status

November 14, 2002 Kaushik De PCAP Review 21 Lessons Learned  Globus, Magda and Pacman make grid production easy!  On the grid - submit anywhere, run anywhere, store data anywhere - really works!  Error reporting, recovery & cleanup very important - will always lose/hang some jobs  Found many unexpected limitations, hangs, software problems - next time, need larger team to quantify these problems and provide feedback to Globus, Condor, and other middleware teams  Large pool of hardware resources available on testbed: BNL Tier 1, LBNL (pdsf), IU & BU prototype Tier 2 sites, UTA (new $1.35M NSF- MRI), OU & UNM CS supercomputing clusters...  Testbed production effort suffering from severe shortage of human resources. Need people to debug middleware problems and provide feedback to middleware developers

November 14, 2002 Kaushik De PCAP Review 22 GRAT in DC1 Production  Important ATLAS deliverable  Main features:  Automatic job submission under full user control  One, many or infinite sequence of jobs  One or many sites (for scheduling of production)  Independent data management scripts to check consistency of production semi-automatically  query production database  check Globus for job completion status  check Magda for output files  recover from many possible production failures  New (and old) Magda features used:  moving and registering output files stored at BNL HPSS and at replica locations on the grid  tracking production status  Minimal requirement on grid production site  Globus & Magda installed on gatekeeper  shared $atlas_scratch disk on batch nodes

November 14, 2002 Kaushik De PCAP Review 23 DC1 samples reconst- ruction simulat-ion Total size TB GB Event size MB Number of events (1) 6.5 (2) x (1) 2.6 (2) x x Total size TB Total size GB Event size MB # of events L

November 14, 2002 Kaushik De PCAP Review 24 Atlsim, atlfast, athena-based reconstruction 5x10 7 events October 2003 thru March 2004 DC2 Atlsim + Athena > 10 7 events November ? DC1-p2 Atlsim (Geant) 10 7 events July 2002 thru August 2002 DC1-p1 Atlfast (Athena) 10 6 events April 2002 DC0TechnologyDataPeriod Challen ge DC1 Schedule

November 14, 2002 Kaushik De PCAP Review 25 GRAT DC1 Script Details  Main tools:  dc1-forever: simple opportunistic scheduler  continuously submits jobs to all sites  dc1-submit: submit multiple production jobs  submit one, many or range of partitions at one site  dc1-data-verify: check production consistency  usually run asynchronously to scheduler  dc1-list-*: generate status and statistics  many scripts to provide information  dc1-data-*: tools to manage data  Limitations found in dc1 phase 1:  Single scheduler < 100 jobs  Single gatekeeper < 50 jobs  major shutdowns required user intervention  1 TeV single particles generated - no input files required (cpu intensive: easier on the grid)

November 14, 2002 Kaushik De PCAP Review 26 GRAT Computing Infrastructure Gatekeeper GridFTP MDS Gatekeeper GridFTP MDS LocalRemote Batch Common Storage BQS Service to service communication

November 14, 2002 Kaushik De PCAP Review 27 GRAT Execution Model 1. Resource Discovery 2. Partition Selection 3. Job Creation 4. Pre-stage 5. Batch Submission 6. Job Parameterization 7. Simulation 8. Post-stage 9. Cataloging 10. Monitoring DC1 Prod. (UTA) Remote Gatekeeper Replica (local) MAGDA (BNL) Param (CERN) Batch Execution scratch 1,4,5,

November 14, 2002 Kaushik De PCAP Review 28 GRAT Job Scheduling Create job script module Replica storage select module Site select module Stage software on atlas_scratch Move files/cleanup module Execute Atlsim Job Query environment Partition select module GRAT Scheduler Gatekeeper Queue Node ATLAS_SCRATCH Virtual Data Catalogue Register Production Magda Database

November 14, 2002 Kaushik De PCAP Review 29 Production Status Checking usertools-0.7]$./dc1-list-jobs Checking the status of 123 globus jobs Active Pending Jobs at LBNL: 39 0 Jobs at OU : 8 0 Jobs at UTA : 24 6 usertools-0.7]$ condor_q -- Submitter: atlas000.uta.edu : ID OWNER ITTED RUN_TIME ST PRI SIZE CMD kaushik 9/8 12: :17:43 R dc1-run kaushik 9/8 12: :13:29 R dc1-run kaushik 9/8 12: :13:27 R dc1-run kaushik 9/8 13: :43:58 R dc1-run kaushik 9/8 13: :37:06 R dc1-run kaushik 9/8 13: :08:37 R dc1-run kaushik 9/9 14: :00:00 I dc1-run kaushik 9/9 14: :00:00 I dc1-run 30 jobs; 6 idle, 24 running, 0 held

November 14, 2002 Kaushik De PCAP Review 30 DC1 Production Statistics $./dc1-list-statistics Wed Nov 13 13:03:46 CST 2002 total number of partitions at BNL HPSS = 2002 total size of zebra files in HPSS = MB Files Size 1 TeV elec. (2107) M 1 TeV photon (2117) M 1 TeV photon (2127) M 1 TeV photon (2137) M Replica locations: UTA M Oklahoma M LBNL M Grid+BNL farm production = 2nd largest in DC1

November 14, 2002 Kaushik De PCAP Review 31 DC1 Statistics

November 14, 2002 Kaushik De PCAP Review 32 DC1 Production Experience  Anything that can go wrong, WILL go wrong  During 18 days of grid production (in August), every system died at least once  Local experts were not always be accessible (many of them on vacation)  Examples: scheduling machines died 5 times (thrice power failure, twice system hung)  Network outages - multiple times  Gatekeeper - died at every site at least 2-3 times  Three databases used - production, magda and virtual data. Each died at least once!  Scheduled maintenance - HPSS, Magda server, LBNL hardware, LBNL Raid array…  These outages should be expected on the grid - software design must be robust  We managed > 100 files/day (~80% efficiency) in spite of these problems!

November 14, 2002 Kaushik De PCAP Review 33 DC1 Phase 2  Pile up production:  plan to start now!  grid-based production using nodes  SUSY simulation - 50k events  Athena reconstruction  grid testbed will participate  store ESD/AOD at BNL, Ntuples on grid testbed  need grid deployable athena package!  Analysis / user access to data  magda_getfile already provides access to ~30k catalogued dc1 files from/to many grid locations (need ATLAS VO to make this universal)  testbed dc1 production database in magda with GRAT provides simple ‘discovery’ mechanism  long term: DIAL being developed at BNL

November 14, 2002 Kaushik De PCAP Review 34 Grid Monitoring  Monitoring is critically important in distributed Grid computing  check system health, debug problems  discover resources using static data  job scheduling and resource allocation decisions using dynamic data from MDS and other monitors  Testbed monitoring priorities  Discover site configuration  Discover software installation  Application monitoring  Grid status/operations monitoring  Visualization  Also need  Well defined data for job scheduling

November 14, 2002 Kaushik De PCAP Review 35 Monitoring - Back End  Publishing MDS information  Glue schema - BNL & UTA  Pippy - Pacman information service provider  BNL ACAS schema  Hierarchical GIIS server  Non-MDS back ends  iPerf, Netlogger, Prophesy, Ganglia  Archiving  MySQL  GridView, BNL ACAS  RRD  Network  Work needed  What to store?  Replication of archived information  Good progress on back end!

November 14, 2002 Kaushik De PCAP Review 36 PIPPY  Publishes information into MDS about software packages installed via Pacman.  Uses LDAP schema elements that describe a package.  Interfaces with GRIS backend of MDS service.  pippy creates two type of entries:  Software Collection  Software Package  Collection allows aggregation of packages.  Package describes contents of a pacman file.  Will be included in next release of VDT

November 14, 2002 Kaushik De PCAP Review 37 Architecture MDS (slapd) GRISGIIS pip.py Pacman.db InvocationLDIF.conf files

November 14, 2002 Kaushik De PCAP Review 38 Monitoring - Front End  MDS based  GridView, Gridsearcher, Gstat  Converting TeraGrid and other toolkits  Non-MDS  Cricket, Ganglia  Example:  Gstat - text version

November 14, 2002 Kaushik De PCAP Review 39 GridView - Old Text Version   Simple visualization tool using Globus Toolkit  First native Globus application for ATLAS grid (March 2001)  Collects information using Globus tools. Archival information is stored in MySQL server on a different machine. Data published through web server on a third machine.  

November 14, 2002 Kaushik De PCAP Review 40 GridView (NG)

November 14, 2002 Kaushik De PCAP Review 41 Fabric Testing

November 14, 2002 Kaushik De PCAP Review 42 MDS Browser

November 14, 2002 Kaushik De PCAP Review 43 Integration!!  Coordination with other grid efforts and software developers - very difficult task!  Project centric:  GriPhyN/iVDGL - Rob Gardner  PPDG/LCG - Torre Wenaus  EDG - Ed May, Jerry Gieraltowski  ATLAS/LHCb - Rich Baker  ATLAS/CMS - Kaushik De  ATLAS/D0 - Jae Yu  Fabric/Middleware centric:  Afs Software installations - Alex Undrus, Shane Canon, Iwona Sakrejda  Networking - Shawn McKee, Rob Gardner  Virtual and Real Data Management - Wendsheng Deng, Sasha Vaniachin, Pavel Nevski, David Malon, Rob Gardner, Dan Engh, Mike Wilde, Yong Zhao, Shava Smallen  Security/Site AA/VO - Rich Baker, Dantong Yu

November 14, 2002 Kaushik De PCAP Review 44 Summary  Testbed ideal for rapid testing of middleware, prototyping and testing of production software, and MC production  Extensive recent experience with grid deployment for DC & SC2002  U.S. Testbed and NorduGrid were the only sites to use grid for ATLAS DC1 Phase 1  Testbed provides large pool of hardware resources through the grid  Severely short of human resources - large list of tasks spanning many WBS items  Need testbed effort to provide feedback to middleware and application developers  Important DC milestones coming up!