HENP Computing at BNL Torre Wenaus STAR Software and Computing Leader BNL RHIC & AGS Users Meeting Asilomar, CA October 21, 1999.

Slides:



Advertisements
Similar presentations
First results from the ATLAS experiment at the LHC
Advertisements

Resources for the ATLAS Offline Computing Basis for the Estimates ATLAS Distributed Computing Model Cost Estimates Present Status Sharing of Resources.
STAR C OMPUTING STAR Overview and OO Experience Torre Wenaus STAR Computing and Software Leader Brookhaven National Laboratory, USA CHEP 2000, Padova February.
Distributed IT Infrastructure for U.S. ATLAS Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Randall Sobie The ATLAS Experiment Randall Sobie Institute for Particle Physics University of Victoria Large Hadron Collider (LHC) at CERN Laboratory ATLAS.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Planning for a Western Analysis Facility Richard P. Mount Planning for a Western Analysis FacilityPage 1.
Assessment of Core Services provided to USLHC by OSG.
Hall D Online Data Acquisition CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental forces of nature. 75.
STAR C OMPUTING The STAR Databases: Objectivity and Post-Objectivity Torre Wenaus BNL ATLAS Computing Week CERN August 31, 1999.
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
25 February 2000Tim Adye1 Using an Object Oriented Database to Store BaBar's Terabytes Tim Adye Particle Physics Department Rutherford Appleton Laboratory.
CERN - European Laboratory for Particle Physics HEP Computer Farms Frédéric Hemmer CERN Information Technology Division Physics Data processing Group.
STAR C OMPUTING ROOT in STAR Torre Wenaus STAR Computing and Software Leader Brookhaven National Laboratory, USA ROOT 2000 Workshop, CERN February 3, 2000.
Event Metadata Records as a Testbed for Scalable Data Mining David Malon, Peter van Gemmeren (Argonne National Laboratory) At a data rate of 200 hertz,
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
Meeting the Data Protection Demands of a 24x7 Economy Steve Morihiro VP, Programs & Technology Quantum Storage Solutions Group
9/16/2000Ian Bird/JLAB1 Planning for JLAB Computational Resources Ian Bird.
Central Reconstruction System on the RHIC Linux Farm in Brookhaven Laboratory HEPIX - BNL October 19, 2004 Tomasz Wlodek - BNL.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
The Cluster Computing Project Robert L. Tureman Paul D. Camp Community College.
Grid Status - PPDG / Magda / pacman Torre Wenaus BNL U.S. ATLAS Physics and Computing Advisory Panel Review Argonne National Laboratory Oct 30, 2001.
Chapter 4 Realtime Widely Distributed Instrumention System.
LHC Computing Review - Resources ATLAS Resource Issues John Huth Harvard University.
PPDG and ATLAS Particle Physics Data Grid Ed May - ANL ATLAS Software Week LBNL May 12, 2000.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
LHC Computing Plans Scale of the challenge Computing model Resource estimates Financial implications Plans in Canada.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Developing & Managing A Large Linux Farm – The Brookhaven Experience CHEP2004 – Interlaken September 27, 2004 Tomasz Wlodek - BNL.
5 May 98 1 Jürgen Knobloch Computing Planning for ATLAS ATLAS Software Week 5 May 1998 Jürgen Knobloch Slides also on:
NOVA Networked Object-based EnVironment for Analysis P. Nevski, A. Vaniachine, T. Wenaus NOVA is a project to develop distributed object oriented physics.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
U.S. ATLAS Tier 1 Planning Rich Baker Brookhaven National Laboratory US ATLAS Computing Advisory Panel Meeting Argonne National Laboratory October 30-31,
STAR Off-line Computing Capabilities at LBNL/NERSC Doug Olson, LBNL STAR Collaboration Meeting 2 August 1999, BNL.
BNL Wide Area Data Transfer for RHIC & ATLAS: Experience and Plans Bruce G. Gibbard CHEP 2006 Mumbai, India.
US ATLAS Tier 1 Facility Rich Baker Brookhaven National Laboratory DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National Laboratory.
Test Results of the EuroStore Mass Storage System Ingo Augustin CERNIT-PDP/DM Padova.
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
1D. Olson, SDM-ISIC Mtg, 26 Mar 2002 Scientific Data Management: An Incomplete Experimental HENP Perspective D. Olson, LBNL 26 March 2002 SDM-ISIC Meeting.
ATLAS WAN Requirements at BNL Slides Extracted From Presentation Given By Bruce G. Gibbard 13 December 2004.
NOVA A Networked Object-Based EnVironment for Analysis “Framework Components for Distributed Computing” Pavel Nevski, Sasha Vanyashin, Torre Wenaus US.
STAR C OMPUTING Plans for Production Use of Grand Challenge Software in STAR Torre Wenaus BNL Grand Challenge Meeting LBNL 10/23/98.
Globus and PlanetLab Resource Management Solutions Compared M. Ripeanu, M. Bowman, J. Chase, I. Foster, M. Milenkovic Presented by Dionysis Logothetis.
STAR C OMPUTING The STAR Databases: From Objectivity to ROOT+MySQL Torre Wenaus BNL ATLAS Computing Week CERN August 31, 1999.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
U.S. ATLAS Computing Facilities Overview Bruce G. Gibbard Brookhaven National Laboratory U.S. LHC Software and Computing Review Brookhaven National Laboratory.
Tier 1 at Brookhaven (US / ATLAS) Bruce G. Gibbard LCG Workshop CERN March 2004.
PDSF and the Alvarez Clusters Presented by Shane Canon, NERSC/PDSF
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
US ATLAS Tier 1 Facility Rich Baker Deputy Director US ATLAS Computing Facilities October 26, 2000.
U.S. ATLAS Computing Facilities DOE/NFS Review of US LHC Software & Computing Projects Bruce G. Gibbard, BNL January 2000.
Ian Bird Overview Board; CERN, 8 th March 2013 March 6, 2013
U.S. ATLAS Computing Facilities U.S. ATLAS Physics & Computing Review Bruce G. Gibbard, BNL January 2000.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Grid Status - PPDG / Magda / pacman Torre Wenaus BNL DOE/NSF Review of US LHC Software and Computing Fermilab Nov 29, 2001.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
1 Particle Physics Data Grid (PPDG) project Les Cottrell – SLAC Presented at the NGI workshop, Berkeley, 7/21/99.
Grid technologies for large-scale projects N. S. Astakhov, A. S. Baginyan, S. D. Belov, A. G. Dolbilov, A. O. Golunov, I. N. Gorbunov, N. I. Gromova, I.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
Hall D Computing Facilities Ian Bird 16 March 2001.
Clouds , Grids and Clusters
Nuclear Physics Data Management Needs Bruce G. Gibbard
OO-Design in PHENIX PHENIX, a BIG Collaboration A Liberal Data Model
Using an Object Oriented Database to Store BaBar's Terabytes
Presentation transcript:

HENP Computing at BNL Torre Wenaus STAR Software and Computing Leader BNL RHIC & AGS Users Meeting Asilomar, CA October 21, 1999

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Content Bruce’s talk ATLAS Linux Mock Data Challenges D0 focus on areas really changing the scale of HENP comp at BNL Mount’s APOGEE talk Security Software ‘attracting good people’ ROOT; Phenix’s online threaded Objectivity, MySQL RIKEN comp center Esnet Open Science

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Historical Perspective Prior to RHIC, BNL has hosted many small to modest scale AGS experiments With RHIC, BNL moves into realm of large collider detectors  computing task at a scale similar to SLAC, Fermilab, CERN etc. Has required a dramatic change in scale of HENP computing at BNL  RHIC Computing Facility (RCF) established Feb 1997 to supply primary (non-simulation) RHIC computing needs  Successful operations in two ‘Mock Data Challenge’ production stress tests and in summer 1999 engineering run  First physics run in early 2000 Presence of RCF a strong factor in the selection of BNL as the principal US computing site for the CERN LHC ATLAS experiment  Requirements and computing plan similar to RCF  Will operate in close coordination with RCF  LHC and ATLAS operations begin in 2005

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 This Talk Will focus on the major growth of HENP computing as a BNL activity brought by these new programs  RHIC computing at BNL  ATLAS computing at BNL  Brief mention of some other programs  Conclusions Thanks to Bruce Gibbard, RHIC computing facility head, and others (indicated on slides) for materials

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 RHIC Computing at RCF Four experiments: PHENIX, STAR, PHOBOS, BRAHMS  4:4:2:1 relative scales of computing task Aggregate raw data recording rate of ~60 MBytes/sec  Annual raw data volume ~600 TBytes l NB. Size of global WWW content estimated at 7 Tbytes Event reconstruction: 13,000 SPECint95 (450MHz PC = 18 SPECint95) Event filtering (data mining) and physics analysis: 7,000 SPECint95  ‘mining’ interesting data off of tape for physics analyses l aggregate access rates of ~200 MBytes/sec  iterative, interactive analysis of disk-based data by hundreds of users l aggregate access rates of ~1000 MBytes/sec Software development and distribution  100’s of developers; many 100k lines of code per experiment  RCF is primary development and distribution (AFS) site

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Computing Strategies Extensive use of community/commercial/commodity products  hardware and software  increasing use of open software (eg. Linux, MySQL database) Exploit ‘embarrassingly parallel’ nature of HENP computing  farms of loosely coupled processors (Linux PCs on Ethernet)  limited use of Sun machines for I/O intensive analysis Hierarchical storage management (disk + tape robot/shelf) and flexible partitioning of event data based on access characteristics  optimize storage cost and access latencies to interesting data Extensive use of OO software technologies  adopted by all four RHIC experiments, ATLAS, other BNL HENP software efforts (eg. D0), and virtually all other forthcoming expts  primarily C++; some Java  Object I/O: Objectivity commercial OO database and ROOT community (CERN) developed tool

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Event Data Storage and Management Major software challenge: event data storage and management  ROOT: HENP community tool (from CERN) l used by all RHIC experiments for event data storage  Objectivity: Commercial object database l Used by PHENIX for conditions database l RCF did Linux port  Relational databases (MySQL, ORACLE) l Many cataloguing applications in experiments, RCF l MySQL developed by STAR as complement to ROOT for event store, replacing Objectivity  Grand Challenge Architecture l Managed access to HPSS-resident data, particularly for data mining l LBNL-led with ANL, BNL participation; deployment at RCF  Particle Physics Data Grid: transparent wide-area data processing l US HENP ‘Next Generation Internet’ project, primarily LHC directed l RCF/RHIC will act as early testbed

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 ATLAS Computing at BNL A Toroidal LHC ApparatuS  One of 4 experiments at LHC  14 TeV pp collider ATLAS computing at CERN estimated to be >10 times that of RHIC Augmented by regional centers outside CERN  Total scale similar to CERN installation US ATLAS will have one primary ‘Tier 1’ regional center, at BNL  ~20% of CERN facility; ~2x RCF BNL also manages the US ATLAS construction project; ~20% of full ATLAS detector Simulation, data mining, physics analysis, and software development will be primary missions of the BNL Tier 1 center

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 ATLAS: Commonality and Synergy with RHIC Qualitative requirements and Tier 1 quantitative requirements similar to RCF  Exploit economies of scale in hardware and software  Share technical expertise  Learn from and build on RHIC computing as a ‘real world testbed’ Commonality:  Complete coincidence of supported platforms l Intel/Linux processor farms, Sun/Solaris  Objectivity -- and shared concerns over Objectivity!  HPSS -- and shared concerns over HPSS!  Data mining, Grand Challenge  ROOT as an interim analysis tool  Particle Physics Data Grid

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Current Status RHIC RCF  Hardware for first year physics in place, except for some tape store hardware (5 drives; IBM server upgrades)  Extensive testing and tuning to be done l performance, reliability, robustness  All year 1 requirements satisfied except for disk capacity (later augmentation an option; not critically needed now)  In production use by experiments  Positive review by Technical Advisory Committee just concluded US ATLAS Tier 1 center  Initial facility in place, usage by US ATLAS ramping up  Operating out of RCF  ATLAS software installed and operating  More hardware on the way; further increases at proposal stage  Dedicated manpower ramping up

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Conclusions RHIC and RCF have brought BNL to the forefront of HENP computing  Computing scale, imminent operation, mainstream approaches and community involvement make RHIC computing an important testbed for today’s technologies and a stepping stone to the next generation  Performance to date gives confidence for RHIC operations  Strong software efforts at BNL in the experiments BNL as host of US ATLAS Tier 1 center will be a leading HENP computing center in the years to come  Leveraging the facilities, expertise and experience of RCF and the RHIC program  Facility installation to be complemented by a software development effort integrated with the local US ATLAS group Programs well supported by Brookhaven as part of an increased attention to scientific computing at the lab Lots of potential for involvement!

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 RIKEN QCDSP Parallel Computer Special purpose massively parallel machine based on DSPs for quantum field theory calculations  4D mesh with nearest neighbor connections 12,288 node, 600 Gflops Custom designed and built  Collaboration centered at Columbia RIKEN BNL Research Center 192 mother boards, 64 processors each

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 CDIC - Center for Data Intensive Computing Newly established BNL Center developing collaborative projects  Close ties to SUNY at Stony Brook Some of the HENP projects proposed or begun  RHIC Visualization l Newly established collaboration with Stony Brook to develop dynamic 3D visualization tools for RHIC interactions and `beam’s eye’ view  RHIC Computing l Proposed collaboration with IBM to use idle PC cycles for RHIC physics simulation (generator level)  Data Mining l New project studying application of `rough sets’ data mining concepts to RHIC event classification and feature extraction  Accelerator Design l Proposed parallel simulation of beam dynamics for accelerator design and optimization

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Visualization RHIC Au-Au collision animation (Quicktime movie available on web) PHENIX event simulation

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 ESnet Utilization

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Open Software/Open Science Conference BNL Oct 2, 1999  Educate scientists on open source projects  Stimulate open source applications in science  Present science applications to open source developers

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 HENP Computing Challenges Craig Tull, LBNL

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 STAR at RHIC RHIC: Relativistic Heavy Ion Collider at Brookhaven National Laboratory  Colliding Au - Au nuclei at 200GeV/nucleon  Principal objective: Discovery and characterization of the Quark Gluon Plasma  Additional spin physics program in polarized p - p  Engineering run 6-8/99; first year physics run 1/00 STAR experiment  One of two large ‘HEP-scale’ experiments at RHIC, >400 collaborators each (PHENIX is the other)  Heart of experiment is a Time Projection Chamber (TPC) drift chamber (operational) together with Si tracker (year 2) and electromagnetic calorimeter (staged over years 1-3)  Hadrons, jets, electrons and photons over large solid angle

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 The STAR Computing Task Data recording rate of 20MB/sec; ~12MB raw data per event (~1Hz)  ~4000+ tracks/event recorded in tracking detectors (factor of 2 uncertainty in physics generators)  High statistics per event permit event by event measurement and correlation of QGP signals such as strangeness enhancement, J/psi attenuation, high Pt parton energy loss modifications in jets, global thermodynamic variables (eg. Pt slope correlated with temperature)  17M Au-Au events (equivalent) recorded in nominal year Relatively few but highly complex events requiring large processing power  Wide range of physics studies: ~100 concurrent analyses in ~7 physics working groups

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 RHIC/STAR Computing Facilities Dedicated RHIC computing center at BNL, the RHIC Computing Facility  Data archiving and processing for reconstruction and analysis  Three production components: Reconstruction (CRS) and analysis (CAS) services and managed data store (MDS)  10,000 (CRS) + 7,500 (CAS) SpecInt95 CPU  ~50TB disk, 270TB robotic tape, 200MB/s I/O bandwidth, managed by High Performance Storage System (HPSS) developed by DOE/commercial consortium (IBM et al)  Current scale: ~2500 Si95 CPU, 3TB disk for STAR Limited resources require the most cost-effective computing possible  Commodity Intel farms (running Linux) for all but I/O intensive analysis (Sun SMPs) Smaller outside resources:  Simulation, analysis facilities at outside computing centers  Limited physics analysis computing at home institutions

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Implementation of RHIC Computing Model Incorporation of Offsite Facilities T3EHPSS Tape storeSP2 Many universities, etc. Berkeley Japan MIT Doug Olson, LBNL

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 HENP Computing: Today’s Realities Very Large Data Volumes Large, Globally Distributed Collaborations Long Lived Projects (>15 years) Large (1-2M LOC), Complex Analyses Distributed, Heterogeneous Systems Very Limited Computing Manpower Most Computing Manpower are not Professionals  Not necessarily a bad thing! Good understanding and direct interest among developers in the problem Reliance on Open and Commercial Software & Standards Evolving Computer Industry & Technology

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Event Data Storage Management of Petabyte data volumes arguably the most difficult task in HENP computing today  Solutions must map effectively onto OO software technology Intensive community effort in Object Database technology in last 5 years  Focus on Objectivity, the only commercial product that scales to PBytes  Great early promise; strong potential to minimize in-house development and match well the OO architecture of experiments  Reality has been more difficult: development effort much greater than expected, and mixed results on scalability In parallel with Objectivity, community solutions have also been developed  Particularly, ROOT system from CERN supporting I/O of C++ based object models  When complemented by a relational database, provides a robust and scalable solution that integrates well with experiment software The jury is still out  STAR and some other experiments have dropped Objectivity in favor of ROOT+RDBMS  BaBar at SLAC is in production with Objectivity, and is working through the problems

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Data Management Coupled to the event data storage problem, but distinct, is the problem of managing effective archiving and retrieval of the data Hierarchical storage management system required, capable of managing  Terabytes of disk-resident rapid-access data  Petabytes of tape-resident data with medium latency access Industry offers very few solutions today  One (only) has been identified: HPSS  Deployed at RCF (and many other sites), successfully but with caveats l Demands high manpower levels for development and 24x7 support l Still under development, particularly in HENP applications, with stability and robustness issues Community HENP solutions under development in this area as well (Fermilab, DESY)

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Distributed Computing In current generation experiments such as RHIC, and to a much greater degree in the next generation such as LHC, distributed computing is essential  Fully empowering physicists not at the experimental site to participate in development and analysis, with effective access to the data  Distributing the computing and data management task among several large sites l The central site can no longer afford to support computing on its own Near and long term efforts underway to address the need  eg. NOVA project at BNL (Networked Object-based enVironment for Analysis): small project to address immediate and near term needs (STAR/RHIC, ATLAS, possibly others)  Large, LHC directed projects such as the Particle Physics Data Grid project and the MONARC regional center modelling project

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Computing Requirements Nominal year processing and data volume requirements: Raw data volume: 200TB Reconstruction: 2800 Si95 total CPU, 30TB DST data  10x event size reduction from raw to reco  1.5 reconstruction passes/event assumed Analysis: 4000 Si95 total analysis CPU, 15TB micro-DST data  Si95-sec/event per MB of DST depending on analysis l Wide range, from CPU-limited to I/O limited  ~100 active analyses, 5 passes per analysis  micro-DST volumes from.1 to several TB Simulation: 3300 Si95 total including reconstruction, 24TB Total nominal year data volume: 270TB Total nominal year CPU: 10,000 Si95

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 STAR Computing Facilities: RCF  Data archiving and processing for reconstruction and analysis (not simulation; done offsite)  General user services ( , web browsing, etc.)  Three production components: Reconstruction and analysis services (CRS, CAS) and managed data store (MDS)  Nominal year scale: l 10,000 (CRS) + 7,500 (CAS) SpecInt95 CPU §Intel farms running Linux for almost all processing; limited use of Sun SMPs for I/O intensive analysis §Cost-effective, productive, well-aligned with the HENP community l ~50TB disk, 270TB robotic tape, 200MB/s, managed by HPSS  Current scale (when new procurements are in place): l ~2500 Si95 CPU, 3TB disk for STAR l ~8TB of data currently in HPSS

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Computing Facilities Dedicated RHIC computing center at BNL, the RHIC Computing Facility  Data archiving and processing for reconstruction and analysi l Simulation done offsite  10,000 (reco) + 7,500 (analysis) Si95 CPU l Primarily Linux; some Sun for I/O intensive analysis  ~50TB disk, 270TB robotic tape, 200MB/s, managed by HPSS  Current scale (STAR allocation, ~40% of total): l ~2500 Si95 CPU l 3TB disk Support for (a subset of) physics analysis computing at home institutions

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Mock Data Challenges MDC1: Sep/Oct ‘98  >200k (2TB) events simulated offsite; 170k reconstructed at RCF (goal was 100k)  Storage technologies exercised (Objectivity, ROOT)  Data management architecture of Grand Challenge project demonstrated  Concerns identified: HPSS, AFS, farm management software MDC2: Feb/Mar ‘99  New ROOT-based infrastructure in production  AFS improved, HPSS improved but still a concern  Storage technology finalized (ROOT)  New problem area, STAR program size, addressed in new procurements and OS updates (more memory, swap) Both data challenges:  Effective demonstration of productive, cooperative, concurrent (in MDC1) production operations among the four experiments  Bottom line verdict: the facility works, and should perform in physics datataking and analysis

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Offline Software Environment Current software base a mix of Fortran (55%) and C++ (45%)  from ~80%/20% (~95%/5% in non-infrastructure code) in 9/98  New development, and all post-reco analysis, in C++ Framework built over ROOT adopted 11/98  Origins in the ‘Makers’ of ATLFAST  Supports legacy Fortran codes, table (IDL) based data structures developed in previous StAF framework without change  Deployed in offline production and analysis in our ‘Mock Data Challenge 2’, 2-3/99 Post-reconstruction analysis: C++/OO data model ‘StEvent’  StEvent interface is ‘generic C++’; analysis codes are unconstrained by ROOT and need not (but may) use it Next step: migrate the OO data model upstream to reco

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Initial RHIC DB Technology Choices A RHIC-wide Event Store Task Force in Fall ‘97 addressed data management alternatives  Requirements formulated by the four experiments  Objectivity and ROOT were the ‘contenders’ put forward  STAR and PHENIX selected Objectivity as the basis for data management l Concluded that only Objectivity met the requirements of their event stores  ROOT selected by the smaller experiments and seen by all as analysis tool with great potential  Issue for the two larger experiments: l Where to draw a dividing line between Objectivity and ROOT in the data model and data processing

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Event Store Requirements -- And Fall ‘97 View

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Requirements: STAR 8/99 View (My Version)

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 RHIC Data Management: Factors For Evaluation My perception of changes in the STAR view from ‘97 to now are shown Objy Root+MySQLFactor  Cost  Performance and capability as data access solution  Quality of technical support  Ease of use, quality of doc   Ease of integration with analysis   Ease of maintenance, risk   Commonality among experiments   Extent, leverage of outside usage   Affordable/manageable outside RCF   Quality of data distribution mechanisms   Integrity of replica copies   Availability of browser tools   Flexibility in controlling permanent storage location   Level of relevant standards compliance, eg. ODMG   Java access   Partitioning DB and resources among groups

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Object Database: Storage Hierarchy vs User View User deals only with ‘object model’ of his own design; storage details are hidden

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 ATLAS and US ATLAS One of two large HEP experiments at CERN’s Large Hadron Collider (LHC)  Proton-proton collider; 14 TeV in center of mass  1 billion events/year  Principal objective: Discovery and characterization of physics ‘beyond the Standard Model’: Higgs, Supersymmetry, …  Startup Brookhaven hosts the US Project Office for US contributions to ATLAS ~$170M; about 20% of the project Brookhaven recently selected as host lab for US ATLAS Computing and site of US Regional Center  Extension of RHIC Computing Facility  US ATLAS Computing projected to grow to ~$15M/yr

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Conclusions HENP is (unfortunately!) still pushing the envelope in the scale of the data processing and management tasks of present and next generation experiments The HENP community has looked to the commercial and open software worlds for tools and approaches, with strong successes in some areas (OO programming), qualified successes in others (HPSS), and the jury is still out on some (Object Databases)  Moore’s Law and the rise of Linux have made provisioning CPU cycles less of an issue  The community has converged on OO as the principal tool to make software development tractable  But solutions to data storage and management are much less clear  A need on the rise is distributed computing, but internet-driven growth in capacities and technologies will be a strong lever Developments within the HENP community continue to be important, either as fully capable solutions or interim solutions pending further commercial/open software developments

Torre Wenaus, BNL RHIC/AGS Users Meeting 10/99 Conclusions The circumstances of STAR  Startup this year  Slow start in addressing event store implementation, C++ migration  Large base of legacy software  Extremely limited manpower and computing resources drive us to very practical and pragmatic data management choices  Beg, steal and borrow from the community  Deploy community and industry standard technologies  Isolate implementation choices behind standard interfaces, to revisit and re-optimize in the future which leverage existing STAR strengths  Component and standards-based software greatly eases integration of new technologies l preserving compatibility with existing tools for selective and fall-back use l while efficiently migrating legacy software and legacy physicists After some course corrections, we have a capable data management architecture for startup that scales to STAR’s data volumes … but Objectivity is no longer in the picture.