1 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Lawrence.

Slides:



Advertisements
Similar presentations
CHEP 2012 – New York City 1.  LHC Delivers bunch crossing at 40MHz  LHCb reduces the rate with a two level trigger system: ◦ First Level (L0) – Hardware.
Advertisements

Distributed IT Infrastructure for U.S. ATLAS Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Grid Collector: Enabling File-Transparent Object Access For Analysis Wei-Ming Zhang Kent State University John Wu, Alex Sim, Junmin Gu and Arie Shoshani.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
Chapter 9: Moving to Design
Magda – Manager for grid-based data Wensheng Deng Physics Applications Software group Brookhaven National Laboratory.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
Ian Fisk and Maria Girone Improvements in the CMS Computing System from Run2 CHEP 2015 Ian Fisk and Maria Girone For CMS Collaboration.
Grid Data Management A network of computers forming prototype grids currently operate across Britain and the rest of the world, working on the data challenges.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
Master Thesis Defense Jan Fiedler 04/17/98
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
A User’s Introduction to the Grand Challenge Software STAR-GC Workshop Oct 1999 D. Zimmerman.
1 Use of SRMs in Earth System Grid Arie Shoshani Alex Sim Lawrence Berkeley National Laboratory.
Grid Status - PPDG / Magda / pacman Torre Wenaus BNL U.S. ATLAS Physics and Computing Advisory Panel Review Argonne National Laboratory Oct 30, 2001.
PPDG and ATLAS Particle Physics Data Grid Ed May - ANL ATLAS Software Week LBNL May 12, 2000.
May PEM status report. O.Bärring 1 PEM status report Large-Scale Cluster Computing Workshop FNAL, May Olof Bärring, CERN.
1 New developments in the HENP-GC HENP-GC Collaboration New Capabilities in the HENP Grand Challenge Storage Access System and its Application at RHIC.
Grand Challenge MDC1 Plans Doug Olson Nuclear Science Division, Berkeley Lab for the HENP-GC Collaboration RCF Meeting September 24, 1998.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
5 May 98 1 Jürgen Knobloch Computing Planning for ATLAS ATLAS Software Week 5 May 1998 Jürgen Knobloch Slides also on:
NOVA Networked Object-based EnVironment for Analysis P. Nevski, A. Vaniachine, T. Wenaus NOVA is a project to develop distributed object oriented physics.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
Using Bitmap Index to Speed up Analyses of High-Energy Physics Data John Wu, Arie Shoshani, Alex Sim, Junmin Gu, Art Poskanzer Lawrence Berkeley National.
Grand Challenge and PHENIX Report post-MDC2 studies of GC software –feasibility for day-1 expectations of data model –simple robustness tests –Comparisons.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
CLASS Information Management Presented at NOAATECH Conference 2006 Presented by Pat Schafer (CLASS-WV Development Lead)
1 GCA Application in STAR GCA Collaboration Grand Challenge Architecture and its Interface to STAR Sasha Vaniachine presenting for the Grand Challenge.
9 Systems Analysis and Design in a Changing World, Fourth Edition.
PPDG update l We want to join PPDG l They want PHENIX to join NSF also wants this l Issue is to identify our goals/projects Ingredients: What we need/want.
US ATLAS Tier 1 Facility Rich Baker Brookhaven National Laboratory Review of U.S. LHC Software and Computing Projects Fermi National Laboratory November.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
STAR C OMPUTING STAR Analysis Operations and Issues Torre Wenaus BNL STAR PWG Videoconference BNL August 13, 1999.
PPDGLHC Computing ReviewNovember 15, 2000 PPDG The Particle Physics Data Grid Making today’s Grid software work for HENP experiments, Driving GRID science.
STAR Collaboration, July 2004 Grid Collector Wei-Ming Zhang Kent State University John Wu, Alex Sim, Junmin Gu and Arie Shoshani Lawrence Berkeley National.
January 26, 2003Eric Hjort HRMs in STAR Eric Hjort, LBNL (STAR/PPDG Collaborations)
PHENIX and the data grid >400 collaborators 3 continents + Israel +Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
NOVA A Networked Object-Based EnVironment for Analysis “Framework Components for Distributed Computing” Pavel Nevski, Sasha Vanyashin, Torre Wenaus US.
Grand Challenge in MDC2 D. Olson, LBNL 31 Jan 1999 STAR Collaboration Meeting
STAR C OMPUTING Plans for Production Use of Grand Challenge Software in STAR Torre Wenaus BNL Grand Challenge Meeting LBNL 10/23/98.
1 fileCatalog, tagDB and GCA A. Vaniachine Grand Challenge STAR fileCatalog, tagDB and Grand Challenge Architecture A. Vaniachine presenting for the Grand.
1 Grid in STAR: Needs and Plans A. Vaniachine STAR Grid Components in STAR: Experience, Needs and Plans A. Vaniachine Lawrence Berkeley National Laboratory.
Securing the Grid & other Middleware Challenges Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
Tier 1 at Brookhaven (US / ATLAS) Bruce G. Gibbard LCG Workshop CERN March 2004.
Magda Distributed Data Manager Prototype Torre Wenaus BNL September 2001.
Participation of JINR in CERN- INTAS project ( ) Korenkov V., Mitcin V., Nikonov E., Oleynik D., Pose V., Tikhonenko E. 19 march 2004.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Grid Status - PPDG / Magda / pacman Torre Wenaus BNL DOE/NSF Review of US LHC Software and Computing Fermilab Nov 29, 2001.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Grid Activities in CMS Asad Samar (Caltech) PPDG meeting, Argonne July 13-14, 2000.
DOE/NSF Quarterly review January 1999 Particle Physics Data Grid Applications David Malon Argonne National Laboratory
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
1 Particle Physics Data Grid (PPDG) project Les Cottrell – SLAC Presented at the NGI workshop, Berkeley, 7/21/99.
G. Russo, D. Del Prete, S. Pardi Kick Off Meeting - Isola d'Elba, 2011 May 29th–June 01th A proposal for distributed computing monitoring for SuperB G.
The HENP Grand Challenge Project and initial use in the RHIC Mock Data Challenge 1 D. Olson DM Workshop SLAC, Oct 1998.
Magda Distributed Data Manager Torre Wenaus BNL October 2001.
Hall D Computing Facilities Ian Bird 16 March 2001.
Business System Development
Nuclear Physics Data Management Needs Bruce G. Gibbard
Presentation transcript:

1 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Lawrence Berkeley National Laboratory June 2, 2000 CDIC Seminar

2 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Outline The Grid Grand Challenge NOVA PPDG Summary

3 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Vision Today: –you don’t care which power plant actually produced the electric power –you just plug your microwave and use it Tomorrow: –you don’t care which actual computer did your computations –you just plug your PDA and use it

4 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Virtual Data Grid OO Computing –encapsulation of data and algorithms –large data transfers Selection Parameters Federation DB1 DB3 DB4 DB5 DB6 CPU Local Remote OODB DB2 Drawing by Federico Carminati

5 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Algorithmic Virtual Data Agent Computing –separation of data and algorithms –small data transfers Algorithmic Virtual Data Selection Parameters DB1 DB4 DB5 DB6 CPU Local Remote Procedure Proc.C PROOF CPU TagDB RDB DB3 DB2 Drawing by Federico Carminati

6 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Data Access –GCA: Grand Challenge Architecture optimized storage access for physics analysis integrating with the STAR experiment at RHIC –NOVA: Networked Object-based enVironment for Analysis prototyped components for physics analysis in distributed environment –PPDG: Particle Physics Data Grid addresses long-term data-management needs of high-energy and nuclear physics rapidly deploys advanced services

7 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine GCA Participants NERSC/Berkeley Lab –L. Bernardo, A. Mueller, H. Nordberg, A.Shoshani, A. Sim, J. Wu Argonne –D. Malon, E. May, G. Pandola Brookhaven Lab –B. Gibbard, S. Johnson, J. Porter, T.Wenaus Nuclear Science/Berkeley Lab –D. Olson, A. Vaniachine, J. Yang, D.Zimmerman

8 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Grand Challenge Architecture An order-optimized prefetch architecture for data retrieval from multilevel storage in a distributed multiuser environment Queries select events and specific event components based upon tag attribute ranges –query estimates are provided prior to execution –support for collections as queries –Because event components are distributed over several files, processing an event requires delivery of a “bundle” of files Events are delivered in an order that takes advantage of what is already on disk, and multi-user policy-based prefetching of further data from tertiary storage GCA inter-component communication is though a CORBA layer (shielded from physicists)

9 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Challenge There are several –Not all data fits on disk ($$) Part of 1 year’s DST’s fit on disk –What about last year, 2 year’s ago? –What about hits, raw? –Available disk bandwidth means data read into memory must be efficiently used ($$) don’t read unused portions of the event Don’t read events you don’t need –Available tape bandwidth means files read from tape must be shared by many users, files should not contain unused bytes ($$$$) –Facility resources are sufficient only if used efficiently Should operate steady-state (nearly) fully loaded

10 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Bottleneks Keep recently accessed data on disk, but manage it so unused data does not waste space. Try to arrange that 90% of file access is to disk and only 10% are retrieved from tape.

11 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Efficient Access Split event into components across different files so that most bytes read are used –Raw, tracks, hits, tags, summary, trigger, … Optimize file size so tape bandwidth is not wasted –1GB files,  means different # of events in each file Coordinate file usage so tape access is shared –Users select all files at once –System optimizes retrieval and order of processing Use disk space & bandwidth efficiently –Operate disk as cache in front of tape

12 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Bundles Event Identifiers (Run#, Event#) Event components Files File bundle 1File bundle 2File bundle 3

13 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Event Components T. Ullrich, Jan. 2000

14 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine STACS: STorage Access Coordination System Bit-Sliced Index File Catalog Policy Module Query Status, CacheMap Query Monitor List of file bundles and events Cache Manager Requests for file caching and purging Query Estimator Estimate pftp and file purge commands File Bundles, Event lists Query

15 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine database Data Flow GC System StIOMaker fileCatalog tagDB Query Monitor Cache Manager Query Estimator STAR Software Index Builder gcaClient FileCatalog IndexFeeder GCA Interface

16 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Eliminating Dependencies StIOMaker ROOT + STAR Software StChallenger ::Challenge() libGCAClient.so libChallenger.so (implementation) CORBA + GCA software libOB.so TNamed > StFileI

17 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine GCA-STAR Integration Workshop March 2000

18 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine GCA Services Status Running on rsun00 since end of March 3.6 TB of simulated data indexed 179 physics tags: –StrangeTag –FlowTag –ScaTag 240 user queries served Next week: index of 1M cosmic events

19 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine NOVA Networked Object-based EnVironment for Analysis Modular components for distributed computing; application-neutral interfaces –Can be used as a coherent framework or in isolation to extend existing analysis systems Focused on support for C++ based analysis –Used for all RHIC, LHC, other large experiments Emphasis on user participation in iterative development; real-world prototyping and testing Use of existing tools and technologies

20 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Architecture

21 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Domains –Regional Center Central management and execution of analysis –Remote Client Mobile Analysis –Middleware Components Data exchange and navigation tools Client/Server object request brokerage –Data Management Data repository, catalogue, and interface Data model for simple objects (C-structs)

22 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Deliverables Configuration manager for analysis jobs Analysis workflow catalog Database for versioned dataObjects Web-based database navigation tool Distributed job submission and monitoring system

23 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Job monitoring system Workflow Cataloguing fileCatalog Job configuration manager

24 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine PPDG Participants

25 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Data Replication LBNL - ESNet: 612Mb/s (OC12) BNL - ESNet: 45Mb/s (T3, does 1 MB/s LBNL/NERSC RCF) BNL - ESNet: 145Mb/s after upgrade to OC3 (local: 100 Mb/s, firewall: 70Mb/s) STAR: BNL to LBNL peak: 3 MB/s sustained: 0.1 TB/day Planned for July 2000 PPDG: ANL to LBNL OC3: 3 MB/s OC12: 8 MB/s

26 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Summary Grand Challenge –optimized access to multi-component event data files stored in HPSS –tested up to 10M events, 250 concurrent queries –has been deployed in the STAR experiment –solutions for data management in PPDG project NOVA –developed tools for distributed physics analysis –technologies for “Algrorithmic Virtual Data” LDRD proposal