Jerome Lauret RCF Advisory Committee Meeting The Data Carousel what problem it’s trying to solve the data carousel and the grand challenge the bits and.

Slides:



Advertisements
Similar presentations
4.4 Page replacement algorithms
Advertisements

16/9/2004Features of the new CASTOR1 Alice offline week, 16/9/2004 Olof Bärring, CERN.
Grid Collector: Enabling File-Transparent Object Access For Analysis Wei-Ming Zhang Kent State University John Wu, Alex Sim, Junmin Gu and Arie Shoshani.
The Google File System. Why? Google has lots of data –Cannot fit in traditional file system –Spans hundreds (thousands) of servers connected to (tens.
Informationsteknologi Friday, November 16, 2007Computer Architecture I - Class 121 Today’s class Operating System Machine Level.
What is it? Hierarchical storage software developed in collaboration with five US department of Energy Labs since 1992 Allows storage management of 100s.
Magda – Manager for grid-based data Wensheng Deng Physics Applications Software group Brookhaven National Laboratory.
CT NIKHEF June File server CT system support.
Web Application Architecture: multi-tier (2-tier, 3-tier) & mvc
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
STAR Software Basics Introduction to the working environment Lee Barnby - Kent State University.
DEMIGUISE STORAGE An Anonymous File Storage System VIJAY KUMAR RAVI PRAGATHI SEGIREDDY COMP 512.
Distributed File Systems
Central Reconstruction System on the RHIC Linux Farm in Brookhaven Laboratory HEPIX - BNL October 19, 2004 Tomasz Wlodek - BNL.
3rd Nov 2000HEPiX/HEPNT CDF-UK MINI-GRID Ian McArthur Oxford University, Physics Department
Ideas to Improve SharePoint Usage 4. What are these 4 Ideas? 1. 7 Steps to check SharePoint Health 2. Avoid common Deployment Mistakes 3. Analyze SharePoint.
Caravan Business Server a viable alternative development platform niti telecom consultancy april 2002.
A User’s Introduction to the Grand Challenge Software STAR-GC Workshop Oct 1999 D. Zimmerman.
An Overview of PHENIX Computing Ju Hwan Kang (Yonsei Univ.) and Jysoo Lee (KISTI) International HEP DataGrid Workshop November 8 ~ 9, 2002 Kyungpook National.
Paul Scherrer Institut 5232 Villigen PSI HEPIX_AMST / / BJ95 PAUL SCHERRER INSTITUT THE PAUL SCHERRER INSTITUTE Swiss Light Source (SLS) Particle accelerator.
Nov 1, 2000Site report DESY1 DESY Site Report Wolfgang Friebel DESY Nov 1, 2000 HEPiX Fall
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
Introduction to U.S. ATLAS Facilities Rich Baker Brookhaven National Lab.
Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
NOVA Networked Object-based EnVironment for Analysis P. Nevski, A. Vaniachine, T. Wenaus NOVA is a project to develop distributed object oriented physics.
Magda Distributed Data Manager Status Torre Wenaus BNL ATLAS Data Challenge Workshop Feb 1, 2002 CERN.
Kurt Mueller San Diego Supercomputer Center NPACI HotPage Updates.
Using Bitmap Index to Speed up Analyses of High-Energy Physics Data John Wu, Arie Shoshani, Alex Sim, Junmin Gu, Art Poskanzer Lawrence Berkeley National.
LHCb-Italy Farm Monitor Domenico Galli Bologna, June 13, 2001.
Grand Challenge and PHENIX Report post-MDC2 studies of GC software –feasibility for day-1 expectations of data model –simple robustness tests –Comparisons.
BNL DDM Status Report Hironori Ito Brookhaven National Laboratory.
LCG Phase 2 Planning Meeting - Friday July 30th, 2004 Jean-Yves Nief CC-IN2P3, Lyon An example of a data access model in a Tier 1.
HPSS for Archival Storage Tom Sherwin Storage Group Leader, SDSC
240-Current Research Easily Extensible Systems, Octave, Input Formats, SOA.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Facilities and How They Are Used ORNL/Probe Randy Burris Dan Million – facility administrator.
1 GCA Application in STAR GCA Collaboration Grand Challenge Architecture and its Interface to STAR Sasha Vaniachine presenting for the Grand Challenge.
GLAST Science Support CenterJuly, 2003 LAT Ground Software Workshop Status of the D1 (Event) and D2 (Spacecraft Data) Database Prototypes for DC1 Robert.
CC-J Monthly Report Shin’ya Sawada (KEK) for CC-J Working Group
Disk Farms at Jefferson Lab Bryan Hess
STAR C OMPUTING STAR Analysis Operations and Issues Torre Wenaus BNL STAR PWG Videoconference BNL August 13, 1999.
STAR Collaboration, July 2004 Grid Collector Wei-Ming Zhang Kent State University John Wu, Alex Sim, Junmin Gu and Arie Shoshani Lawrence Berkeley National.
January 26, 2003Eric Hjort HRMs in STAR Eric Hjort, LBNL (STAR/PPDG Collaborations)
PHENIX and the data grid >400 collaborators 3 continents + Israel +Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
Grand Challenge in MDC2 D. Olson, LBNL 31 Jan 1999 STAR Collaboration Meeting
STAR C OMPUTING Plans for Production Use of Grand Challenge Software in STAR Torre Wenaus BNL Grand Challenge Meeting LBNL 10/23/98.
1 fileCatalog, tagDB and GCA A. Vaniachine Grand Challenge STAR fileCatalog, tagDB and Grand Challenge Architecture A. Vaniachine presenting for the Grand.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
Liverpool Experience of MDC 1 MAP (and in our belief any system which attempts to be scaleable to 1000s of nodes) broadcasts the code to all the nodes.
November 10, 1999PHENIX CC-J Updates in Nov.991 PHENIX CC-J Updates in Nov New Hardware - N.Hayashi / RIKEN November 10, 1999 PHENIX Computing Meeting.
Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.
SPI NIGHTLIES Alex Hodgkins. SPI nightlies  Build and test various software projects each night  Provide a nightlies summary page that displays all.
RCF Status Extended outage of the Mass Storage System (HPSS) last Wednesday –Latest transaction logs of namespace DB were erroneously deleted in the production.
(ITI310) By Eng. BASSEM ALSAID SESSIONS 10: Internet Information Services (IIS)
November 1, 2004 ElizabethGallas -- D0 Luminosity Db 1 D0 Luminosity Database: Checklist for Production Elizabeth Gallas Fermilab Computing Division /
10 March Andrey Grid Tools Working Prototype of Distributed Computing Infrastructure for Physics Analysis SUNY.
Oct. 6, 1999PHENIX Comp. Mtg.1 CC-J: Progress, Prospects and PBS Shin’ya Sawada (KEK) For CCJ-WG.
/16 Final Project Report By Facializer Team Final Project Report Eagle, Leo, Bessie, Five, Evan Dan, Kyle, Ben, Caleb.
CASTOR new stager proposal CASTOR users’ meeting 24/06/2003 The CASTOR team.
A Web Based Job Submission System for a Physics Computing Cluster David Jones IOP Particle Physics 2004 Birmingham 1.
1 Policy Based Systems Management with Puppet Sean Dague
File-System Management
CASTOR: possible evolution into the LHC era
Sanjay Ghemawat, Howard Gobioff, Shun-Tak Leung
Service Challenge 3 CERN
Interoperability of Digital Repositories
A Web-Based Data Grid Chip Watson, Ian Bird, Jie Chen,
Distributed P2P File System
Overview Assignment 12: solution Distributed file systems 1/13/2019.
Presentation transcript:

Jerome Lauret RCF Advisory Committee Meeting The Data Carousel what problem it’s trying to solve the data carousel and the grand challenge the bits and pieces: how it all works what’s ready now; what’s left to do

Jerome Lauret RCF Advisory Committee Meeting the problem we’re facing PHENIX program heavy in “ensemble” physics –typical day (or week) at the office: get lots of events, make foreground and background distributions, compare, improve code, repeat until published needs to move lots of data very efficiently needs to be comprehensible to PHENIX physicists –people are accustomed to “staging files” needs to work with the CAS analysis architecture –lots of linux boxes with 30 GB disk on each –main NFS server with 3 TB disk solution: optimized batch file mover –similar to Fermilab data “freight train” –works with existing tools: HPSS, ssh, pftp, perl

Jerome Lauret RCF Advisory Committee Meeting the carousel and the grand challenge complementary tools for accessing event data works at lower level of abstraction than GC –files, not objects can work with non-event data files –important since it doesn’t take much to clog access to tapes –11 MB/sec/drive in principle; 6 MB/sec/drive in practice –best case: Eagles take ~20 seconds to load, seek: read ~100 MB files at random and you’ll see no better than 50% bandwidth –MDC1,2 naive ftp only saw ~1 MB/sec effective bandwidth for reads already works with disjoint staging areas can, in principle, work over the WAN doesn’t reorganize data, doesn’t provide event iterator, isn’t coupled to analysis code –good or bad, depends on what you’re expecting

Jerome Lauret RCF Advisory Committee Meeting the bits and pieces split-brain server –part which knows HPSS, part which knows PHENIX HPSS batch queue (Jae Kerr, IBM) –optimizes tape mounts for a given set of file requests –once file is staged to cache, used NFS write to non-cache disk –modified to use pftp call-back (Tom Throwe, J.K.) carousel server (J. Lauret, SUNYSB) –feeds sets of files to batch queue at measured pace –knows about groups, does group-level accounting –implements file retrieval policy –maintains all state info in external database client side scripts –implements file deletion policy (defaults to LRU cache) client side requirements are kept ALARA –ssh +.shosts, perl + few modules, pftp

Jerome Lauret RCF Advisory Committee Meeting carousel architecture “ORNL” software carousel server mySQL database filelist HPSS tape HPSS cache pftp rmine0x client pftp CAS data mover CAS local disk NFS disk

Jerome Lauret RCF Advisory Committee Meeting showing carousel info via the web

Jerome Lauret RCF Advisory Committee Meeting accounting tables group-level accounting information provides possibility of tailoring access to HPSS resources

Jerome Lauret RCF Advisory Committee Meeting current state and future directions works (has basically worked since MDC2) two main sources for code – –PHENIX CVS repository there remains one PHENIX-ism to be exorcised –HPSS batch queue is currently hardwired to suid to “phnxreco” –instead could select uid, gid based on COS lots of future improvements are possible –have worked to make system “good enough” to use –could use more sophisticated server/client communication –check for available space before staging file to HPSS cache