February 28, 2003Eric Hjort PDSF Status and Overview Eric Hjort, LBNL STAR Collaboration Meeting February 28, 2003.

Slides:



Advertisements
Similar presentations
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Advertisements

23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
STAR Software Walk-Through. Doing analysis in a large collaboration: Overview The experiment: – Collider runs for many weeks every year. – A lot of data.
Grid Collector: Enabling File-Transparent Object Access For Analysis Wei-Ming Zhang Kent State University John Wu, Alex Sim, Junmin Gu and Arie Shoshani.
Slides for Grid Computing: Techniques and Applications by Barry Wilkinson, Chapman & Hall/CRC press, © Chapter 1, pp For educational use only.
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
Data Management for Physics Analysis in PHENIX (BNL, RHIC) Evaluation of Grid architecture components in PHENIX context Barbara Jacak, Roy Lacey, Saskia.
1-2.1 Grid computing infrastructure software Brief introduction to Globus © 2010 B. Wilkinson/Clayton Ferner. Spring 2010 Grid computing course. Modification.
DataGrid Kimmo Soikkeli Ilkka Sormunen. What is DataGrid? DataGrid is a project that aims to enable access to geographically distributed computing power.
The B A B AR G RID demonstrator Tim Adye, Roger Barlow, Alessandra Forti, Andrew McNab, David Smith What is BaBar? The BaBar detector is a High Energy.
DIRAC API DIRAC Project. Overview  DIRAC API  Why APIs are important?  Why advanced users prefer APIs?  How it is done?  What is local mode what.
Enabling Grids for E-sciencE Medical image processing web portal : Requirements analysis. An almost end user point of view … H. Benoit-Cattin,
Workload Management WP Status and next steps Massimo Sgaravatto INFN Padova.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
Grid Computing - AAU 14/ Grid Computing Josva Kleist Danish Center for Grid Computing
Computing and LHCb Raja Nandakumar. The LHCb experiment  Universe is made of matter  Still not clear why  Andrei Sakharov’s theory of cp-violation.
SUMS ( STAR Unified Meta Scheduler ) SUMS is a highly modular meta-scheduler currently in use by STAR at there large data processing sites (ex. RCF /
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
Central Reconstruction System on the RHIC Linux Farm in Brookhaven Laboratory HEPIX - BNL October 19, 2004 Tomasz Wlodek - BNL.
ESP workshop, Sept 2003 the Earth System Grid data portal presented by Luca Cinquini (NCAR/SCD/VETS) Acknowledgments: ESG.
MySQL and GRID Gabriele Carcassi STAR Collaboration 6 May Proposal.
100 Million events, what does this mean ?? STAR Grid Program overview.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
Overview Why are STAR members encouraged to use SUMS ? Improvements and additions to SUMS Research –Job scheduling with load monitoring tools –Request.
1 Use of SRMs in Earth System Grid Arie Shoshani Alex Sim Lawrence Berkeley National Laboratory.
STAR Analysis Meeting, BNL, Dec 2004 Alexandre A. P. Suaide University of Sao Paulo Slide 1 BEMC software and calibration L3 display 200 GeV February.
Grid Workload Management Massimo Sgaravatto INFN Padova.
Costin Grigoras ALICE Offline. In the period of steady LHC operation, The Grid usage is constant and high and, as foreseen, is used for massive RAW and.
PHENIX Simulation System 1 December 7, 1999 Simulation: Status and Milestones Tarun Ghosh, Indrani Ojha, Charles Vanderbilt University.
Offline shifter training tutorial L. Betev February 19, 2009.
MAGDA Roger Jones UCL 16 th December RWL Jones, Lancaster University MAGDA  Main authors: Wensheng Deng, Torre Wenaus Wensheng DengTorre WenausWensheng.
August 13, 2003Eric Hjort Getting Started with Grid Computing in STAR Eric Hjort, LBNL STAR Collaboration Meeting August 13, 2003.
Using Bitmap Index to Speed up Analyses of High-Energy Physics Data John Wu, Arie Shoshani, Alex Sim, Junmin Gu, Art Poskanzer Lawrence Berkeley National.
PHENIX and the data grid >400 collaborators Active on 3 continents + Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
STAR Software Walk-Through. Doing analysis in a large collaboration: Overview The experiment: – Collider runs for many weeks every year. – A lot of data.
STAR Collaboration Meeting, BNL, Feb 2005 Alexandre A. P. Suaide University of Sao Paulo Slide 1 BEMC software update L3 display 200 GeV February.
PPDG update l We want to join PPDG l They want PHENIX to join NSF also wants this l Issue is to identify our goals/projects Ingredients: What we need/want.
CC-J Monthly Report Shin’ya Sawada (KEK) for CC-J Working Group
5/2/  Online  Offline 5/2/20072  Online  Raw data : within the DAQ monitoring framework  Reconstructed data : with the HLT monitoring framework.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
STAR Collaboration, July 2004 Grid Collector Wei-Ming Zhang Kent State University John Wu, Alex Sim, Junmin Gu and Arie Shoshani Lawrence Berkeley National.
January 26, 2003Eric Hjort HRMs in STAR Eric Hjort, LBNL (STAR/PPDG Collaborations)
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
PHENIX and the data grid >400 collaborators 3 continents + Israel +Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
AliEn AliEn at OSC The ALICE distributed computing environment by Bjørn S. Nilsen The Ohio State University.
FTPC status and results Summary of last data taken AuAu and dAu calibration : Data Quality Physic results with AuAu data –Spectra –Flow Physic results.
Jean-Roch Vlimant, CERN Physics Performance and Dataset Project Physics Data & MC Validation Group McM : The Evolution of PREP. The CMS tool for Monte-Carlo.
RCF Status - Introduction PHENIX and STAR Counting Houses are connected to RCF at a Network Bandwidth of 20 Gbits/sec each –Redundant (Bandwidth-wise and.
November 10, 1999PHENIX CC-J Updates in Nov.991 PHENIX CC-J Updates in Nov New Hardware - N.Hayashi / RIKEN November 10, 1999 PHENIX Computing Meeting.
STAR J/  Trigger in dA Manuel Calderon for the Heavy-Flavor Group Trigger Workshop at BNL October 21, 2002.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
STAR Collaboration Meeting, BNL – march 2003 Alexandre A. P. Suaide Wayne State University Slide 1 EMC Update Update on EMC –Hardware installed and current.
M. Oldenburg GridPP Metadata Workshop — July 4–7 2006, Oxford University 1 Markus Oldenburg GridPP Metadata Workshop July 4–7 2006, Oxford University ALICE.
AliRoot survey: Analysis P.Hristov 11/06/2013. Are you involved in analysis activities?(85.1% Yes, 14.9% No) 2 Involved since 4.5±2.4 years Dedicated.
1 NCC Task Force Richard Seto NCC Task Force Meeting Jan 8, 2007 BNL.
Predrag Buncic CERN Future of the Offline. Data Preparation Group.
Production Mode Data-Replication Framework in STAR using the HRM Grid CHEP ’04 Congress Centre Interlaken, Switzerland 27 th September – 1 st October Eric.
10 March Andrey Grid Tools Working Prototype of Distributed Computing Infrastructure for Physics Analysis SUNY.
Joe Foster 1 Two questions about datasets: –How do you find datasets with the processes, cuts, conditions you need for your analysis? –How do.
BaBar & Grid Eleonora Luppi for the BaBarGrid Group TB GRID Bologna 15 febbraio 2005.
Year 3 Reconstruction and Calibration Status David Hardtke LBNL 1.What’s new? 2.Status of Calibrations 3.Why we need ITTF.
U.S. ATLAS Grid Production Experience
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Simulation use cases for T2 in ALICE
Leigh Grundhoefer Indiana University
Near Real Time Reconstruction of PHENIX Run7 Minimum Bias Data From RHIC Project Goals Reconstruct 10% of PHENIX min bias data from the RHIC Run7 (Spring.
Preparations for Reconstruction of Run7 Min Bias PRDFs at Vanderbilt’s ACCRE Farm (more substantial update set for next week) Charles Maguire et al. March.
Grid Computing Software Interface
Presentation transcript:

February 28, 2003Eric Hjort PDSF Status and Overview Eric Hjort, LBNL STAR Collaboration Meeting February 28, 2003

Eric Hjort PDSF Overview STAR has production facilities at RCF and PDSF –PDSF needs to transfer data: Replicate all DSTs and some raw data from RCF to PDSF –PDSF computing: Embedding production, data analysis, simulations PDSF infrastructure people: –Doug Olson (STAR computing at PDSF coordinator) –Iwona Sakrejda (PDSF user support, STAR libraries, accounts, etc.) Contact by filing a PDSF support request –Eric Hjort (Embedding, data management) Contact by or hypernews: etc. –PDSF oversight committee: P. Jacobs (chair), D. Olson, K. Schweda, D. Hardtke, I. Sakrejda, S. Canon (PDSF project leader), J. Lauret, E. Hjort

February 28, 2003Eric Hjort File Replication: RCF to PDSF (A simple example of Grid Tools) HPSS Disk Cache HPSS Disk Cache Request_to_GET Replica Coordinator HRM LBNL BNL GridFTP CORBA Interface Request_to_PUT Status, errors, etc. PFTP File in cache HRM = Hierarchical Resource Manager

February 28, 2003Eric Hjort Grid computing For STAR data transfer: –Authenticate with grid certificate Convenient: not necessary to log in at RCF Use grid-proxy-init (requires password) –Automatic caching of data by HRMs large cache disk not necessary enables continous automatic transfers –Result: Grid tools improve net transfer rate, reduce effort What does it take to do grid computing? –Get a DOE Science Grid certificate –Get it installed on STAR grid nodes at RCF and PDSF –Easy to do some simple, convenient things with Globus –Alpha testers wanted – Iwona has prepared instructions

February 28, 2003Eric Hjort Disk Resources at PDSF High performance disk (1 TB on /aztera) –Heavily used datasets; embedding input files Distributed disks (14 TB on 70 nodes) –For large MuDst productions. –Data access by the STAR job scheduler. Production data vaults (11 TB) –Embedding output; simulations data; selected MuDst’s; etc. –Managed by production people PWG data vaults (5.3 TB) –Managed by PWG’s. –More space available on a by-request basis. Scratch space (1 TB on pdsfdv15)

February 28, 2003Eric Hjort How to find STAR Data at PDSF Summary of STAR data at PDSF: Embedding data on disk MuDst's and simulation data on disk Job scheduler instructions and data on distributed disks.Job scheduler These pages update automatically every 24 hours. If you can’t find it on these pages… New links on the STAR PDSF help page:

February 28, 2003Eric Hjort Job submission at PDSF Job scheduler in use for data on distributed disks –No special queue for scheduler –Not fully function without a PDSF file catalog –Uses pre-made filelists Job scheduler not used for data on NFS disks –Data is filtered for sanity=1 Queues/priorities: –Short (1hr), medium (24hr), long (5 days) –Production account (user starofl) has a higher priority –At present production is not run on distributed disk nodes –Important to balance production vs. users’ resources

February 28, 2003Eric Hjort Embedding Overview Embedding production done at PDSF –“embed” simulated particles into real data at the raw data level –Reconstruction yields efficiencies –Important test of STAR software: simulations meet real data 20 TB, 10 M events in 2002 People involved: –Eric Hjort (infrastructure and development; spectra production) –Matt Lamont (strangeness-specific infrastructure, strangeness production) –Patricia Fachini (development; miscellaneous production) –Christina Markert (development; miscellaneous production) –Olga Barranikova (QA) –STAR collaborators (SOFI, calibrations, simulations, etc.)

February 28, 2003Eric Hjort Embedding Methods Year 2 AuAu TPC embedding –20 GeV –200 GeV (P02gd) running without problems for about 1 year At least 29 different embedded particles Central, minbias, both fields, various pt, y ranges, etc. pp embedding –Hijing pp -> zerobias (vertex reconstruction studies) –Hijing pp + embedded particle -> zerobias (Jon G.) –Embedded particle -> pp data (Matt/strangeness) RICH embedding (Boris H.): Tested and in production dAu embedding status –Working in a testing mode –Need to test/understand dE/dx shift –Many dAu and zerobias daq files are at PDSF –Initial production setup ready next week dAu FTPC (Frank S.): Working; needs details + testing

February 28, 2003Eric Hjort Embedding Requests Ask your PWG convenor to submit request to Simulations Production Request Page: –Organizes and documents the work –Specifies job parameters for reference –Allows for prioritization of jobs by Jamie –Protects against resource misuse If not prioritized, jobs order = submissions order, but… –Multiple operators mix order –Technical reasons mix order –Some requests take much longer than others

February 28, 2003Eric Hjort Summary/Future plans Data transfer –Grid tools serve us well –Data transfer needs are met in general –Goal for this year’s run: reduce latency to 1 week or less? Data management –PDSF data discovery webpages overhauled –Job scheduler in use –Next big step: file catalog at PDSF Embedding and Simulations –AuAu, pp, RICH embedding all working –dAu TPC and FTPC embedding almost in production –Future: new detectors, understand triggers, etc. Bigger picture –Seamless, more automated data transfer RCF PDSF –Distributed grid computing with the Job Scheduler.