Jan. 17, 2002DØRAM Proposal DØRACE Meeting, Jae Yu 1 Proposal for a DØ Remote Analysis Model (DØRAM) IntroductionIntroduction Remote Analysis Station ArchitectureRemote.

Slides:



Advertisements
Similar presentations
4/2/2002HEP Globus Testing Request - Jae Yu x Participating in Globus Test-bed Activity for DØGrid UTA HEP group is playing a leading role in establishing.
Advertisements

Amber Boehnlein, FNAL D0 Computing Model and Plans Amber Boehnlein D0 Financial Committee November 18, 2002.
Reliability Week 11 - Lecture 2. What do we mean by reliability? Correctness – system/application does what it has to do correctly. Availability – Be.
The Google File System. Why? Google has lots of data –Cannot fit in traditional file system –Spans hundreds (thousands) of servers connected to (tens.
Client-Server Computing in Mobile Environments
Microsoft ® Application Virtualization 4.5 Infrastructure Planning and Design Series.
Microsoft ® Application Virtualization 4.6 Infrastructure Planning and Design Published: September 2008 Updated: February 2010.
Shilpa Seth.  Centralized System Centralized System  Client Server System Client Server System  Parallel System Parallel System.
Data Center Infrastructure
The D0 Monte Carlo Challenge Gregory E. Graham University of Maryland (for the D0 Collaboration) February 8, 2000 CHEP 2000.
Microsoft ® Application Virtualization 4.6 Infrastructure Planning and Design Published: September 2008 Updated: November 2011.
DØ IB Meeting, Nov. 8, 2001 J. Yu, UTA, Remote Analysis Status Remote Analysis Coordination Computing hardware is rather inexpensive –CPU and storage media.
High Energy Physics At OSCER A User Perspective OU Supercomputing Symposium 2003 Joel Snow, Langston U.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
UTA Site Report Jae Yu UTA Site Report 4 th DOSAR Workshop Iowa State University Apr. 5 – 6, 2007 Jae Yu Univ. of Texas, Arlington.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
Hadoop Hardware Infrastructure considerations ©2013 OpalSoft Big Data.
3rd Nov 2000HEPiX/HEPNT CDF-UK MINI-GRID Ian McArthur Oxford University, Physics Department
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Snapshot of the D0 Computing and Operations Planning Process Amber Boehnlein For the D0 Computing Planning Board.
Status of DØ Computing at UTA Introduction The UTA – DØ Grid team DØ Monte Carlo Production The DØ Grid Computing –DØRAC –DØSAR –DØGrid Software Development.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
A Design for KCAF for CDF Experiment Kihyeon Cho (CHEP, Kyungpook National University) and Jysoo Lee (KISTI, Supercomputing Center) The International Workshop.
DØ RAC Working Group Report Progress Definition of an RAC Services provided by an RAC Requirements of RAC Pilot RAC program Open Issues DØRACE Meeting.
DØ RACE Introduction Current Status DØRAM Architecture Regional Analysis Centers Conclusions DØ Internal Computing Review May 9 – 10, 2002 Jae Yu.
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
7April 2000F Harris LHCb Software Workshop 1 LHCb planning on EU GRID activities (for discussion) F Harris.
Status of UTA IAC + RAC Jae Yu 3 rd DØSAR Workshop Apr. 7 – 9, 2004 Louisiana Tech. University.
Spending Plans and Schedule Jae Yu July 26, 2002.
D0RACE: Testbed Session Lee Lueking D0 Remote Analysis Workshop February 12, 2002.
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
DØSAR a Regional Grid within DØ Jae Yu Univ. of Texas, Arlington THEGrid Workshop July 8 – 9, 2004 Univ. of Texas at Arlington.
From DØ To ATLAS Jae Yu ATLAS Grid Test-Bed Workshop Apr. 4-6, 2002, UTA Introduction DØ-Grid & DØRACE DØ Progress UTA DØGrid Activities Conclusions.
GridPP11 Liverpool Sept04 SAMGrid GridPP11 Liverpool Sept 2004 Gavin Davies Imperial College London.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
CDMS Computing Project Don Holmgren Other FNAL project members (all PPD): Project Manager: Dan Bauer Electronics: Mike Crisler Analysis: Erik Ramberg Engineering:
UTA Site Report DØrace Workshop February 11, 2002.
DØRACE Workshop Jae Yu Feb.11, 2002 Fermilab Introduction DØRACE Progress Workshop Goals Arrangements.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
Feb. 14, 2002DØRAM Proposal DØ IB Meeting, Jae Yu 1 Proposal for a DØ Remote Analysis Model (DØRAM) Introduction Partial Workshop Results DØRAM Architecture.
CD FY09 Tactical Plan Status FY09 Tactical Plan Status Report for Neutrino Program (MINOS, MINERvA, General) Margaret Votava April 21, 2009 Tactical plan.
Remote Institute Tasks Frank Filthaut 11 February 2002  Monte Carlo production  Algorithm development  Alignment, calibration  Data analysis  Data.
Maria Girone CERN - IT Tier0 plans and security and backup policy proposals Maria Girone, CERN IT-PSS.
Feb. 13, 2002DØRAM Proposal DØCPB Meeting, Jae Yu 1 Proposal for a DØ Remote Analysis Model (DØRAM) IntroductionIntroduction Partial Workshop ResultsPartial.
A UK Computing Facility John Gordon RAL October ‘99HEPiX Fall ‘99 Data Size Event Rate 10 9 events/year Storage Requirements (real & simulated data)
D0 File Replication PPDG SLAC File replication workshop 9/20/00 Vicky White.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
DØRACE Workshop Agenda Feb. 11, 2002 (Morning – 1West) 8:30 Registration at High Rise next to 1 West Conference room 9:00 Welcome (John W.) - Session Chair:
DØRACE Workshop Agenda Feb. 11, 2002 (Morning – 1West) 8:30 Registration at High Rise next to 1 West Conference room 9:00 Welcome Womersley 9:10-9:30 DØRACE.
UTA Site Report Jae Yu UTA Site Report 7 th DOSAR Workshop Louisiana State University Apr. 2 – 3, 2009 Jae Yu Univ. of Texas, Arlington.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
Apr. 25, 2002Why DØRAC? DØRAC FTFM, Jae Yu 1 What do we want DØ Regional Analysis Centers (DØRAC) do? Why do we need a DØRAC? What do we want a DØRAC do?
July 10, 2002DØRACE Status Report, Jae Yu DØ IB Meeting, OU Workshop 1 DØ RACE Status Report Introduction Software Distribution (DØRACE Setup) Hardware.
Grid Computing at NIKHEF Shipping High-Energy Physics data, be it simulated or measured, required strong national and trans-Atlantic.
Status report NIKHEF Willem van Leeuwen February 11, 2002 DØRACE.
Database Services at CERN Status Update
5th DOSAR Workshop Louisiana Tech University Sept. 27 – 28, 2007
Nuclear Physics Data Management Needs Bruce G. Gibbard
News & Action Items Automatic Release Ready notification system is in place and working (Thanks Alan!! ) Needs refinements!!! New Fermi RH CD-Rom released.
UM D0RACE STATION Status Report Chunhui Han June 20, 2002
DØRACE Workshop Summary
Status report NIKHEF Willem van Leeuwen February 11, 2002 DØRACE.
Remote Analysis Coordination
DØ Internal Computing Review
Comments on #3: “Motivation for Regional Analysis Centers and Use Cases” Chip Brock 3/13/2.
DØ RAC Working Group Report
Proposal for a DØ Remote Analysis Model (DØRAM)
Presentation transcript:

Jan. 17, 2002DØRAM Proposal DØRACE Meeting, Jae Yu 1 Proposal for a DØ Remote Analysis Model (DØRAM) IntroductionIntroduction Remote Analysis Station ArchitectureRemote Analysis Station Architecture Requirement for Regional Analysis CentersRequirement for Regional Analysis Centers Suggested Storage Equipment DesignSuggested Storage Equipment Design What Do I Think We Should Do?What Do I Think We Should Do? ConclusionsConclusions DØRACE Workshop Feb. 12, 2002 Jae Yu

Jan. 17, 2002DØRAM Proposal DØRACE Meeting, Jae Yu 2 Why do we need a DØRAM? Total Run IIa data sizes are –350TB for RAW – TB for Reco + root –1.4x10 9 Events total At the fully optimized 10sec/event reco.  1.4x10 10 Seconds for one time reprocessing Takes one full year w/ 500 machines –Takes ~8mos to transfer raw data for dedicated gigabit network Centralized system will do a lot of good but not sufficient (DØ analysis model proposal should be complemented with DØRAM) Need to allow remote locations to work on analysis efficiently Sociological benefits within the institutes Regional Analysis Centers should be established

Jan. 17, 2002DØRAM Proposal DØRACE Meeting, Jae Yu 3 Categorized remote analysis system set up by the functionality –Desk top only –A modest analysis server –Linux installation –UPS/UPD Installation and deployment –External package installation via UPS/UPD CERNLIB Kai-lib Root –Download and Install a DØ release Tar-ball for ease of initial set up? Use of existing utilities for latest release download –Installation of cvs –Code development –KAI C++ compiler –SAM station setup DØRACE Strategy Phase I Rootuple Analysis Phase 0 Preparation Phase II Executables Phase III Code Dev. Phase IV Data Delivery

Jan. 17, 2002DØRAM Proposal DØRACE Meeting, Jae Yu 4 Progressive

Jan. 17, 2002DØRAM Proposal DØRACE Meeting, Jae Yu 5 Central Analysis Center (CAC) Desktop Analysis Stations DAS …. DAS …. RAC …. Regional Analysis Centers Store & Process 10~20% of All Data IAC …. Institutional Analysis Centers IAC …. IAC Normal Interaction Communication Path Occasional Interaction Communication Path Proposed DØRAM Architecture

Jan. 17, 2002DØRAM Proposal DØRACE Meeting, Jae Yu 6 Regional Analysis Centers A few geographically selected sites that satisfy requirements Provide almost the same level of service as FNAL to a few institutional analysis centers Analyses carried out within the regional center –Store 10~20% of statistically random data permanently –Most the analyses performed on these samples with the regional network –Refine the analyses using the smaller but unbiased data set –When the entire data set is needed  Underlying Grid architecture provide access to remaining data set

Jan. 17, 2002DØRAM Proposal DØRACE Meeting, Jae Yu 7 Regional Analysis Center Requirements Become a Mini-CAC Sufficient computing infrastructure –Large bandwidth (gagibit or better) –Sufficient Storage Space to hold 10~20% of data permanently and expandable to accommodate data increase >30TB just for Run IIa RAW data –Sufficient CPU resources to provide regional or Institutional analysis requests and reprocessing Geographically located to avoid unnecessary network traffic overlap Software Distribution and Support –Mirror copy of CVS database for synchronized update between RAC’s and CAC –Keep the relevant copies of data bases –Act as SAM service station

Jan. 17, 2002DØRAM Proposal DØRACE Meeting, Jae Yu 8 Regional Storage Cache Disk Server IDE-RAID Gbit Switch IDE Hard drives are $1.5~$2./Gb Each IDE RAID array gives up to ~1TByte – hot swappable Can be configured to have up to 10TB in a rack Modest server can manage the entire system Gbit network switch provides high throughput transfer to outside world Flexible and scalable system Need an efficient monitoring and error recovery system Communication to resource management

Jan. 17, 2002DØRAM Proposal DØRACE Meeting, Jae Yu 9 What Do I Think We Should Do? Most the students and postDocs are at FNAL, thus it is important to provide them sufficient computing and cache resources for their analysis.  The Current suggestion for backend analysis clusters should be built!! In the mean time, we should select a few sites as RACs and prepare sufficient hardware and infrastructure –My rough map scan gives FNAL+3RACs in the US, and a few in Europe Software effort for Grid should proceed as fast as we can to supplement the hardware –We cannot afford to spend time for Test beds –Our set ups should be the Test Bed and the actual Grid A working group to determine number of RAC sites, their requirements, and select RACs within the next couple of months.

Jan. 17, 2002DØRAM Proposal DØRACE Meeting, Jae Yu 10 Suggestions and Comments from The Working Group Data characteristics –Specialized data set, in addition to service data set for reprocessing –Some level of replication should be allowed –Consistency of data must be ensured –Centralized organization of reprocessing –Book keeping of reprocessing Two staged approach: –Before Full gridification  All data kept in the CAC –After full gridification Fully distributed within the network Data sets are mutually exclusive

Jan. 17, 2002DØRAM Proposal DØRACE Meeting, Jae Yu 11 In Europe, some institutions are already in the works to become an RAC –Karlsruhe (Germany) –NIKHEF (Netherlands) –IN2P3, Lyon (France) We want more US participation Agreed to form a group to formulate RAC more systematically  Write up a document within 1-2 mos. –Functions –Services –Requirements –Etc.

Jan. 17, 2002DØRAM Proposal DØRACE Meeting, Jae Yu 12 Conclusions DØ must prepare for large data set era Need to expedite analyses in timely fashion Need to distribute data set throughout the collaboration Establishing regional analysis centers will be the first step toward DØ Grid Will write up a proposal for your perusal