CDF SAM Deployment Status Doug Benjamin Duke University (for the CDF Data Handling Group)

Slides:



Advertisements
Similar presentations
Status of BESIII Distributed Computing BESIII Workshop, Mar 2015 Xianghu Zhao On Behalf of the BESIII Distributed Computing Group.
Advertisements

Proposal for dCache based Analysis Disk Pool for CDF presented by Doug Benjamin Duke University on behalf of the CDF Offline Group.
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
Batch Production and Monte Carlo + CDB work status Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
Workload Management Massimo Sgaravatto INFN Padova.
F Fermilab Database Experience in Run II Fermilab Run II Database Requirements Online databases are maintained at each experiment and are critical for.
Database Infrastructure Major Current Projects –CDF Connection Metering, codegen rewrite, hep w/ TRGSim++ – Dennis –CDF DB Client Monitor Server and MySQL.
JIM Deployment for the CDF Experiment M. Burgon-Lyon 1, A. Baranowski 2, V. Bartsch 3,S. Belforte 4, G. Garzoglio 2, R. Herber 2, R. Illingworth 2, R.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
18 Feb 2004Computing Division Project Status Report1 Project Status Report : SAMGrid  SAMGrid Management, Status, Operations – Merritt  SAMGrid Development.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
L3 Filtering: status and plans D  Computing Review Meeting: 9 th May 2002 Terry Wyatt, on behalf of the L3 Algorithms group. For more details of current.
SOFTWARE & COMPUTING Durga Rajaram MICE PROJECT BOARD Nov 24, 2014.
GLAST LAT ProjectDOE/NASA Baseline-Preliminary Design Review, January 8, 2002 K.Young 1 LAT Data Processing Facility Automatically process Level 0 data.
S. Veseli - SAM Project Status SAMGrid Developments – Part I Siniša Veseli CD/D0CA.
The SAMGrid Data Handling System Outline:  What Is SAMGrid?  Use Cases for SAMGrid in Run II Experiments  Current Operational Load  Stress Testing.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London.
CDF Grid Status Stefan Stonjek 05-Jul th GridPP meeting / Durham.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
MiniBooNE Computing Description: Support MiniBooNE online and offline computing by coordinating the use of, and occasionally managing, CD resources. Participants:
1 st December 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
GridPP18 Glasgow Mar 07 DØ – SAMGrid Where’ve we come from, and where are we going? Evolution of a ‘long’ established plan Gavin Davies Imperial College.
International Workshop on HEP Data Grid Nov 9, 2002, KNU Data Storage, Network, Handling, and Clustering in CDF Korea group Intae Yu*, Junghyun Kim, Ilsung.
CDF Offline Production Farms Stephen Wolbers for the CDF Production Farms Group May 30, 2001.
Tape logging- SAM perspective Doug Benjamin (for the CDF Offline data handling group)
06/22/2005CDF Taking Stock CDF Taking Stock By Anil Kumar CD/CSS/DSG June 22, 2005.
Support in setting up a non-grid Atlas Tier 3 Doug Benjamin Duke University.
22 nd September 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
São Paulo Regional Analysis Center SPRACE Status Report 22/Aug/2006 SPRACE Status Report 22/Aug/2006.
Remote Control Room and SAM DH Shifts at KISTI for CDF Experiment 김현우, 조기현, 정민호 (KISTI), 김동희, 양유철, 서준석, 공대정, 김지은, 장성현, 칸 아딜 ( 경북대 ), 김수봉, 이재승, 이영장, 문창성,
16 September GridPP 5 th Collaboration Meeting D0&CDF SAM and The Grid Act I: Grid, Sam and Run II Rick St. Denis – Glasgow University Act II: Sam4CDF.
4 March 2004GridPP 9th Collaboration Meeting SAMGrid:JIM and CDF Development CDF Accepts the Need for the Grid –Requirements How to Meet the Need –Status.
Computing Division Requests The following is a list of tasks about to be officially submitted to the Computing Division for requested support. D0 personnel.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
Outline: Tasks and Goals The analysis (physics) Resources Needed (Tier1) A. Sidoti INFN Pisa.
GridPP11 Liverpool Sept04 SAMGrid GridPP11 Liverpool Sept 2004 Gavin Davies Imperial College London.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
SAM Overview (training session) for CDF Users Doug Benjamin Duke University Krzysztof Genser Fermilab/CD.
CD FY09 Tactical Plan Status FY09 Tactical Plan Status Report for Neutrino Program (MINOS, MINERvA, General) Margaret Votava April 21, 2009 Tactical plan.
The GridPP DIRAC project DIRAC for non-LHC communities.
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
Analysis Tools at D0 PPDG Analysis Grid Computing Project, CS 11 Caltech Meeting Lee Lueking Femilab Computing Division December 19, 2002.
General SAM Hints and Tips Adam Lyon (FNAL/CD/D0) GCAS Meeting 12/19/02 Removing Special RunsChecking Sam’s Health Cutting on Detector QualitySamRoot DB.
Adapting SAM for CDF Gabriele Garzoglio Fermilab/CD/CCF/MAP CHEP 2003.
MC Production in Canada Pierre Savard University of Toronto and TRIUMF IFC Meeting October 2003.
Pavel Nevski DDM Workshop BNL, September 27, 2006 JOB DEFINITION as a part of Production.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
The GridPP DIRAC project DIRAC for non-LHC communities.
Simulation Production System Science Advisory Committee Meeting UW-Madison March 1 st -2 nd 2007 Juan Carlos Díaz Vélez.
DBS Monitor and DAN CD Projects Report July 9, 2003.
Software framework and batch computing Jochen Markert.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
Joe Foster 1 Two questions about datasets: –How do you find datasets with the processes, cuts, conditions you need for your analysis? –How do.
DØ Grid Computing Gavin Davies, Frédéric Villeneuve-Séguier Imperial College London On behalf of the DØ Collaboration and the SAMGrid team The 2007 Europhysics.
Status of BESIII Distributed Computing BESIII Collaboration Meeting, Nov 2014 Xiaomei Zhang On Behalf of the BESIII Distributed Computing Group.
U.S. ATLAS Grid Production Experience
 YongPyong-High Jan We appreciate that you give an opportunity to have this talk. Our Belle II computing group would like to report on.
Level 1 Processing Pipeline
Presentation transcript:

CDF SAM Deployment Status Doug Benjamin Duke University (for the CDF Data Handling Group)

8-Nov-05D Benjamin - GDM mtg1 CDF’s definition of SAM Deployment Raw data logging only into SAM database schema Production Farm writes only to SAM User analysis of new data (collision/ simulated) available SAM only ( from my GDM talk 30-Aug-05) Beyond Deployment Common Root ntuples from production data CDF has two widely used (~ 80% of the collaboration) ntuple formats. Should be considered as production- level datasets.

8-Nov-05D Benjamin - GDM mtg2 SAM access to production data (Users) Major Success!!! Been using SAM v7 client and db servers since 15 September. > 300 TB collision data. Ntupling of data proceeding well. (> 26 TB of common ntuples produced) Applied the strategy of minimizing the impact on the users - to keep their productivity up User’s scripts required small changes Fully tested the examples ahead general usage. Good documentation

8-Nov-05D Benjamin - GDM mtg3 Deployment Status Completed Production farm User access to production data via SAM Incomplete Raw data logging Calibration ntuple creation executable still uses DFC schema Monitor backup cron job (Predator) to verify it does not move metadata; 4-6 weeks of monitoring (raw data too valuable to lose) (started on 1-Nov-05) MC upload - testing has begun

8-Nov-05D Benjamin - GDM mtg4 Common Ntuples and SAM (further issues) Root Ntuples are how most CDF users access the data. Two major ntuples types (Standard Ntuple and Top Ntuple ~ 80% CDF users) These Ntuples should be in the data handling system SAM use cases Batch - Users macros loop over files (like production files - diskcache_i) - want minimal changes to users’ macros Requires SAM interface in Root (based on C++ API) Interactive tests to ensure successful batch jobs (mimic batch tests interactively as a test) Interactive data exploration Requires SAM tool to download files to desktop

8-Nov-05D Benjamin - GDM mtg5 Appendix - Slides with further details

8-Nov-05D Benjamin - GDM mtg6 Current Status - Raw Data Logging Raw Data logging still requires three computers until CDF Consumer Server Logger (CSL) is upgraded (‘06 - ‘07) - CDF online responsibility. Dehong Zhang has made the system more robust Cross mounted NFS mounted disk used to transfer metadata (ascii files) between sgi machines and Linux SAM machine Sam meta data writing has automatic retry in case of errors (V7 client). Raw Data metadata still logged into Data File Catalogue (DFC) and SAM DB schemas Job to create ntuples for Calibrations - still uses DFC - Currently offline operations managers (Aidan Robson and Bernd Steltzer) volunteered to fix situation. ( Extraordinary effort that we are very grateful for!!! )

8-Nov-05D Benjamin - GDM mtg7 Current Status - Production Farm Production Farm - SAM based Can process > 22 M events/day Issues: Running v6 of SAM client/ DB server Working w/ SAMGrid team to test v6 SAM client/ v7 DB server. SAMGrid team as identified two changes to client API Migration of Prod. Farm to SAM v7 client proceeding slowly - Production group responsibility moving from the Taiwan group ( Suen H & Tsan L) to the UNM (Elena Vataga) Several weeks of running v7 SAM on test farm is required before using v7 client on the production farm. - to maintain robust farm performance. With smaller farm need to process > 50 M events to ensure success.

8-Nov-05D Benjamin - GDM mtg8 Current Status - MC Data Upload Monte Carlo generated offsite Generation 5 (v5.3.* offline software ) - uses DFC tools to save MC data on tape (and DFC schema) DFC schema is copied to SAM schema daily The Physics groups set the schedule for Gen 5 MC Generation 6 (v6.1.2> offline software 2005 > ) will use SAM tools (SAM_upload) Tools developed/ maintained by Armando Fella (and others from Italy) (SAM_upload ) CDF MC production group starting tests of SAM upload tools ahead of large scale MC production

8-Nov-05D Benjamin - GDM mtg9 Current Status - MC Upload (2) Currently CDF private Autodest server. The official SAM autodest server needed CDF features added Steve White, Randolph Herber and Valeria Bartsch worked implementing the CDF specific needs into the SAM version Testing has just begun My estimate: at least 6 weeks before CDF is using official SAM autodest server. Due to testing and script modifications (Sam_upload and Production farm scripts)