18-sep-021 Computing for CDF Status and Requests for 2003 Stefano Belforte INFN – Trieste.

Slides:



Advertisements
Similar presentations
Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
Advertisements

Resources for the ATLAS Offline Computing Basis for the Estimates ATLAS Distributed Computing Model Cost Estimates Present Status Sharing of Resources.
Duke Atlas Tier 3 Site Doug Benjamin (Duke University)
10 May 2002Report & Plans on computers1 Status Report for CDF Italy computing fcdfsgi2 disks Budget report CAF: now, next year GRID etc. CDF – Italy meeting.
Batch Production and Monte Carlo + CDB work status Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
David Adams ATLAS DIAL Distributed Interactive Analysis of Large datasets David Adams BNL March 25, 2003 CHEP 2003 Data Analysis Environment and Visualization.
Title US-CMS User Facilities Vivian O’Dell US CMS Physics Meeting May 18, 2001.
Sep Donatella Lucchesi 1 CDF Status of Computing Donatella Lucchesi INFN and University of Padova.
22/04/2005Donatella Lucchesi1 CDFII Computing Status OUTLINE:  New CDF-Italy computing group organization  Usage status at FNAL and CNAF  Towards GRID:
A tool to enable CMS Distributed Analysis
Windows Server 2008 Chapter 11 Last Update
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
L3 Filtering: status and plans D  Computing Review Meeting: 9 th May 2002 Terry Wyatt, on behalf of the L3 Algorithms group. For more details of current.
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
The D0 Monte Carlo Challenge Gregory E. Graham University of Maryland (for the D0 Collaboration) February 8, 2000 CHEP 2000.
08/06/00 LHCb(UK) Meeting Glenn Patrick LHCb(UK) Computing/Grid: RAL Perspective Glenn Patrick Central UK Computing (what.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
The SAMGrid Data Handling System Outline:  What Is SAMGrid?  Use Cases for SAMGrid in Run II Experiments  Current Operational Load  Stress Testing.
October, Scientific Linux INFN/Trieste B.Gobbo – Compass R.Gomezel - T.Macorini - L.Strizzolo INFN - Trieste.
LcgCAF:CDF submission portal to LCG Federica Fanzago for CDF-Italian Computing Group Gabriele Compostella, Francesco Delli Paoli, Donatella Lucchesi, Daniel.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
BaBar Grid Computing Eleonora Luppi INFN and University of Ferrara - Italy.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London.
3rd Nov 2000HEPiX/HEPNT CDF-UK MINI-GRID Ian McArthur Oxford University, Physics Department
EGEE is a project funded by the European Union under contract IST Testing processes Leanne Guy Testing activity manager JRA1 All hands meeting,
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
8th November 2002Tim Adye1 BaBar Grid Tim Adye Particle Physics Department Rutherford Appleton Laboratory PP Grid Team Coseners House 8 th November 2002.
Nick Brook Current status Future Collaboration Plans Future UK plans.
A Design for KCAF for CDF Experiment Kihyeon Cho (CHEP, Kyungpook National University) and Jysoo Lee (KISTI, Supercomputing Center) The International Workshop.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
GridPP18 Glasgow Mar 07 DØ – SAMGrid Where’ve we come from, and where are we going? Evolution of a ‘long’ established plan Gavin Davies Imperial College.
9 February 2000CHEP2000 Paper 3681 CDF Data Handling: Resource Management and Tests E.Buckley-Geer, S.Lammel, F.Ratnikov, T.Watts Hardware and Resources.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
3 Apr 2002Stefano Belforte – INFN Trieste Necessita’ CDF al Tier11 CDF needs at Tier 1 Many details in slides for (your) future reference Will move faster.
Febryary 10, 1999Stefano Belforte - INFN Trieste1 CDF Run II Computing Workshop. A user’s perspective Stefano Belforte INFN - Trieste.
16 September GridPP 5 th Collaboration Meeting D0&CDF SAM and The Grid Act I: Grid, Sam and Run II Rick St. Denis – Glasgow University Act II: Sam4CDF.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
Stefano Belforte INFN Trieste 1 CMS Simulation at Tier2 June 12, 2006 Simulation (Monte Carlo) Production for CMS Stefano Belforte WLCG-Tier2 workshop.
Outline: Tasks and Goals The analysis (physics) Resources Needed (Tier1) A. Sidoti INFN Pisa.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
GridPP11 Liverpool Sept04 SAMGrid GridPP11 Liverpool Sept 2004 Gavin Davies Imperial College London.
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
AliEn AliEn at OSC The ALICE distributed computing environment by Bjørn S. Nilsen The Ohio State University.
May Donatella Lucchesi 1 CDF Status of Computing Donatella Lucchesi INFN and University of Padova.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
Outline: Status: Report after one month of Plans for the future (Preparing Summer -Fall 2003) (CNAF): Update A. Sidoti, INFN Pisa and.
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
12-nov-021 CDF Computing Politics and Technicalities Stefano Belforte INFN – Trieste.
Adapting SAM for CDF Gabriele Garzoglio Fermilab/CD/CCF/MAP CHEP 2003.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Hans Wenzel CDF CAF meeting October 18 th -19 th CMS Computing at FNAL Hans Wenzel Fermilab  Introduction  CMS: What's on the floor, How we got.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
CDF SAM Deployment Status Doug Benjamin Duke University (for the CDF Data Handling Group)
CDF ICRB Meeting January 24, 2002 Italy Analysis Plans Stefano Belforte - INFN Trieste1 Strategy and present hardware Combine scattered Italian institutions.
BaBar & Grid Eleonora Luppi for the BaBarGrid Group TB GRID Bologna 15 febbraio 2005.
1 P. Murat, Mini-review of the CDF Computing Plan 2006, 2005/10/18 An Update to the CDF Offline Plan and FY2006 Budget ● Outline: – CDF computing model.
Readiness of ATLAS Computing - A personal view
Proposal for the LHCb Italian Tier-2
Support for ”interactive batch”
Gridifying the LHCb Monte Carlo production system
Linux Cluster Tools Development
Presentation transcript:

18-sep-021 Computing for CDF Status and Requests for 2003 Stefano Belforte INFN – Trieste

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing2 The CDF-Italy Computing Plan Presented on June 24, 2002 Referees (and CSN1) postponed discussion/approval until November 2002: decide based on experience  Collecting experience now  No reason to modify plan so far Today:  Status report on analysis farm at FNAL  Update on work toward de-centralization  GRID - CNAF  Progress toward MOU/MOF  Rational for 2003 requests

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing3 Status of CAF FNAL Central Analysis Farm (CAF): a big success so far  Easy to use  Effective  Convenient Measure of success  100% used now  Upgrade in progress  Many institutions spending their $$$ there  Cloning started (Korea)

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing4 CDF Central Analysis Farm Compile/link/debug everywhere Submit from everywhere FNAL  Submission of N parallel jobs with single command  Access data from CAF disks now  Access tape data via transparent cache soon now Get job output everywhere Store small output on local scratch area for later analysis Access to scratch area from everywhere IT WORKS NOW FNAL Local Data serversA pile of PC’s My Desktop My favorite Computer gateway ftp switch job Log out NFS rootd N jobs rootd scratch server

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing5 Tape to Disk to CPU 2TB/day From disk From tape Days in September 2002 “Spec. from 2000 review”: Disk cache should satisfy 80% of all data requests

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing6 CAF promise fulfilled Giorgio Chiarelli runs 100 section jobs and integrates 120x7x24x3% = 600 CPU hours in a few days using up to more then half the full CAF at the same time  Go through 1TB of data in a few hours All of this with one single few lines script that automatically divides the input among the various job sections Made in Italy

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing7 Monitoring jobs and sections on the Web Made in Italy

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing8 Managing user’s area on CAF O(100GB) Made in Italy

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing9 CAF this summer CAF stage 1 saved the day for summer conferences  61 duals (10 INFN 16Pitt/CMU)  15 fileservers (4 INFN 1 MIT) CPU usage ~90% since June Users happy Made in Italy

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing10 CAF today Wait times get longer Users want more Ready for Stage 2 New hardware ready this fall for ski conferences Made in Italy

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing11 CAF Stage 2 (Stage1  x4) FNAL/CD centralized bid ~ two times/year CDF procurement for Stage 2 this summer  JustInTime to catch INFN funds released in June (x3) Bids are in  Hope for HW up and running in November CSN1  users = 6 months Many others will join CAF in Stage2  KEK-Japan: 2 fileservers 38 duals  Korea : 0.5 fileserver (+ 2 later)  Spain : 1 fileserver  Canada : 1 fileserver  US (8 universities) : 10 fileservers 4 duals  More to come

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing12 Why is CAF a success CAF is more than a pile of PC’s Integrated hw/sw design for farm and access tools Designed for optimized access to data  Lots of disk resident data  Large transparent disk cache in front of tape robot  Tuning of disk access (data striping, minimal NFS,…) Designed for users convenience  Simple GUI’s, Kerberos based authentication, large local user areas Professional system management and close loop with vendors  Several hw/firmware/sw problems solved so far  RAID controller, defective RAM, file system or kernel bugs …  Plus the normal failure rate of disks, power supplies etc.  2 FTE on CAF infrastructure

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing13 Will CAF success last ? User community:  ramping up in these days: 20  200  From the pioneers to the masses  Exposure to all kinds of access patterns Hardware expansion:  up to a factor 10 over the next 2 years Only experience will tell  CAF is build with the cheapest hardware  Will have to learn to live with 10~20% of hardware broken at any given time

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing14 Beyond CAF FERMILAB wants to join the GRID  FNAL will be Tier1 for CMS-US Foreign CDF institutions want to integrate their local farms  Spain, Korea, UK, Germany, Canada, Italy  In many case to exploit LHC/GRID hardware So far no big offer of help for common work, unlikely D0  Exception: Canada: 224 nodes “now” for CDF MC No software tool to do this integration “transparently” yet Not clear how much this will help CDF analysis

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing15 Decentralizing analysis computing FNAL-CD working hard to promote SAM for remote work  SAM: Metadata catalog + distributed disk caches  Run analysis locally  Copy data as needed (only 1 st time)  Works in Trieste (as other places)  SAM to become “the” CDF data access tool  SAM integration with (EuroData)GRID being tried CDF working on “packaging CAF for export”  Decentralized CAFs  Each handling data independently  Cloning FNAL CAF is the easiest way (Korea choice) Remote farms = extra costs for FNAL

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing16 CDF computing outside US (approx) Notes TBdualsTBduals Spain Shared with CMS, plan for EDG tools No plan for shared access Germany Tier1 (shared with LHC) + Tier3 (CDF) No plan for shared access Testing SAM on Tier3 UK (4 sites) Maybe 5x the CPU if 8-way  duals No EDG, Kerberos for user access, SAM for data. maybe open Korea Want to clone CAF by end of 2002 Kerberos for user access, open to all Start w/o SAM Canada No GRID tools Run official CDF MC and copy to FNAL Italy15729 No plan for shared access Exploring SAM on single node

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing17 MOU/MOF Moving to a way to recognize foreign contribution IFC and Scrutiny Group to work on this  INFN present in both Issues being talked about:  Computing will have to enter MOF somehow  Allow and encourage contribution  Take into account history and present situation No indication of a “crisis” that has to be dumped on the collaborators for help

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing requests detailed: 5 items Stick to June plan : 1)Invest majority of resources on FNAL CAF 2)Modest growth in Italy for interactive work Summer experience: needs do not scale down with luminosity  No reason to expect large variation from June numbers  Requested resources well within June forecast  Nevertheless, prudent, incremental approach (  referees) New in )Start MC 4)Interactive work at FNAL 5)Start transition to CNAF

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing19 Tevatron keeps us busy By next summer tune analysis to same level as Run1:  Alignments, precision tracking, secondary vertexes, B-tag  Jet energy corrections, underlying event Do interesting physics in the meanwhile Example: All italian D  hh  By end of year (100pb -1 )  10^6 events in the mass peak, 10^7 in the histogram  4TB of data by spring, 16TB by end 2003  This channel alone saturates disk financed so far (15TB)  Learning field for B  hh

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing20 Monte Carlo CDF has talked about central production  But no overall estimate of needs yet Next year safe bet: everybody on his/her own  Just the same as Run 1 Italian groups starting on this now Plan for capacity of 10^7 events/months  Modest hw need: 10 dual cpu nodes  Adequate for most analysis (10x a given dataset)  Future growth should be small Further requests only on basis of clear “cases”

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing21 Interactive work at FNAL When at FNAL can not run root on Italy’s machines  Need “some” “better then desktop” PC (Cfr. June’s talk) Referees asked for central management:  Defined total cap at 10 “power PCs”  Asked for 5 in 2003  4 full time physicists doing analysis at FNAL P.Azzi, R.Carosi, S.Giagu, M.Rescigno  Explore central alternative in 2003  Interactive login pool in CAF  Some ideas so far, will try and see

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing22 Moving CAF to CNAF Spend money in Italy Join INFN effort in building world class computing center Easier access to 3 rd data and/or interactive resources  GARR vs WAN Tap on GRID/LHC hardware pool for peak needs Import here tools and experience learnt on CAF Not an “experiment need”  FNAL CAF may be enough Costs more Poor access to main data repository (FNAL tapes) Need to replicate easy of use and operation of FNAL CAF Different hardware = different problems Have to divert time and effort from data analysis PRO’sCON’s

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing23 Moving CAF to CNAF: the proposal Start with limited, but significant hardware  2003 at CNAF  ½ of private share of CAF in 2002  7TB of disk and 29 dual processor estimated on the basis of expected data needs for top  6j and Z  bbar Explore effectiveness of work environment  Don’t give up on CAF features  Look for added value  Will need help (manpower) Will try and see, decision to leave FNAL will have to be based on proof of existence of valid alternative here

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing24 Summary of requests June 24 “plan” After CSN1’s June decision Analysis at FNAL FNAL CAF: 22TB disk + 63 dual nodes = =306KEu Monte Carlo: 10 dual nodes = 28KEu (FNAL price) CNAF: 7TB disk + 29 dual nodes = 70+96=166KEu Interactive FNAL: 5 “power PC” = 22.5KEu Interactive Italy: disk and cpu Pd/Pi/Rm/Ts/… = 50KEu total

18-sep-0225 SPARE Spare slides from here on

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing26 Working on CDF CAF is easy 1.Pick a dataset by name 2.Decide how many parallel execution threads (sections) 3.Prepare 1 executable, 1 tcl and 1 script file Submit from anywhere via simple GUI Query CAF status at any time via web monitor Retrieve log/data anywhere via simple GUI 2 step submission of 100 sections 1) In the script: setenv TOT_SECT section = $1 - 1 setenv CAF_SECTION $section 2) In the tcl file (only one tcl file) module talk DHInput include dataset bhmu03 setInput cache=DCACHE splitInput slots=$env(TOT_SECT) this=$env(CAF_SECTION)

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing27 Working on CAF is effective Quickly go through any CDF dataset (disk or tape) Create personalized output and store it locally Run on that output (data file or root ntuple)  Locally on CAF nodes  Remotely via rootd (e.g. Root from desktop)

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing28 CAF is convenient: can work from anywhere All needed code and tools for CDF offline via anonymous ftp or simply from /afs/infn.it  Everything runs on plain RedHat 6.x, 7.x  even on GRID testbed  no need for customized system install Need Kerberos ticket to talk to FNAL, but..  One click install of kerberos client from the web  No need for system manager  Just type “kinit” and your Fermilab password Many people work from their laptop !

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing29 CAF future

CSN1 18-sep-02 Stefano Belforte – INFN Trieste CDF computing30 Little data ? No way ! DAQ runs at full speed Typical Luminosity better then Run1 2 track trigger from SVT is full of charm We are refocusing attention on samples that in the default scenario would have been limited in statistics  Low Pt jets (20GeV) and leptons (8GeV)  Charm  Interesting for physics  improve on PDG in charm sector  Fundamental control samples  Particle ID on D  hh as learning field for B  hh  Heavy flavor content in jets  B-jet tagging  Jets resolution  …