ALICE analysis at GSI (and FZK) Kilian Schwarz CHEP 07.

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
T1 at LBL/NERSC/OAK RIDGE General principles. RAW data flow T0 disk buffer DAQ & HLT CERN Tape AliEn FC Raw data Condition & Calibration & data DB disk.
GSIAF "CAF" experience at GSI Kilian Schwarz. GSIAF Present status Present status installation and configuration installation and configuration usage.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
ALICE Operations short summary and directions in 2012 WLCG workshop May 19-20, 2012.
1 Status of the ALICE CERN Analysis Facility Marco MEONI – CERN/ALICE Jan Fiete GROSSE-OETRINGHAUS - CERN /ALICE CHEP Prague.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
LHCb Quarterly Report October Core Software (Gaudi) m Stable version was ready for 2008 data taking o Gaudi based on latest LCG 55a o Applications.
The ALICE Analysis Framework A.Gheata for ALICE Offline Collaboration 11/3/2008 ACAT'081A.Gheata – ALICE Analysis Framework.
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
CHEP – Mumbai, February 2006 The LCG Service Challenges Focus on SC3 Re-run; Outlook for 2006 Jamie Shiers, LCG Service Manager.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
ALICE Roadmap for 2009/2010 Patricia Méndez Lorenzo (IT/GS) Patricia Méndez Lorenzo (IT/GS) On behalf of the ALICE Offline team Slides prepared by Latchezar.
Experience with analysis of TPC data Marian Ivanov.
ALICE Upgrade for Run3: Computing HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
Status of the production and news about Nagios ALICE TF Meeting 22/07/2010.
What is expected from ALICE during CCRC’08 in February.
1. Maria Girone, CERN  Q WLCG Resource Utilization  Commissioning the HLT for data reprocessing and MC production  Preparing for Run II  Data.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
AliRoot survey P.Hristov 11/06/2013. Offline framework  AliRoot in development since 1998  Directly based on ROOT  Used since the detector TDR’s for.
The CMS CERN Analysis Facility (CAF) Peter Kreuzer (RWTH Aachen) - Stephen Gowdy (CERN), Jose Afonso Sanches (UERJ Brazil) on behalf.
23.March 2004Bernd Panzer-Steindel, CERN/IT1 LCG Workshop Computing Fabric.
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
Computing for Alice at GSI (Proposal) (Marian Ivanov)
Predrag Buncic ALICE Status Report LHCC Referee Meeting CERN
ATLAS Distributed Computing perspectives for Run-2 Simone Campana CERN-IT/SDC on behalf of ADC.
Predrag Buncic CERN ALICE Status Report LHCC Referee Meeting 01/12/2015.
Alignment in real-time in current detector and upgrade 6th LHCb Computing Workshop 18 November 2015 Beat Jost / Cern.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Analysis experience at GSIAF Marian Ivanov. HEP data analysis ● Typical HEP data analysis (physic analysis, calibration, alignment) and any statistical.
Data transfers and storage Kilian Schwarz GSI. GSI – current storage capacities vobox LCG RB/CE GSI batchfarm: ALICE cluster (67 nodes/480 cores for batch.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
Alien and GSI Marian Ivanov. Outlook GSI experience Alien experience Proposals for further improvement.
1 Reconstruction tasks R.Shahoyan, 25/06/ Including TRD into track fit (JIRA PWGPP-1))  JIRA PWGPP-2: Code is in the release, need to switch setting.
Main parameters of Russian Tier2 for ATLAS (RuTier-2 model) Russia-CERN JWGC meeting A.Minaenko IHEP (Protvino)
LHCb Current Understanding of Italian Tier-n Centres Domenico Galli, Umberto Marconi Roma, January 23, 2001.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
ALICE Physics Data Challenge ’05 and LCG Service Challenge 3 Latchezar Betev / ALICE Geneva, 6 April 2005 LCG Storage Management Workshop.
Some topics for discussion 31/03/2016 P. Hristov 1.
ALICE Full Dress Rehearsal ALICE TF Meeting 02/08/07.
Predrag Buncic CERN Plans for Run2 and the ALICE upgrade in Run3 ALICE Tier-1/Tier-2 Workshop February 2015.
Pledged and delivered resources to ALICE Grid computing in Germany Kilian Schwarz GSI Darmstadt ALICE Offline Week.
1 P. Murat, Mini-review of the CDF Computing Plan 2006, 2005/10/18 An Update to the CDF Offline Plan and FY2006 Budget ● Outline: – CDF computing model.
Grid Operations in Germany T1-T2 workshop 2015 Torino, Italy Kilian Schwarz WooJin Park Christopher Jung.
Availability of ALICE Grid resources in Germany Kilian Schwarz GSI Darmstadt ALICE Offline Week.
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
Kilian Schwarz ALICE Computing Meeting GSI, October 7, 2009
ALICE Computing Data Challenge VI
Workshop Computing Models status and perspectives
Report PROOF session ALICE Offline FAIR Grid Workshop #1
Status of the CERN Analysis Facility
INFN-GRID Workshop Bari, October, 26, 2004
GSIAF & Anar Manafov, Victor Penso, Carsten Preuss, and Kilian Schwarz, GSI Darmstadt, ALICE Offline week, v. 0.8.
ALICE Physics Data Challenge 3
Bernd Panzer-Steindel, CERN/IT
Status and Prospects of The LHC Experiments Computing
Readiness of ATLAS Computing - A personal view
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
AliRoot status and PDC’04
MC data production, reconstruction and analysis - lessons from PDC’04
GSIAF "CAF" experience at GSI
Analysis framework - status
ALICE Computing Model in Run3
ALICE Computing Upgrade Predrag Buncic
R. Graciani for LHCb Mumbay, Feb 2006
Support for ”interactive batch”
The LHCb Computing Data Challenge DC06
Presentation transcript:

ALICE analysis at GSI (and FZK) Kilian Schwarz CHEP 07

ALICE T2 use cases (see computing model) Three kinds of data analysis Fast pilot analysis of the data “just collected” to tune the first reconstruction at CERN Analysis Facility (CAF) Scheduled batch analysis using GRID (Event Summary Data and Analysis Object Data) End-user interactive analysis using PROOF and GRID (AOD and ESD) CERN Does: first pass reconstruction Stores: one copy of RAW, calibration data and first-pass ESD’s T1 Does: reconstructions and scheduled batch analysis Stores: second collective copy of RAW, one copy of all data to be kept, disk replicas of ESD’s and AOD’s T2 Does: simulation and end-user interactive analysis Stores: disk replicas of AOD’s and ESD’s

ALICE T2 – present status vobox LCG RB/CE GSI Batchfarm (39 nodes/252 cores for ALICE) & GSIAF(14 nodes) Directly attached disk storage (55 TB) ALICE::GSI::SE_tactical ::xrootd 30 TB ALICE::GSI::SE ::xrootd PROOF/ Batch Grid CERN GridKa 150 Mbps GSI

ALICE T2 – short term plans Extend GSIAF to all 39 nodes Extend GSIAF to all 39 nodes Study coexistence of interactive and batch processes on the same machines. Develop possibility to increase/decrease the number of batch jobs on the fly to give advantage to analysis. Study coexistence of interactive and batch processes on the same machines. Develop possibility to increase/decrease the number of batch jobs on the fly to give advantage to analysis. Add newly bought fileservers (about 120 TB disk space) to ALICE::LCG::SE::xrootd Add newly bought fileservers (about 120 TB disk space) to ALICE::LCG::SE::xrootd

Plans for the Alice Tier 2&3 at GSI: Remarks: Remarks: 2/3 of that capacity is for the tier 2 (ALICE central, fixed via WLCG MoU) 2/3 of that capacity is for the tier 2 (ALICE central, fixed via WLCG MoU) 1/3 for the tier 3 (local usage, may be used via Grid) 1/3 for the tier 3 (local usage, may be used via Grid) according to the Alice computing model no tape for tier2 according to the Alice computing model no tape for tier2 tape for tier3 independent of MoU tape for tier3 independent of MoU hi run in October -> upgrade operational: 3Q each year hi run in October -> upgrade operational: 3Q each year Year ramp-up CPU (kSI2k) 400/ / / / Disk (TB) 120/80300/200390/260510/ WAN (Mb/s)

Computing for Alice at GSI (Proposal) (Marian Ivanov)

Priorities ( ) Detector calibration and alignment (TPC- ITS-TRD) Detector calibration and alignment (TPC- ITS-TRD) –First test – Cosmic and Laser – October 2007 –To be ready for first pp collision First paper First paper –Time scale - Depends on success of October tests –Goal : ~ 1 week (statistic about 10^4- 10^5 events) ==> The calibration and alignment has the TOP priority ( ) ==> The calibration and alignment has the TOP priority ( )

Assumptions ● CPU requirements – Relative ● Simulation ~ 400 a.u ● Reconstruction ~ 100 a.u ● Alignment ~ 1 a.u ● Calibration ~ 1 a.u ● To verify and improve the calibration and alignment several passes through data are necessary The time scale for one iteration ~ minutes, hours ==> The time scale for one iteration ~ minutes, hours ==> ● The calibration and alignment algorithms should be decoupled from the simulation and reconstruction ● The reconstruction algorithm should be repeated after retuning of the calibration

Assumptions Type of analysis (requirements) Type of analysis (requirements) First priority First priority –Calibration of TPC – 10 ^4 -10^5 pp –Validation of the reconstruction - 10^4-10^5 pp –Alignment TPC, TPC-ITS – 10^5 pp + 10^4- 10^5 cosmic

Assumptions Alice test in October – (in one month) Alice test in October – (in one month) ● Full stress test of system ● Significant data volume ~20 Tby of raw data from test of 2 sectors (2006) ~20 Tby of raw data from test of 2 sectors (2006) Bottleneck (2006) – The processing time given by time of the data access - CPU time negligible Bottleneck (2006) – The processing time given by time of the data access - CPU time negligible We should be prepared for different scenarios We should be prepared for different scenarios ● We would like to start with the data copied at GSI and reconstruct/calibrate/align locally, later switch to GRID (The same we did in 2006) ● This approach enables several fast iteration over data

data transfers CERN GSI motivation: calibration modell and algorithms need to be tested before October motivation: calibration modell and algorithms need to be tested before October test the functionality of current T0/T1  T2 transfer methods. test the functionality of current T0/T1  T2 transfer methods. At GSI the CPU and storage resources are available, but how do we bring the data here ? At GSI the CPU and storage resources are available, but how do we bring the data here ?

analysis of TPC test data copy to GSI: copy to GSI: –store at ALICE::GSI::SE::xrootd out2547.list Cosmic Scan A0&1 77 out2548.list Cosmic Scan A0&1 67 out2557.list Cosmic Scan A0&1 82 out2574.list Cosmic Stability A0&1 265 out2660.list Cosmic Stability A4&5 313 out2641.list Cosmic Scan A4&5 138 out2642.list Cosmic Scan A4&5 97 out2643.list Cosmic Scan A4& Laser: out2572.list 31 out2657.list 171 out2728.list 195 out2798.list 215 out2906.list 90 out3094.list 6 directories out3189.list 4 directories out2612.list 114 out2686.list 177 out2746.list 41 out2851.list Job test data transfer to T2 and test SE preconditon:

data transfer CERN GSI

Software development Write component Write component Software validation - Sequence: Software validation - Sequence: 1)Local environment (first filter) 1) Stability – debugger 2) Memory consumption – valgrind, memstat (root) 3) CPU profiling - callgrind, vtune 4) Output – rough, quantitative – if possible 2)PROOF 1) For rapid development – fast user feedback 2) Iterative improvement of algorithms, selection criterias... 3) Improve statistic 3)production using GRID/ALIEN 1) Improve statistic 4)alternative scenario: local batch system 1) Memory consumption – valgrind, memstat 2) CPU profiling 3) Output - better statistic

analysis of test TPC data using various analysis techniques using various analysis techniques –local batch farm at GSI (read from ALICE::GSI::SE) (GSIAF) – copy data to ALICE::GSI::SE_tactical (PROOF cluster – directly attached disks) (submit to AliEn – jobs should arrive at GSI – since this is where the data are)

Proposal Algorithmic part of our analysis, calibration software should be independent of the running environment Algorithmic part of our analysis, calibration software should be independent of the running environment –TPC calibration classes (components) as example (running, tuning OFFLINE, used in HLT, DAQ and Offline) Analysis and calibration code should be written following a component based model Analysis and calibration code should be written following a component based model –TSelector (for PROOF) and AliAnalysisTask (at GRID/ALIEN) – just simple wrapper

analysis of test TPC data using Grid methods using Grid methods –analysis partition: GSI should be included –JDL: specify that CERN CE should not be used since data of interest are stored at CERN and GSI. Job should then take the other alternative.

analysis of TPC test data Executable="tpcRecAlienLocal.sh"; Executable="tpcRecAlienLocal.sh"; InputFiles={"LF:/afs/cern.ch/alice/tpctest/AliRoot/HEAD/TPC/recTPC.C","/afs/cern.ch /alice/tpctest/AliRoot/HEAD/TPC/AnalyzeESDtracks.C"}; InputFiles={"LF:/afs/cern.ch/alice/tpctest/AliRoot/HEAD/TPC/recTPC.C","/afs/cern.ch /alice/tpctest/AliRoot/HEAD/TPC/AnalyzeESDtracks.C"}; InputDataCollection="LF:/alice/cern.ch/user/h/haavard/jdl/runs/run$1.xml"; InputDataCollection="LF:/alice/cern.ch/user/h/haavard/jdl/runs/run$1.xml"; InputDataList="tpcRec.xml"; InputDataList="tpcRec.xml"; InputDataListFormat="xml-single"; InputDataListFormat="xml-single"; "tpc_archive.zip:FitSignal.root,TimeRoot.root,TPCsignal.root,TPCtracks.root,TPCdebu "tpc_archive.zip:FitSignal.root,TimeRoot.root,TPCsignal.root,TPCtracks.root,TPCdebu OutputDir="/alice/cern.ch/user/h/haavard/$2/$1/#alienfirstfilename#_dir"; OutputDir="/alice/cern.ch/user/h/haavard/$2/$1/#alienfirstfilename#_dir"; Split="file"; Split="file"; SplitArguments = {"#alienfirstfilename#"}; SplitArguments = {"#alienfirstfilename#"}; Arguments = " $1 $2 $3 "; Arguments = " $1 $2 $3 "; Workdirectorysize={"4000MB"}; Workdirectorysize={"4000MB"}; Requirements = ( !other.CE=="ALICE::CERN::LSF" ); Requirements = ( !other.CE=="ALICE::CERN::LSF" );

analyse TPC test data tpcrecAlienLocal.sh... command aliroot -q -b "$ALICE_ROOT/TPC/recTPC.C($runtype)" command aliroot -q -b "$ALICE_ROOT/TPC/recTPC.C($runtype)" command aliroot -q -b "$ALICE_ROOT/TPC/AnalyzeESDtracks.C+($run)" command aliroot -q -b "$ALICE_ROOT/TPC/AnalyzeESDtracks.C+($run)"...recTPC.C... AliReconstruction rec; AliReconstruction rec; rec.SetDefaultStorage("local://$ALICE_ROOT"); rec.SetDefaultStorage("local://$ALICE_ROOT"); rec.SetRunReconstruction("TPC"); rec.SetRunReconstruction("TPC"); rec.SetFillESD("TPC"); rec.SetFillESD("TPC"); rec.Run(0); rec.Run(0);...

0,003

Analysis at GridKa T1 centrally steered batch analysis centrally steered batch analysis preconditions: preconditions: –functioning SE with xrootd interface technically feasible with dCache technically feasible with dCache –GridKa hesitates so far to open xrootd ports to the world on all xrd doors security concept should be rediscussed security concept should be rediscussed

summary at GSI ALICE analysis will be done using Grid, PROOF, and local batch. at GSI ALICE analysis will be done using Grid, PROOF, and local batch. batch cluster and GSIAF are on the same machines. We will be able to increase/decrease batch/Grid jobs dynamically to give advantage to GSIAF/PROOF analysis if needed batch cluster and GSIAF are on the same machines. We will be able to increase/decrease batch/Grid jobs dynamically to give advantage to GSIAF/PROOF analysis if needed data transfer to GSI has still to be exercised data transfer to GSI has still to be exercised analysis performance on the Grid has to improve with respect to data I/O. analysis performance on the Grid has to improve with respect to data I/O. xrootd access to T1 centres has to be discussed xrootd access to T1 centres has to be discussed