Download presentation
Presentation is loading. Please wait.
Published byLillian Kennedy Modified over 8 years ago
1
1 Studies for setting up an Analysis Facility for ATLAS Data Analysis at CNRST and its application to Top Physics Acuerdo Bilateral CSIC-CNRST: 2007MA0057 José Salt Overview of the Spanish ATLAS Distributed TIER-2/IFIC
2
2 Overview 1.- Description of the Spanish ATLAS Tier-2 2.- The ATLAS Computing Model and our Tier-2 3.- Project Activities 4.- Ramp-up of resources 5.- Role of TIER-2 in the MA-SP Agreement
3
3 1.- Description of the Spanish ATLAS TIER-2 –The ATLAS distributed TIER-2 is conceived as a Computing Infrastructure for ATLAS experiment. Its main goals are: Enable Physics Analysis by Spanish ATLAS Users Continuous production of ATLAS MC events To contribute to ATLAS + LCG Computing Common Tasks Sustainable growth of infrastructure according to the scheduled ATLAS ramp-up and stable operation –Sequence of projects: ‘Acción Especial’ (2000-2001) LCG- oriented project (2002-2004) ATLAS Tier-2 project (phase I): 2005-2007 (2 years) ATLAS Tier-2 project (phase II): 2008-2010 (3 years)
4
4 The sites… IFIC IFAE UAM … and the resources CPU = 608 KSI2k Disk = 244 TBytes 14 FTE Equipment: Human Resources: (as of October 2007)
5
5 IFIC New Worker Nodes + Disk servers PC Farm (former one) Tape Storage Robot
6
6 UAM Storage devices Worker Nodes
7
7 IFAE Top view of the PIC Computer Room: where TIER-2 equipment is placed Worker Nodes
8
8 UAM: –José del Peso (PL) –Juanjo Pardo (*) –Luís Muñoz –Pablo Fernández (*T.MEC) IFAE: –Andreu Pacheco (PL) –Jordi Nadal (*) –Carlos Borrego (*) (<-) –Marc Campos –Hegoi Garaitogandia ( NIKHEF) IFIC –José Salt (PL and Coordinator) –Javier Sánchez (Tier-2 Operation Coordination) –Santiago González (Tier-3) –Alvaro Fernández (EGEE) –Mohammed Kaci (*T.MEC) –Gabriel Amorós (EGEE) –Alejandro Lamas –Elena Oliver (Ph. D. student) –Miguel Villaplana (Ph. D. student) –Luis March (postdoc, CERN) –Farida Fassi ( FR Tier-1) Total FTE = 14 FTEs Human Resources of the Spanish ATLAS TIER-2 MA_ES Agreement
9
9 UAM: –Construction and Commissioning of the ATLAS Electromagnetic End Cap Calorimeter –ATLAS Physics: Higgs search through 4 leptons decay mode IFIC: –Construction and Commissioning of the ATLAS Hadronic Calorimeter (TileCal) –Construction and Commissioning of the ATLAS SCT-Fwd. –Alignment of the Inner Detector –ATLAS Physics: b tagging algorithms for event selection; MC studies of different process beyond SM ( Exotics, Little Higgs and Extra Dimensions models); Top Physics IFAE: –Construction and Commissioning of ATLAS Hadronic Calorimeter (TileCal) –Development and Deployment of the ATLAS High Level Trigger: third level trigger software, infrastructure (Event Filter Farm); on-line commisioning of event selection software, tau trigger. –ATLAS Physics: TileCal Calibration, Reconstruction and Calibration of Jet/Tau/Missing Transverse Energy, SUSY searches, Standard Model processes, Charged Higgs. ATLAS Physics & Detectors (2008) Involvement of Spain in ATLAS (Det. & Phys)
10
10 2.- The ATLAS Computing Model and our TIER-2 Data will undergo transformations to obtain a reduction in size and the extraction of the relevant information Commitment of the participant ATLAS Institutes to contribute with their resources TIERed Structure: starting from a simple hierarchical organization of geographically scattered centers
11
11 RAL IN2P3 BNL FZK CNAF PIC ICEPP FNAL USC NIKHEF Krakow Legnaro IFAE Taipei TRIUMF IFIC UAM UB IFCA CIEMAT MSU Prague Budapest Cambridge small centres desktops portables Services of disk storage for data files and data bases To provide analysis capacity for the Physics Working groups. To conduct the operation of an installation of an ‘end-user’ Data Analysis System that gives service to at least 20 physics topics running in pararel To provide simulation according to the requirements of the experiments To provide Network services for the interchange of data with TIER-1 centers Tier-1 5 to ensure a sustainable growth of the TIER-2 Infrastructure distributed between these centers and its stable operations Tier-2 Tier-3 TIER-2 Funcionalities
12
12 Last grid use report @ WLCG Workshop
13
13 SAT-2 Activity Groups: SA: System Administrators MCP: Monte Carlo Production DAS: Distributed Analysis System US: User Support DMM: Data Movement Manager OM: Operation Manager PM: Project Management Contribution to CHEP’07: ‘Experience Running a Distributed Tier-2 for the ATLAS Experiment’ Spanish ATLAS Tier-2 3.- Project activities
14
14 Infrastructure: Hardware & Operations OM SA Operation Manager: OM System Administrators: SA Responsible of the Overall Coordination of TIER-2 (J.Sánchez) Design and development of the technical specification and policies to ensure the distributed TIER-2 will be seen as an unique virtual TiER-2 Technical link with the TIER-1’s Constellation Coordination : –Processing and storage resources in order to achieve an efficient and optimal operation of individual centers (fault-tolerance,response speed, etc) –Global policies of security and access –Establishment of global monitoring tools and policies to obtain usage metrics, estability of TIER-2 –Usage statistics and QoS to the responsibles of the project To manage the local cluster: hardware installation and maintenance, installation and configuration of the OS releases, GRID Mw updates, monitoring the cluster, solve problems detected by the local monitoring or the Global GRID Operation Center, etc Service Requirements: 12h/day, 5 days/week Now: 1’5 FTE/ site needed; usual service during ATLAS Data Taking: 6 FTE will be required
15
15 User Support (US) Site User Support has started at the end of 2006: first feedback from ‘end-users’ was studied at the 3 sites In next months it’s needed to go towards a ‘coordinated User Support’ for our distributed Tier-2: On-line/direct support Off-line support using a simple ticketing system and providing an internal tree-structure to solve different class of problems (T-2 Helpdesk To provide periodical tutorials/hands on software tools for analysis
16
16 DB EX CE SE CE Production System for Simulated Data (MC) : The ATLAS Production System manages the official massive productions of simulated data. The system consists of: - a data base (definition of jobs) - supervisors/executors (they take the jobs from the DB and are managed by the computing resources of ATLAS ) Since january 2008, the ATLAS collaboration has migrated to the unique executor : PANDA - Management of Distributed Data (the produced data are recorded and stored at different centers) Production of Simulated Data
17
17 Spanish physicists are doing physics analysis using GANGA on data distributed around the world This tool has been installed in our Tier2 infrastructure GANGA is an easy-to-use front-end for job definition, management and submission. Users interaction via: Its own Python shell (command line) A Graphical User Interface (GUI) In our case the jobs are sent to the LCG/EGEE Grid flavour. We are doing performance test: LCG Resource Broker (RB) gLite Workload Management System (WMS). New RB from EGEE Distributed Analysis System (DAS)
18
18 Data Movement Management(DMM) Data Management is one of the main activities inside the ATLAS Spanish federated Tier-2. AOD physics data are going to be stored at these sites (distributed geographically) Data Management main activities: –Data monitoring at federated Tier-2: –http://ific.uv.es/atlas-t2-eshttp://ific.uv.es/atlas-t2-es –Old and unnecessary data cleaning –Inconsistency check and cleaning up (catalogue entries and/or file size mismatch) –Data replication to Tier-2 for Tier-2 users (using DQ2 client and doing subscriptions) Coordination with Tier-1: –Bi-weekly Tier-1/Tier-2 coordination meetings where Tier- 1/Tier-2 data management issues are discussed –There is a collaboration project between Tier-1/Tier-2 to monitor some ATLAS/Grid services at both sites
19
19 Main objective: to achieve a more federative structure Organization & management of the project is responsibility of the subproject leaders –SL constitute the Project Management Board A. Pacheco (IFAE), J. del Peso (UAM) and J. Salt (IFIC) –PMB chairman is the Coordinator of the Project J. Salt (Tier-2 Coordinator) PMB meeting every 4-5 weeks Tier-2 Operation Meetings: Virtual, biweakly Presential Tier-2 meetings: every 6 months –February’06 : Valencia –October’06: Madrid –May’07: Barcelona –October’07: foreseen in Valencia –May’08: Madrid –November’08: Barcelona PROJECT MANAGEMENT (PM)
20
20 4.- Ramp-up of the TIER-2 Resources Ramp-up of Tier-2 Resources (after LHC rescheduling) numbers are cumulative Evolution of ALL ATLAS T-2 resources according to the estimations made by ATLAS CB (October 2006) Spanish ATLAS T-2 assuming a contribution of a 5% to the whole effort Strong increase of resources Spanish Tier-2 Size (October 2007): CPU = 608 KSI2k Disk = 244 TB
21
21 IFIC Networking UAM IFAE Connection at 1 Gbps to University backbone (10 Gbps) Universitat de València hosts the RedIris PoP in the Comunitat Valenciana Bought a new equipment: Cisco Catalyst 3506 Will connect at 10 Gpbs to University backbone Aggregate WNs and Diskservers Public IP addresses in the subnet 147.156.116.0 / 24 (to be increase to 23 bits) reserved to Tier2 Connection at 1 Gbps between UAM and RedIRIS new switch needed in UAM site to connect all servers at 1Gb/s Direct gigabit connection to PIC backbone (Cisco 6509E 100x1Gb, 8x10Gb) PIC hosts the 10 Gbs link to LHCOPN (CERN) and 2 Gbs to Tier2 / Internet
22
22 To increase surface from 90 m**2 to 150 m**2 Upgrade UPS: 50 KVA to 250 KVA To install 70 lines of 16 Amps ( 3 lines/rack) To increase the power for the building (Electrical 20KV transformer, DIESEL generator, Low Voltage distribution, …) To change the air conditioning (impulsion on technical floor) New racks Redistribution of all machines located at Computer Center Execution: in progress Upgrade of IFIC Computing Room new area
23
23 5.- Role of the Spanish ATLAS-T2 in the MA_ES Agreement ATLAS Collaboration Infrastructure (GRID framework) to plug in the Analysis Facilities (TIER-3) Support in term of expertise such as installing, configuring, tune, troubleshooting of a GRID Insfrastructure Project Activities To enable the work of Physics groups of both countries. – Analysis Use cases – collaboration in (several) Physics Analysis (Top Physics) Interface to the EGEE/EGI (European GRID Infrastructure) South West Federation ROC
24
24 Infrastructure Maintenance Production of Sim. Data (Public) Production of Sim. Data (Private) User Support Data Management Distributed Analysis Interactive Analysis Local Batch Facility End Users Interaction Scheme TIER-2 TIER-3
25
25 BACKUP SLIDES
26
26 Service Level Agreement To ensure the service availability of our TIER-2 : –12 hours/day ; 5 days/week To ensure a delay time in responding to operational problems Contribution to the M&O A and B (in 1 or 2 people) Service Maximum delay in responding to operational problems Averag e availab ility Measu red on annual basis Prime Time Other periods End-user analysis facility 2 hours 72 hours 95% Other services 12 hours 72 hours 95% Fellowhships: our groups have a high level training ability TTP: Technical Training Program
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.