Physics Data Management at CERN

Slides:



Advertisements
Similar presentations
Bernd Panzer-Steindel, CERN/IT WAN RAW/ESD Data Distribution for LHC.
Advertisements

Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
Resources for the ATLAS Offline Computing Basis for the Estimates ATLAS Distributed Computing Model Cost Estimates Present Status Sharing of Resources.
T1 at LBL/NERSC/OAK RIDGE General principles. RAW data flow T0 disk buffer DAQ & HLT CERN Tape AliEn FC Raw data Condition & Calibration & data DB disk.
Randall Sobie The ATLAS Experiment Randall Sobie Institute for Particle Physics University of Victoria Large Hadron Collider (LHC) at CERN Laboratory ATLAS.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Business Continuity and DR, A Practical Implementation Mich Talebzadeh, Consultant, Deutsche Bank
CERN IT Department CH-1211 Genève 23 Switzerland Visit of Professor Jerzy Szwed Under Secretary of State Ministry of Science and Higher.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 15 th April 2009 Visit of Spanish Royal Academy.
Les Les Robertson WLCG Project Leader WLCG – Worldwide LHC Computing Grid Where we are now & the Challenges of Real Data CHEP 2007 Victoria BC 3 September.
Les Les Robertson LCG Project Leader LCG - The Worldwide LHC Computing Grid LHC Data Analysis Challenges for 100 Computing Centres in 20 Countries HEPiX.
Introduction and Overview Questions answered in this lecture: What is an operating system? How have operating systems evolved? Why study operating systems?
Frédéric Hemmer, CERN, IT DepartmentThe LHC Computing Grid – October 2006 LHC Computing and Grids Frédéric Hemmer IT Deputy Department Head October 10,
José M. Hernández CIEMAT Grid Computing in the Experiment at LHC Jornada de usuarios de Infraestructuras Grid January 2012, CIEMAT, Madrid.
Frédéric Hemmer, CERN, IT Department The LHC Computing Grid – June 2006 The LHC Computing Grid Visit of the Comité d’avis pour les questions Scientifiques.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
INFSO-RI Enabling Grids for E-sciencE Geant4 Physics Validation: Use of the GRID Resources Patricia Mendez Lorenzo CERN (IT-GD)
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 25 th April 2012.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
Ian Bird LCG Deployment Manager EGEE Operations Manager LCG - The Worldwide LHC Computing Grid Building a Service for LHC Data Analysis 22 September 2006.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
1 The LHC Computing Grid – February 2007 Frédéric Hemmer, CERN, IT Department LHC Computing and Grids Frédéric Hemmer Deputy IT Department Head January.
CERN IT Department CH-1211 Genève 23 Switzerland Visit of Professor Karel van der Toorn President University of Amsterdam Wednesday 10 th.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
CERN – IT Department CH-1211 Genève 23 Switzerland t Working with Large Data Sets Tim Smith CERN/IT Open Access and Research Data Session.
Service, Operations and Support Infrastructures in HEP Processing the Data from the World’s Largest Scientific Machine Patricia Méndez Lorenzo (IT-GS/EIS),
23.March 2004Bernd Panzer-Steindel, CERN/IT1 LCG Workshop Computing Fabric.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
David Foster LCG Project 12-March-02 Fabric Automation The Challenge of LHC Scale Fabrics LHC Computing Grid Workshop David Foster 12 th March 2002.
CASTOR project status CASTOR project status CERNIT-PDP/DM October 1999.
The LHC Computing Grid Visit of Dr. John Marburger
1 The LHC Computing Grid – April 2007 Frédéric Hemmer, CERN, IT Department The LHC Computing Grid A World-Wide Computer Centre Frédéric Hemmer Deputy IT.
GDB, 07/06/06 CMS Centre Roles à CMS data hierarchy: n RAW (1.5/2MB) -> RECO (0.2/0.4MB) -> AOD (50kB)-> TAG à Tier-0 role: n First-pass.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
IT-DSS Alberto Pace2 ? Detecting particles (experiments) Accelerating particle beams Large-scale computing (Analysis) Discovery We are here The mission.
A Computing Tier 2 Node Eric Fede – LAPP/IN2P3. 2 Eric Fede – 1st Chinese-French Workshop Plan What is a Tier 2 –Context and definition To be a Tier 2.
Top 5 Experiment Issues ExperimentALICEATLASCMSLHCb Issue #1xrootd- CASTOR2 functionality & performance Data Access from T1 MSS Issue.
CASTOR: possible evolution into the LHC era
Dynamic Extension of the INFN Tier-1 on external resources
“Replica Management in LCG”
WLCG Tier-2 Asia Workshop TIFR, Mumbai 1-3 December 2006
“A Data Movement Service for the LHC”
Grid site as a tool for data processing and data analysis
IT Department and The LHC Computing Grid
The LHC Computing Grid Visit of Mtro. Enrique Agüera Ibañez
Jan 12, 2005 Improving CMS data transfers among its distributed Computing Facilities N. Magini CERN IT-ES-VOS, Geneva, Switzerland J. Flix Port d'Informació.
Data Challenge with the Grid in ATLAS
Bernd Panzer-Steindel, CERN/IT
The LHC Computing Challenge
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
The LHC Computing Grid Visit of Her Royal Highness
Ákos Frohner EGEE'08 September 2008
The INFN Tier-1 Storage Implementation
Large Scale Test of a storage solution based on an Industry Standard
Visit of US House of Representatives Committee on Appropriations
LHC Data Analysis using a worldwide computing grid
Using an Object Oriented Database to Store BaBar's Terabytes
The LHC Computing Grid Visit of Prof. Friedrich Wagner
Gridifying the LHCb Monte Carlo production system
INFNGRID Workshop – Bari, Italy, October 2004
Overview & Status Al-Ain, UAE November 2007.
The LHC Computing Grid Visit of Professor Andreas Demetriou
The LHCb Computing Data Challenge DC06
Presentation transcript:

Physics Data Management at CERN Alberto Pace IT Data Management Group Leader January 2009

150 million sensors deliver data … … 40 million times per second View of the ATLAS detector (under construction) 150 million sensors deliver data … … 40 million times per second Alberto Pace, CERN, IT Department

Distribution of CERN users (Feb 2008) Alberto Pace, CERN, IT Department

Alberto Pace, CERN, IT Department The LHC Data Challenge The accelerator will be completed in 2008 and run for 10-15 years Experiments will produce about 15 Million Gigabytes (PB) of data each year (about 2 million DVDs!) LHC data analysis requires a computing power equivalent to ~100,000 of today's fastest PC processors Requires many cooperating computer centres, as CERN can only provide ~20% of the capacity Alberto Pace, CERN, IT Department

Frédéric Hemmer, CERN, IT Department

Frédéric Hemmer, CERN, IT Department

Alberto Pace, CERN, IT Department CPU Disk Tape Alberto Pace, CERN, IT Department

Alberto Pace, CERN, IT Department Solution: the Grid Use the Grid to unite computing resources of particle physics institutes around the world The World Wide Web provides seamless access to information that is stored in many millions of different geographical locations The Grid is an infrastructure that provides seamless access to computing power and data storage capacity distributed over the globe Alberto Pace, CERN, IT Department

Alberto Pace, CERN, IT Department How does the Grid work? It makes multiple computer centres look like a single system to the end-user Advanced software, called middleware, automatically finds the data the scientist needs, and the computing power to analyse it. Middleware balances the load on different resources. It also handles security, accounting, monitoring and much more. Alberto Pace, CERN, IT Department

Alberto Pace, CERN, IT Department LCG Service Hierarchy Tier-0: the accelerator centre Data acquisition & initial processing Long-term data curation Distribution of data  Tier-1 centres Canada – Triumf (Vancouver) France – IN2P3 (Lyon) Germany – Forschunszentrum Karlsruhe Italy – CNAF (Bologna) Netherlands – NIKHEF/SARA (Amsterdam) Nordic countries – distributed Tier-1 Spain – PIC (Barcelona) Taiwan – Academia SInica (Taipei) UK – CLRC (Oxford) US – FermiLab (Illinois) – Brookhaven (NY) Tier-1: “online” to the data acquisition process  high availability Managed Mass Storage –  grid-enabled data service Data-heavy analysis National, regional support Tier-2: ~200 centres in ~35 countries Simulation End-user analysis – batch and interactive Alberto Pace, CERN, IT Department

Frédéric Hemmer, CERN, IT Department

Alberto Pace, CERN, IT Department WLCG Grid Activity in 2007 WLCG ran ~ 44 million jobs in 2007 – workload has continued to increase Distribution of work across Tier0 / Tier1 / Tier 2 really illustrates the importance of the grid system Tier 2 contribution is around 50%; > 85% is external to CERN Data distribution from CERN to Tier-1 sites Latest test in February show that the data rates required for LHC start-up have been reached and can be sustained over long periods http://gridview.cern.ch Alberto Pace, CERN, IT Department

Data Management Areas of action Tier-0 Data Management and Castor Software for the CERN computer Centre Grid Data Management middleware Software for Tier1 and Tier2 centres Physics Database services Database services for the software above and analysis Persistency Framework Software to ensure that physics application are independent from database vendors

Atlas Storage setup T1 – T2 2 TB MC TAPE 15 TB DISK 6 TB GROUPDISK USERDISK HIT CPUs @Tier-1 @Tier-2 PRODDISK Pile-up digitization reconstruction G4 and ATLFAST Simulation 120 TB AOD HITS AOD from ATLFAST HITS from G4 25 TB EVNT All other T1’s User analysis Group analysis On request DPD1 making DPD1 DPD2 Courtesy of Kors Bos

Storage disk pools for analysis Courtesy of Bernd Panzer-Steindel

Dataflow working model of the LHC experiments Courtesy of Bernd Panzer-Steindel

Data management challenge Provide basic building blocks to empower the experiments to build custom data workflows (especially for analysis) Data pools with different quality of services Also called the “Storage Element” (SE) Tools for “aggregated” data transfer and data migration between pools

Components of Storage Elements Store Data (in the form of files) Make the data available to the computing nodes (CE = Computing Element) Interface with the grid Standard and well defined I/O Protocols to access data RFIO, XROOTD, GridFtp, Mountable file system Standard and well defined protocols to manage the SE SRM Integrate with other systems, in the context of a particular sie Offline storage (i.e., MSS) Tape access D1T0

Typical physics analysis scenario Computing Elements Random I/O Storage Elements Bulk / Sequential I/O The Grid (other sites)

Storage Element Software Linearly Scalable Limited only by network bandwidth To increases capacity or performance, just add hardware through-put proportional to the number of clients Secure Easy to install, configure, and maintain Independent from hardware changes, from OS upgrades and from third party software Integrated monitoring and extensive understandable logging to understand performance issues Hardware requirements based on low cost commodity items

Sometime SE becomes complicated Castor implementation at CERN

Types of Storage Elements Data pools of different quality of services D1T0 – Disk only pool (no tape copy, or with tape copy implemented by the experiment using transfer tools on next slide) D1T1 – Disk pool with automated tape backup DnT0 / DnT1 – replicated disk with or without tape copy D0T1 – Here it get’s tricky – See later Disk cache GC Tape write Tape Read D0T1 D1T0 D2T1 D2T0 D1T0 D1T1 D1T0

D0T1 is tricky What does D0 means ? That the disk pool is (arbitrarily) smaller than the tape pool What is the role of the small disk pool ? a “buffer” to tape operation ? a “cache” of tape media ? The software policy (garbage collector) decides which files (and in which order) are delete when the small disk pool becomes full. Can be: Files that have been written to tape Files that have been recalled from tape and accessed Files that are larger in size Files that are older GC Tape write Tape Read D0T1 Disk cache

The complexity of D0T1 The garbage collector requires tuning and optimization to avoid severe, non-linear, performance drops. It is the core of the Data Management project itself ! One size fit all is not good enough there are many parameters to tune We have multiple “pre-packaged” scenarios. Example: “D0T1 for write” “D0T1 for read” “D0T1 generic” (for backward compatibility) ... And possibly others “D0T1 Write” Disk buffer Tape write GC Simple Garbage collection policy (written files can be deleted)

Important constraints... Avoid both reading and writing from the same D0T1 storage class. Allow combining two classes on the same tape pool as a workaround If tape access is aggregated, there will be a reduced need for “D0T1 Generic” Disk is more a temporary buffer rather then a cache to tape access “D0T1 generic” “D0T1 Write” + “D0T1 Read” Disk cache Disk buffer Disk buffer GC GC GC Tape write Simpler Garbage collection policies, easier to understand & debug Tape Read Tape write Tape Read

Data transfer and synchronization Data transfer and data migration between pools Across WAN and LAN One way (master / slave), Two ways (multiple masters) Two ways is straightforward if files are not modified. Can be also done to support file modifications ... Understand “aggregated” data transfers Concept of “Data Sets” Offer data movements “immediately after data change” (data are synchronized) “At periodic intervals” (“pool A” contains “pool B” data from yesterday/last week/...) “Manually” (the experiment recalls on disk pool data from 3 years ago from tape pool) “Custom” (the experiment scripts his transfer policy)

General strategy ... Provide basic building blocks (storage class and transfer/replication services) to empower experiments to build their data analysis process Building blocks with functionalities easy to understand Building blocks that could be instantiated “on the fly” Building blocks that are easily interconnected with basic set of transfer/replication services Scriptable/customizable transfer/replication services for the most demanding

Alberto Pace, CERN, IT Department For more information: www.cern.ch/lcg www.eu-egee.org www.infn.it www.eu-egi.org/ Alberto Pace, CERN, IT Department 28