LCG LHC Computing Grid Project From the Web to the Grid 23 September 2003 Jamie Shiers, Database Group IT Division, CERN, Geneva, Switzerland

Slides:



Advertisements
Similar presentations
Particle physics – the computing challenge CERN Large Hadron Collider –2007 –the worlds most powerful particle accelerator –10 petabytes (10 million billion.
Advertisements

The LHC experiments AuthZ Interoperation requirements GGF16, Athens 16 February 2006 David Kelsey CCLRC/RAL, UK
Highest Energy e + e – Collider LEP at CERN GeV ~4km radius First e + e – Collider ADA in Frascati GeV ~1m radius e + e – Colliders.
Randall Sobie The ATLAS Experiment Randall Sobie Institute for Particle Physics University of Victoria Large Hadron Collider (LHC) at CERN Laboratory ATLAS.
N° 1 LCG EDG Data Management Catalogs in LCG James Casey LCG Fellow, IT-DB Group, CERN
The LHC Computing Grid Project Tomi Kauppi Timo Larjo.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 15 th April 2009 Visit of Spanish Royal Academy.
Site report: CERN Helge Meinhard (at) cern ch HEPiX fall SLAC.
Exploiting the Grid to Simulate and Design the LHCb Experiment K Harrison 1, N Brook 2, G Patrick 3, E van Herwijnen 4, on behalf of the LHCb Grid Group.
CERN/IT/DB Multi-PB Distributed Databases Jamie Shiers IT Division, DB Group, CERN, Geneva, Switzerland February 2001.
D. Duellmann, CERN Data Management at the LHC1 Data Management at CERN’s Large Hadron Collider (LHC) Dirk Düllmann CERN IT/DB, Switzerland
13 October 2014 Eric Grancher, head of database services, CERN IT Manuel Martin Marquez, data scientist, CERN openlab.
1. 2 CERN European Organization for Nuclear Research Founded in 1954 by 12 countries – Norway one of them Today: 20 member states, around 2500 staff –
CERN - European Laboratory for Particle Physics HEP Computer Farms Frédéric Hemmer CERN Information Technology Division Physics Data processing Group.
CERN-IT-DB Exabyte-Scale Data Management Using an Object-Relational Database: The LHC Project at CERN Jamie Shiers CERN, Switzerland
ITEP participation in the EGEE project NEC’2005, Varna, Bulgaria Ivan Korolko (ITEP Moscow)
Particle Physics and the Grid Randall Sobie Institute of Particle Physics University of Victoria Motivation Computing challenge LHC Grid Canadian requirements.
1 Chasing the Higgs boson with a worldwide distributed trigger system Sander Klous NIKHEF VENI proposal 2006.
Miguel Branco CERN/University of Southampton Enabling provenance on large-scale e-Science applications.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Copyright © 2000 OPNET Technologies, Inc. Title – 1 Distributed Trigger System for the LHC experiments Krzysztof Korcyl ATLAS experiment laboratory H.
May 2004Sverre Jarp1 Preparing the computing solutions for the Large Hadron Collider (LHC) at CERN Sverre Jarp, openlab CTO IT Department, CERN.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
Storage Tank in Data Grid Shin, SangYong(syshin, #6468) IBM Grid Computing August 23, 2003.
LHC Computing Plans Scale of the challenge Computing model Resource estimates Financial implications Plans in Canada.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Quick Introduction to NorduGrid Oxana Smirnova 4 th Nordic LHC Workshop November 23, 2001, Stockholm.
… where the Web was born 11 November 2003 Wolfgang von Rüden, IT Division Leader CERN openlab Workshop on TCO Introduction.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
GridPP Building a UK Computing Grid for Particle Physics Professor Steve Lloyd, Queen Mary, University of London Chair of the GridPP Collaboration Board.
CERN What Happens at CERN? "In the matter of physics, the first lessons should contain nothing but what is experimental and interesting to see. A pretty.
…building the next IT revolution From Web to Grid…
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
CERN-IT Oracle Database Physics Services Maria Girone, IT-DB 13 December 2004.
ATLAS is a general-purpose particle physics experiment which will study topics including the origin of mass, the processes that allowed an excess of matter.
Cluster Configuration Update Including LSF Status Thorsten Kleinwort for CERN IT/PDP-IS HEPiX I/2001 LAL Orsay Tuesday, December 08, 2015.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch.
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 24 January 2005.
Computing for LHC Physics 7th March 2014 International Women's Day - CERN- GOOGLE Networking Event Maria Alandes Pradillo CERN IT Department.
LHC Computing, CERN, & Federated Identities
Data Processing and the LHC Computing Grid (LCG) Jamie Shiers Database Group, IT Division CERN, Geneva, Switzerland
David Foster LCG Project 12-March-02 Fabric Automation The Challenge of LHC Scale Fabrics LHC Computing Grid Workshop David Foster 12 th March 2002.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
CNAF Database Service Barbara Martelli CNAF-INFN Elisabetta Vilucchi CNAF-INFN Simone Dalla Fina INFN-Padua.
05 Novembre years of research in physics European Organization for Nuclear Research.
International Collaboration for Data Preservation and Long Term Analysis in High Energy Physics RECODE - Final Workshop - January.
ATLAS Distributed Analysis DISTRIBUTED ANALYSIS JOBS WITH THE ATLAS PRODUCTION SYSTEM S. González D. Liko
The ATLAS detector … … is composed of cylindrical layers: Tracking detector: Pixel, SCT, TRT (Solenoid magnetic field) Calorimeter: Liquid Argon, Tile.
Grid technologies for large-scale projects N. S. Astakhov, A. S. Baginyan, S. D. Belov, A. G. Dolbilov, A. O. Golunov, I. N. Gorbunov, N. I. Gromova, I.
CrossGrid Workshop, Kraków, 5 – 6 Nov-2001 Distributed Data Analysis in HEP Piotr MALECKI Institute of Nuclear Physics Kawiory 26A, Kraków, Poland.
European Organization for Nuclear Research
The EDG Testbed Deployment Details
Introduction to CERN F. Hahn / CERN PH-DT1 10. May 2007.
Grid site as a tool for data processing and data analysis
The LHC Computing Grid Visit of Mtro. Enrique Agüera Ibañez
European Organization for Nuclear Research
IT-DB Physics Services Planning for LHC start-up
CERN The world’s largest Particle Physics Research Center in Geneva
CERN, the LHC and the Grid
Welcome to the CMS Virtual Visit
What is CERN?.
Computing activities at Victoria
Presentation transcript:

LCG LHC Computing Grid Project From the Web to the Grid 23 September 2003 Jamie Shiers, Database Group IT Division, CERN, Geneva, Switzerland

LCG Overview  Very brief overview of CERN  Use of Oracle at CERN – a partnership lasting two decades  From the Large Electron Positron collider (LEP) to the Large Hadron Collider (LHC)  The LHC Computing Grid (LCG) and Oracle’s role

LCG The European Organisation for Nuclear Research The European Laboratory for Particle Physics  Fundamental research in particle physics  Designs, builds & operates large accelerators  Financed by 20 European countries (member states) + others (US, Canada, Russia, India, ….)  1MSF budget - operation + new accelerators  2000 staff users (researchers) from all over the world  LHC (starts ~2007) experiment: 2000 physicists, 150 universities, apparatus costing ~€300M, computing ~€250M to setup, ~€60M/year to run  10+ year lifetime

LCG airport Computer Centre Geneva  27km 

LCG LEP : 1989 – 2000 (RIP)  27km ring with counter-circulating electrons-positrons  Oracle Database selected to help with LEP construction  Originally ran on PDP-11, later VAX, IBM, Sun, now Linux  Oracle now used during LEP dismantling phase  Data on LEP components must be kept forever  Oracle is now used across entire spectrum of lab’s activities  Several Sun-based clusters (8i OPS, 9i RAC)  Many stand-alone Linux-based systems  Both database and increasingly Application Server

LCG High-lights of the LEP Era  LEP Computing started with the MAINFRAME  Initially IBM running VM/CMS, large VAXcluster, also Cray  In 1989, first proposal of what led to Web was made  Somewhat heretical at the time: strongly based on e.g. use of Internet protocols, whereas official line was OSI…  Goal was to simplify task of sharing information amongst physicists: by definition distributed across the world  Technology convergence: explosion of Internet – explosion of Web  In early 1990s, first steps towards fully distributed computing with farms of RISC processors running Unix  The “SHIFT” project, winner of ComputerWorld Honors Award

LCG The Large Hadron Collider (LHC) A New World-Class Machine in the LEP Tunnel (First proposed in 1979!)

LCG

The LHC machine Two counter- circulating proton beams Collision energy TeV 27 Km of magnets with a field of 8.4 Tesla Super-fluid Helium cooled to 1.9°K The world’s largest superconducting structure

LCG The ATLAS detector – the size of a 6 floor building!

LCG The Atlas Cavern – January 03

LCG Data Acquisition Multi-level trigger Filters out background Reduces data volume Record data 24 hours a day, 7 days a week Equivalent to writing a CD every 2 seconds Level 3 – Giant PC Cluster 160 Hz 160 Hz (320 MB/sec) Data Recording & Offline Analysis Level 2 - Embedded Processors 40 MHz interaction rate equivalent to 2 PetaBytes/sec Level 1 - Special Hardware Atlas detector

LCG Oracle for Physics Data  Work on LHC Computing started ~1992 (some would say earlier…)  Numerous projects kicked off 1994/5 to look at handling multi-PB of data; move from Fortran to OO (C++) etc.  Led to production solutions from ~1997  Always said that ‘disruptive technology’, like Web, would have to be taken into account  In 2002, major project started to move 350TB of data out of ODBMS solution; >100MB/s for 24 hour periods  Now ~2TB of physics data stored in Oracle on Linux servers  A few % of total data volume; expected to double in 2004

LCG Linux for Physics Computing  First steps with Linux started ~1993  Port of Physics Application Software to Linux on PCs  1996: proposal to setup Windows-based batch farms for Physics Data Processing  Overtaken by developments in Linux:  Windows a poor match for batch environment  Linux essential trivial to port to from Solaris, HP/UX etc  Convergence of technologies: PC h/w offers unbeatable price- performance; Linux becomes robust  ~All Physics Computing at CERN now based on Linux / Intel  Strategic platform for the LHC

LCG The Grid The Solution to LHC Computing? LHC Computing Project = LHC Computing Grid (LCG)

LCG LHC Computing Grid (LCG)  Global requirements: handle processing and data handling needs of 4 main LHC collaborations  Total of PB of data per year (>20 million CDs); lifetime 10+ years  Analysis will require equivalent of 70,000 of today’s fastest PCs  LCG project established to meet these unprecedented requirements  Builds on work of European DataGrid (EDG) and Virtual Data Toolkit (US)  Physicists access world-wide distributed data & resources as if local  System determines where job runs, based on resources required/available  Initial partners include sites in CH, F, D, I, UK, US, Japan, Taiwan & Russia

LCG Centres taking part in the initial LCG service ( ) around the world  around the clock

LCG LCG and Oracle  Current thinking is that bulk data will be streamed to files  RDBMS backend also being studied for ‘analysis data’  File catalog (10 9 files) and file-level metadata will be stored in Oracle in a Grid-aware catalog  In longer term, event level metadata may also be stored in the database, leading to much larger data volumes  A few PB, assuming total data volume of PB  Current storage management system – CASTOR at CERN – also uses a database to manage the naming / location of files  bulk data stored in tape silos and faulted in to huge disk caches

LCG Storage Element Replica Location Services Replica Manager Local Replica Catalog Replica Metadata Catalog Storage Element Files have replicas stored at many Grid sites on Storage Elements. Each file has a unique GUID. Locations corresponding to the GUID are kept in the Replica Location Service. Users may assign aliases to the GUIDs. These are kept in the Replica Metadata Catalog. The Replica Manager provides atomicity for file operations, assuring consistency of SE and catalog contents.

LCG Today’s Deployment at CERN rlsatlasrlsalicerlscmsrlslhcbrlsdteamrlscert02rlscert01 lxshare071d rlstest lxshare169d lxshare183d lxshare069d  Oracle Application Server hosting Grid Middleware per VO  Shared Oracle Database for LHC Experiments  Based on ‘standard parts’ out of CERN stores  Disk server (1TB mirrored disk); Farm node (dual processor)

LCG Future Deployment  Currently studying 9iRAC on supported h/w configurations  Expect Grid infrastructure to move to AS cluster + RAC in Q1/  Expect CASTOR databases to move to RAC also in 2004  May also move few TB of event-level metadata (COMPASS) to a single RAC  All based on Linux / Intel

LCG Summary  During past decade, have moved from era of Web to Grid  Rise of Internet Computing, move from mainframes to RISC to farms of dual-processor Intel boxes running Linux  Use of Oracle has expanded from small, dedicated service for LEP construction to all areas of the lab’s work, including handling Physics Data  Both Oracle DB and AS, including for Grid infrastructure  The Grid viewed as ‘disruptive technology’ in that it will change the way we think about computing, much like the Web