5 December 2013, Eric Grancher CERN, head of database services.

Slides:



Advertisements
Similar presentations
Welcome to CERN Accelerating Science and Innovation 2 nd March 2015 – Bidders Conference – DO-29161/EN.
Advertisements

CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000.
Grids: Why and How (you might use them) J. Templon, NIKHEF VLV T Workshop NIKHEF 06 October 2003.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
12. March 2003Bernd Panzer-Steindel, CERN/IT1 LCG Fabric status
Replication Technologies at WLCG Lorena Lobato Pardavila CERN IT Department – DB Group JINR/CERN Grid and Management Information Systems, Dubna (Russia)
Welcome to CERN Research Technology Training Collaborating.
SAP Student Interest Group
Experiences and Challenges running CERN's High-Capacity Tape Archive 14/4/2015 CHEP 2015, Okinawa2 Germán Cancio, Vladimír Bahyl
Accelerating Science and Innovation. Science for Peace CERN was founded in 1954 as a Science for Peace Initiative by 12 European States Member States:
EDUCATION & RESEARCH Training CERN and the Oracle Value to Research February 3rd, Eva Dafonte Perez, Deputy.
13 October 2014 Eric Grancher, head of database services, CERN IT Manuel Martin Marquez, data scientist, CERN openlab.
CERN - IT Department CH-1211 Genève 23 Switzerland t The High Performance Archiver for the LHC Experiments Manuel Gonzalez Berges CERN, Geneva.
From Open Science to (Open) Innovation Markus Nordberg, Marzio Nessi (CERN) About CERN, Open Physics, Innovation and IdeaSquare.
José M. Hernández CIEMAT Grid Computing in the Experiment at LHC Jornada de usuarios de Infraestructuras Grid January 2012, CIEMAT, Madrid.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
Rackspace Analyst Event Tim Bell
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 25 th April 2012.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration.
Ian Bird Trigger, Online, Offline Computing Workshop CERN, 5 th September 2014.
6 May 2014 CERN openlab IT Challenges workshop, Kacper Szkudlarek, CERN Manuel.
The LHC Computing Grid – February 2008 The Challenges of LHC Computing Dr Ian Bird LCG Project Leader 6 th October 2009 Telecom 2009 Youth Forum.
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
CERN – IT Department CH-1211 Genève 23 Switzerland t Working with Large Data Sets Tim Smith CERN/IT Open Access and Research Data Session.
Company small business cloud solution Client UNIVERSITY OF BEDFORDSHIRE.
WelcomeWelcome CSEM – CERN Day 23 rd May 2013 CSEM – CERN Day 23 rd May 2013 to Accelerating Science and Innovation to Accelerating Science and Innovation.
CERN as a World Laboratory: From a European Organization to a global facility CERN openlab Board of Sponsors July 2, 2010 Rüdiger Voss CERN Physics Department.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
CERN openlab Overview CERN openlab Introduction Alberto Di Meglio.
Brocade Flow Optimizer CERN openlab
Status report of the KLOE offline G. Venanzoni – LNF LNF Scientific Committee Frascati, 9 November 2004.
Computing for LHC Physics 7th March 2014 International Women's Day - CERN- GOOGLE Networking Event Maria Alandes Pradillo CERN IT Department.
LHC Computing, CERN, & Federated Identities
The LHC Computing Grid – February 2008 The Challenges of LHC Computing Frédéric Hemmer IT Department 26 th January 2010 Visit of Michael Dell 1 Frédéric.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Enterprise Manager at CERN Andrei Dumitru IT Department / Database Services / openlab.
CERN IT Department CH-1211 Genève 23 Switzerland t The Tape Service at CERN Vladimír Bahyl IT-FIO-TSI June 2009.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 1 st March 2011 Visit of Dr Manuel Eduardo Baldeón.
CERN openlab Overview CERN openlab Summer Students 2015 Fons Rademakers.
The Mission of CERN  Push back  Push back the frontiers of knowledge E.g. the secrets of the Big Bang …what was the matter like within the first moments.
25-September-2005 Manjit Dosanjh Welcome to CERN International Workshop on African Research & Education Networking September ITU, UNU and CERN.
The Worldwide LHC Computing Grid Frédéric Hemmer IT Department Head Visit of INTEL ISEF CERN Special Award Winners 2012 Thursday, 21 st June 2012.
Jiri Chudoba for the Pierre Auger Collaboration Institute of Physics of the CAS and CESNET.
Germany and CERN / June 2009Germany and CERN | May Welcome - Willkommen CERN: to CERN: Accelerating Science and Innovation Professor Wolfgang A.
WLCG after 1 year with data: Prospects for the future Ian Bird; WLCG Project Leader openlab BoS meeting CERN4 th May 2011.
Big Data for Big Discoveries How the LHC looks for Needles by Burning Haystacks Alberto Di Meglio CERN openlab Head DOI: /zenodo.45449, CC-BY-SA,
Tape archive challenges when approaching Exabyte-scale CHEP 2010, Taipei G. Cancio, V. Bahyl, G. Lo Re, S. Murray, E. Cano, G. Lee, V. Kotlyar CERN IT-DSS.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Building Preservation Environments with Data Grid Technology Reagan W. Moore Presenter: Praveen Namburi.
IT-DSS Alberto Pace2 ? Detecting particles (experiments) Accelerating particle beams Large-scale computing (Analysis) Discovery We are here The mission.
Product Lifecycle Management with the CERN EDMS David Widegren CERN, TS/CSE 8 Nov 2005EDMS:
WLCG – Status and Plans Ian Bird WLCG Project Leader openlab Board of Sponsors CERN, 23 rd April 2010.
Grid technologies for large-scale projects N. S. Astakhov, A. S. Baginyan, S. D. Belov, A. G. Dolbilov, A. O. Golunov, I. N. Gorbunov, N. I. Gromova, I.
Computing infrastructures for the LHC: current status and challenges of the High Luminosity LHC future Worldwide LHC Computing Grid (WLCG): Distributed.
Report on availability of the JINR networking, computing and information infrastructure for real data taking and processing in LHC experiments Ivanov V.V.
Bidders’ Conference IT-4130/EN Replacement of Personnel Lifts (up to 2 t Capacity) I. Rühl / EN-HE.
5-minutes tour of CERN (based on official CERN slides) 5-minutes tour of CERN (based on official CERN slides) Christian Joram / CERN EIROfrum Topical Workshop.
European Organization for Nuclear Research
The 5 minutes tour of CERN The 5 minutes race of CERN
Experiences and Outlook Data Preservation and Long Term Analysis
The LHC Computing Challenge
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
ALIGNMENT RULE PROCEDURE
Ákos Frohner EGEE'08 September 2008
LHC Data Analysis using a worldwide computing grid
ALIGNMENT RULE PROCEDURE
The LHC Computing Grid Visit of Professor Andreas Demetriou
CERN: from fundamental sciences to daily applications
Presentation transcript:

5 December 2013, Eric Grancher CERN, head of database services

Outlook 3 CERN History of using Oracle Current usage Collaboration Why using Oracle in our research environment?

CERN European Organization for Nuclear Research Founded in 1954 Research: Seeking and finding answers to questions about the Universe Technology, International collaboration, Education Twenty Member States Austria, Belgium, Bulgaria, Czech Republic, Denmark, Finland, France, Germany, Greece, Italy, Hungary, Netherlands, Norway, Poland, Portugal, Slovakia, Spain, Sweden, Switzerland, United Kingdom Seven Observer States European Commission, USA, Russian Federation, India, Japan, Turkey, UNESCO People ~2400 Staff, ~900 Students, post-docs and undergraduates, ~9000 Users, ~2000 Contractors Associate Member States Israel, Serbia Candidate State Romania 4

A European Laboratory with Global reach 5

LHC 17 miles (27km) long tunnel Thousands of superconducting magnets Coldest place in the Universe: 1.9 degrees kelvin Ultra vacuum: 10x emptier than on the Moon 600 million collisions per second / analysis is like finding a needle in 20 million haystacks The largest particle accelerator & detectors

Events at LHC J. Knobloch/CERN: European Grid Initiative – EGI Luminosity : cm -2 s MHz – every 25 ns 20 events overlaying

Trigger & Data Acquisition J. Knobloch/CERN: European Grid Initiative – EGI

Data Recording J. Knobloch/CERN: European Grid Initiative – EGI

World’s largest computing grid - WLCG 1 PB raw data per second before filtering / >20 PB of new data annually 68,889 physical CPUs / 305,935 logical CPUS 157 computer centres around the world 10

Computing and storage needs 11 Data volume 25 PB per year (in files) > 5.2 * rows in an Oracle table (IOT, in one of the databases) Computing and storage capacity, world-wide distributed > 150 sites (grid computing) > CPU cores > 269 PB disk capacity > 210 PB tape capacity Distributed analysis with costs spread in the different sites (« LHC Computing Grid »)

LS1: no stop for the computing ! Activity on 1 January 2013 Running Jobs: Transfer rate: ~14 GiB/s CERN Tape Archive Reaching 100 PB ! Grid CPU Workloads (HEPSpec06/Month) ~ CPU cores in use at any point in time! : Data to be reprocessed Simulation for higher Energy Analysis continuing …to be repacked on new media Credit: F. Hemmer

Oracle at CERN, 1982 accelerator control 13

Oracle at CERN, version Credit: N. Segura Chinchilla

Accelerator logging 15 Credit: C. Roderick

Accelerator logging 16 Credit: C. Roderick

Accelerator logging 17 50TB/year, rate to increase to 100 – 150 TB in 2014 (Quench Protection System) Credit: C. Roderick

Administrative systems 18 AIS has standardized on Oracle as database and uses the Oracle database as interface between the tools Java EE and Apex, deployment with Weblogic Oracle E-Business HR, managed as Oracle Managed Cloud Services (“OnDemand”)

Engineering applications An integrated PLM platform based on commercial tools. Simplified web interfaces for precise tasks 3D CAD CATIA Design data management Manufacturing follow-up Installation follow-up Maintenance management Data publishing Workflow actions 19 Credit: D. Widegren

Design data management Design baseline with full configuration management - Workflows, versioning rules and access control based on project dependent contexts. - Fully web-based and distributed approval processes. 20 Credit: D. Widegren

Manufacturing follow-up Follow-up of each manufactured component - Manufacturing & test data captured at manufacturing sites. - Predefined manufacturing workflows for each equipment type. 21 Credit: D. Widegren

Installation follow-up Detailed logging of Installation & Commissioning tasks - Over jobs logged – allows detailed progress reporting - Resolution of non-conformities with distributed approval processes. 22 Credit: D. Widegren

CERN in numbers Document & Drawings (incl. CAD): ~1, documents & drawings ~7,000 new documents & drawings created per month Components: ~1,300,000 registered individually followed equipment ~3,000,000 equipment interventions/jobs logged ~ 15,000 equipment interventions/jobs logged per month 23 Credit: D. Widegren

CASTOR and Oracle, tapes 24 Home made mass storage system, relies on Oracle databases for name server, request handling and staging 4 libraries, SL x4 = 40K slots (4500 free) Occupancy: 65PB worth of data Drives: 20 T10KB legacy drives; 40 T10KC drives (to be replaced by T10KD's) Credit: German Cancio Melia

CERN Disk/Tape Storage storage-day.ch CASTOR Archive in Numbers 25 Data: ~90PB of data on tape; 250M files Up to 4.5 PB new data per month Over 10GB/s (R+W) peaks Credit: German Cancio Melia

CERN Disk/Tape Storage storage-day.ch Writing to tape is well managed and understood Detector data can’t wait for us! High aggregate transfer rates are not a problem System designed to split the write stream onto several tape drives However, low per-drive performance “Tape Mark” overhead per tape file up to 2s CASTOR has 3 tape files per payload file No significant overhead 14y ago on 20MB/s drives…! Today with avg files of 350 MB: < 50% efficiency! Tape Writing 26 Changed CASTOR tape layer to write “logical” (or “buffered”) tape marks between files (via CERN-prototyped Linux kernel extension) Now, achieving near-native drive speed by writing data in effective chunks of ~32GB –Significant savings in tape drive purchases Credit: German Cancio Melia

CERN Disk/Tape Storage storage-day.ch Data Write Rate Increase 27 Boosted per-drive performance: from ~55 MB/s to ~150 MB/s Credit: German Cancio Melia

CERN Disk/Tape Storage storage-day.ch Archive Verification Data in the archive cannot just be written and forgotten about. Q: can you retrieve my file? A: let me check… err, sorry, we lost it. Proactive and regular verification of archive data required Ensure cartridges can be mounted Check data can be read+verified against metadata (checksum/size, …) Do not wait until media migration to detect problems Several commercial solutions available on the market Difficult integration with our application Not always check your metadata Recently implemented and deployed a background scanning engine: Read back all newly filled tapes Scan the whole archive over time, starting with least recent accessed tapes 28

CERN Disk/Tape Storage storage-day.ch Verification: first round completed! Up to drives (~10%) for 90% efficiency Turnaround time: ~2.6(!!) ~1.26GB/s Data loss: ~ 65GB lost over 69 tapes 29

CERN Disk/Tape Storage storage-day.ch 1970’s – GB’s 1990’s – TB’s Today – PB’s Mass media migration or “repacking” required for –Higher-density media generations, and / or –Higher-density tape drives (enterprise media rewriting) –Liberating space - tape library slots Media itself can last for 30 years, but not the infrastructure! Media Repacking 30

CERN Disk/Tape Storage storage-day.ch Mass media migration or “repacking” required for –Higher-density media generations, and / or –Higher-density tape drives (enterprise media rewriting) –Liberating space - tape library slots Media itself can last for 30 years, but not the infrastructure! Repack exercise is proportional to the total size of archive - and not to the fresh or active data –Scale infrastructure accordingly! Next Repack run ~ 2014 – 2016 –~100PB to migrate from over 50’000 cartridges Repack data rates exceed LHC data rates… –Over 3 GB/s sustained –Must be compatible with “regular” data taking (60PB/year in ) …. but we need to share the tape drives – which become the bottleneck –“Opportunistic” repack activity handling needed Media Repacking 31

CERN Disk/Tape Storage storage-day.ch Reading is The Real Challenge CASTOR was designed as a “classic” file-based HSM. If user file is not on disk -> recall it from tape ASAP Works OK for “controlled” activities such as RECO and Data Export But Analysis is unpredictable access to data: up to 20K(!) tape (re)mounts / day for reading 1-2 files each Every mount is wasted drive/robotics resources (~2 min for mounting / unmounting)! Mounting and unmounting is the highest risk operation for tapes, robotics and drives. Mechanical (robotics) failure can affect access to a large amount of media. Tape technology evolution moves against HSM: Bigger tapes -> more files -> more mounts per tape -> reduced media lifetime Increased transfer speed, but little increases for mounting/positioning – unsuitable for random access Conclusion: Split Analysis from other Use Cases into a separate system 32

Experiment online systems 33 Experiments rely on a SCADA system for their control Up to 150,000 changes / second stored in Oracle databases

Experiment offline systems 34 Geometry DB Conditions DB Credit: Vakho Tsulaia

Oracle at CERN 35 From accelerator control to accelerator logging, administration, engineering systems, access control, laboratory infrastructure (cabling, network configuration, etc.), mass storage system, experiment online systems, experiment offline systems, Etc.

36 Credit: M. Piorkowski

openlab (1/3) 37 Public-private partnership between CERN and leading ICT companies, currently in fourth phase (started in 2003) Its mission is to accelerate the development of cutting-edge solutions to be used by the worldwide LHC community Innovative ideas aligned between CERN and the partners, for products “you make it, we break it”

openlab (2/3) 38 Many successes: RAC on Linux x86 (9.2 PoC and 10.1 production with ASM), Additional required functionality (IEEE numbers, OCCI, instant client, etc.), PVSS and RAC scalability, Monitoring with Grid Control, Streams world wide distribution, Active DG, GoldenGate, Analytics for accelerator, experiment and IT, Etc. Regular feedback with joint selection of topics, some of the projects are common with more than one partner

openlab (3/3) 39 Publications (web, paper) and presentations of results, visitors Maaike Limper, best poster award at The International Conference on Computing in High Energy and Nuclear Physics 2013

Oracle in our research environment 40 Even if computing is critical for HEP, it is not the goal, there is a lot to do using solutions from commercial vendors which are industry supported and scalable Oracle has provided solutions along the years We have worked with Oracle to improve the tools to our (and others’) needs with success Good for staff to work on industry standards for their future career

Conclusion 41 Not every day you build a 30+ years collaboration A long way since 1982, now very wide usage with applications, tape and database Oracle has proven to be reliable partner who cares and supports research Provide feedback and ideas for enhancements Helps focus on our core challenges A collaboration which works!