CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/i t Experience with NetApp at CERN IT/DB Giacomo Tenaglia on behalf of Eric Grancher Ruben.

Slides:



Advertisements
Similar presentations
Archive Task Team (ATT) Disk Storage Stuart Doescher, USGS (Ken Gacke) WGISS-18 September 2004 Beijing, China.
Advertisements

Data Storage Solutions Module 1.2. Data Storage Solutions Upon completion of this module, you will be able to: List the common storage media and solutions.
NAS vs. SAN 10/2010 Palestinian Land Authority IT Department By Nahreen Ameen 1.
High Availability Group 08: Võ Đức Vĩnh Nguyễn Quang Vũ
CERN IT Department CH-1211 Geneva 23 Switzerland t Marcin Blaszczyk, IT-DB Atlas standby database tests February.
Scale-out Central Store. Conventional Storage Verses Scale Out Clustered Storage Conventional Storage Scale Out Clustered Storage Faster……………………………………………….
Keith Burns Microsoft UK Mission Critical Database.
IBM® Spectrum Storage Virtualize™ V V7000 Unified in a nutshell
CERN IT Department CH-1211 Geneva 23 Switzerland t Sequential data access with Oracle and Hadoop: a performance comparison Zbigniew Baranowski.
CERN IT Department CH-1211 Geneva 23 Switzerland t CERN IT Department CH-1211 Geneva 23 Switzerland t
1 RAL Status and Plans Carmine Cioffi Database Administrator and Developer 3D Workshop, CERN, November 2009.
BNL Oracle database services status and future plans Carlos Fernando Gamboa RACF Facility Brookhaven National Laboratory, US Distributed Database Operations.
CERN IT Department CH-1211 Genève 23 Switzerland t Data Protection with Oracle Data Guard Jacek Wojcieszuk, CERN/IT-DM Distributed Database.
High Availability & Oracle RAC 18 Aug 2005 John Sheaffer Platform Solution Specialist
CERN IT Department CH-1211 Genève 23 Switzerland t Next generation of virtual infrastructure with Hyper-V Michal Kwiatek, Juraj Sucik, Rafal.
Experience and Lessons learnt from running High Availability Databases on Network Attached Storage Ruben Gaspar Manuel Guijarro et al IT/DES.
Object-based Storage Long Liu Outline Why do we need object based storage? What is object based storage? How to take advantage of it? What's.
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
Nexenta Proprietary Global Leader in Software Defined Storage Nexenta Technical Sales Professional (NTSP) COURSE CONTENT.
ASGC 1 ASGC Site Status 3D CERN. ASGC 2 Outlines Current activity Hardware and software specifications Configuration issues and experience.
Experience in running relational databases on clustered storage CERN, IT Department CHEP 2015, Okinawa, Japan 13/04/2015.
Luca Canali, CERN Dawid Wojcik, CERN
Ray Pasetes, Andy Romero CSS Storage Services CD/CSS Central Services Activities Coordination Meeting 9/19/2006.
CERN - IT Department CH-1211 Genève 23 Switzerland t Tier0 database extensions and multi-core/64 bit studies Maria Girone, CERN IT-PSS LCG.
CERN - IT Department CH-1211 Genève 23 Switzerland t Experience and Lessons learnt from running High Availability Databases on Network Attached.
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
1 U.S. Department of the Interior U.S. Geological Survey Contractor for the USGS at the EROS Data Center EDC CR1 Storage Architecture August 2003 Ken Gacke.
CERN IT Department CH-1211 Geneva 23 Switzerland t IT/DB Tests and evolution SSD as flash cache.
Fast Crash Recovery in RAMCloud. Motivation The role of DRAM has been increasing – Facebook used 150TB of DRAM For 200TB of disk storage However, there.
CERN - IT Department CH-1211 Genève 23 Switzerland t Oracle Real Application Clusters (RAC) Techniques for implementing & running robust.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
CERN IT Department CH-1211 Genève 23 Switzerland t Possible Service Upgrade Jacek Wojcieszuk, CERN/IT-DM Distributed Database Operations.
ONStor Pantera 3110 ONStor NAS. Copyright 2008 · ONStor Confidential Pantera 3110 – An Integrated Channel only NAS  Integrated standalone NAS system.
CERN Database Services for the LHC Computing Grid Maria Girone, CERN.
Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS Castor incident (and follow up) Alberto Pace.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
Control System Virtualization for the LHCb Online System ICALEPCS – San Francisco Enrico Bonaccorsi, (CERN)
CERN IT Department CH-1211 Genève 23 Switzerland t Storage Overview and IT-DM Lessons Learned Luca Canali, IT-DM DM Group Meeting
Backup Exec System Recovery. 2 Outline Introduction Challenges Solution Implementation Results Recommendations Q & A.
CERN IT Department CH-1211 Geneva 23 Switzerland t WLCG Operation Coordination Luca Canali (for IT-DB) Oracle Upgrades.
CERN IT Department CH-1211 Geneva 23 Switzerland t Eva Dafonte Perez IT-DB Database Replication, Backup and Archiving.
Operating Systems & Information Services CERN IT Department CH-1211 Geneva 23 Switzerland t OIS Drupal at CERN Juraj Sucik Jarosław Polok.
ORACLE & VLDB Nilo Segura IT/DB - CERN. VLDB The real world is in the Tb range (British Telecom - 80Tb using Sun+Oracle) Data consolidated from different.
Maria Girone CERN - IT Tier0 plans and security and backup policy proposals Maria Girone, CERN IT-PSS.
CERN IT Department CH-1211 Genève 23 Switzerland t Next generation of virtual infrastructure with Hyper-V Juraj Sucik, Michal Kwiatek, Rafal.
Scalable Oracle 10g for the Physics Database Services Luca Canali, CERN IT January, 2006.
LHC Logging Cluster Nilo Segura IT/DB. Agenda ● Hardware Components ● Software Components ● Transparent Application Failover ● Service definition.
Database CNAF Barbara Martelli Rome, April 4 st 2006.
PIC port d’informació científica Luis Diaz (PIC) ‏ Databases services at PIC: review and plans.
CERN IT Department CH-1211 Geneva 23 Switzerland t Distributed Database Operations Workshop November 17 th, 2010 Przemyslaw Radowiecki CERN.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
VVols with Adaptive Flash and InfoSight Analytics 1 Manchester Virtualisation User Group Rich Fenton (Nimble North Senior Systems Engineer)
CERN IT Department CH-1211 Genève 23 Switzerland t Load testing & benchmarks on Oracle RAC Romain Basset – IT PSS DP.
CERN IT Department CH-1211 Geneva 23 Switzerland t OIS Operating Systems & Information Services CERN IT Department CH-1211 Geneva 23 Switzerland.
Journey to the HyperConverged Agile Infrastructure
About ProLion CEO, Robert Graf Headquarter in Austria
RHEV Platform at LHCb Red Hat at CERN 17-18/1/17
Database Services Katarzyna Dziedziniewicz-Wojcik On behalf of IT-DB.
Backup & Recovery of Physics Databases
iSCSI Storage Area Network
The demonstration of Lustre in EAST data system
Database Services at CERN Status Update
Scalable Database Services for Physics: Oracle 10g RAC on Linux
The INFN Tier-1 Storage Implementation
Oracle Storage Performance Studies
ASM-based storage to scale out the Database Services for Physics
Scalable Database Services for Physics: Oracle 10g RAC on Linux
About ProLion CEO, Robert Graf Headquarter in Austria
DataOptimizer Transparent File Tiering for NetApp Storage Robert Graf
Presentation transcript:

CERN IT Department CH-1211 Geneva 23 Switzerland t Experience with NetApp at CERN IT/DB Giacomo Tenaglia on behalf of Eric Grancher Ruben Gaspar Aparicio

CERN IT Department CH-1211 Geneva 23 Switzerland t Outline NAS-based usage at CERN Key features Future plans Experience with NetApp at CERN IT/DB - 2

Storage for Oracle at CERN 1982: Oracle at CERN, PDP-11, mainframe, VAX VMS, Solaris SPARC 32 and : Solaris SPARC with OPS, then RAC 2000: Linux x86 on single node, DAS 2005: Linux x86_64 / RAC / SAN –Experiment and part of WLCG on SAN until : Linux x86_64 / RAC / NFS (IBM/NetApp) 2012: all production primary Oracle databases (*) on NFS (*) apart from ALICE and LHCb online Experience with NetApp at CERN IT/DB - 3

Network topology All 10Gb/s Ethernet Same network for storage and cluster interconnect filer1filer2filer3filer4 serverB serverA serverCserverE serverD Ethernet switch Private 1 Ethernet switch Private 2 Internal HA pair interconnect Private network, both CRS and storage “public network” Ethernet switch Public

Domains: space/filers Total size (TB)Used for backup (TB)# of Filers des-nas shosts2044 gen3974 rac10596 rac11596 castor15418 acc2818 db disk10002 TOTAL Experience with NetApp at CERN IT/DB - 5

Typical setup

Impact of storage architecture on Oracle stability at CERN Experience with NetApp at CERN IT/DB - 7

Key features Flash cache RaidDP Snapshots Compression Experience with NetApp at CERN IT/DB - 8

Flash cache Help to increase random IOPs on disks –Very good for OLTP-like workload Don’t get wiped when servers reboot For databases –Decide what volumes to cache: fas3240>priority on fas3240>priority set volume volname cache=[reuse|keep] 512 GB modules 1 per controller Experience with NetApp at CERN IT/DB - 9

IOPs and Flash cache Experience with NetApp at CERN IT/DB - 10

IOPs and Flash cache Experience with NetApp at CERN IT/DB - 11

Key features Flash cache RaidDP Snapshots Compression Experience with NetApp at CERN IT/DB - 12

Disk and redundancy (1/2) Disks are larger and larger –speed stay ~constant → issue with performance –bit error rate stay constant ( to ), increasing issue with availability With x as the size and α the “bit error rate” Experience with NetApp at CERN IT/DB - 13

Disks, redundancy comparison (2/2) 1 TB SATA desktop Bit error rate 10^-14 RAID 17.68E-02 RAID 5 (n+1)3.29E E E-01 ~RAID 6 (n+2)1.60E E E-13 ~triple mirror8.00E-16 1TB SATA enterprise Bit error rate 10^-15 RAID 17.96E-03 RAID 5 (n+1)3.92E E E-01 ~RAID 6 (n+2)1.60E E E-15 ~triple mirror8.00E GB FC Bit error rate 10^-16 RAID 14.00E-04 RAID 5 (n+1)2.00E E E-02 ~RAID 6 (n+2)7.20E E E-17 ~triple mirror3.60E TB SATA enterprise Bit error rate 10^-15 RAID 17.68E-02 RAID 5 (n+1)3.29E E E-01 ~RAID 6 (n+2)1.60E E E-14 ~triple mirror8E-17 Experience with NetApp at CERN IT/DB - 14 Data loss probability for different disk types and groups

Key features Flash cache RaidDP Snapshots Compression Experience with NetApp at CERN IT/DB - 15

Snapshots Experience with NetApp at CERN IT/DB - 16 T0: take snapshot 1

Snapshots Experience with NetApp at CERN IT/DB - 17 T0: take snapshot 1 T1: file changed

Snapshots Experience with NetApp at CERN IT/DB - 18 T0: take snapshot 1 T1: file changed T2: take snapshot 2

Snapshots for backups With data growth, restoring databases in reasonable amount of time is impossible using “traditional” restore/backup techniques 100TB, 10GbE, 4 tape drives Tape drive restore performance ~120MB/s Restore ~ 58 hours (but it can be much longer) Experience with NetApp at CERN IT/DB - 19

Snapshots and Real Application Testing Capture insert… PL/SQL update … delete … Original Clone Upgrade Replay insert… PL/SQL update … delete … Experience with NetApp at CERN IT/DB - 20

Snapshots and Real Application Testing Capture insert… PL/SQL update … delete … Original Clone Upgrade Replay insert… PL/SQL update … delete … SnapRestore® Replay insert… PL/SQL update … delete … Replay insert… PL/SQL update … delete … Experience with NetApp at CERN IT/DB - 20

Key features Flash cache RaidDP Snapshots Compression Experience with NetApp at CERN IT/DB - 21

NetApp compression factor Experience with NetApp at CERN IT/DB - 22

Compression: backup on disk RMAN File backup 1x tape copy + Disk buffer Raw: ~1700 TiB (576 3TB disks) Usable: 1000 TiB (to hold ~2PiB uncompressed data) Experience with NetApp at CERN IT/DB - 23

Future: OnTap Cluster Mode Non-disruptive upgrades/operations: the immortal cluster Interesting new features –Internal DNS load balancing –Export policies: fine-grained access for NFS exports –Encryption and compression at storage level –NFS 4.1 implementation, parallel NFS Scale-out architecture: up to 24 (512 theoretical) Seamless data moves for capacity, performance rebalancing or hardware replacement Experience with NetApp at CERN IT/DB - 24

Architecture view – Ontap cluster mode Experience with NetApp at CERN IT/DB - 25

Possible implementation Experience with NetApp at CERN IT/DB - 26

Logical components Experience with NetApp at CERN IT/DB - 27

pNFS NFS 4.1 standard (client caching, Kerberos, ACL) Coming with Ontap 8.1RC2 Not natively supported by Oracle yet In RHEL 6.2 Control protocol: provides synchronization among data and metadata server pNFS between client and MDS, get where information is store Storage access protocols: file-based, block-based and object- based pNFS Storage access protocols Experience with NetApp at CERN IT/DB - 28

CERN IT Department CH-1211 Geneva 23 Switzerland t Summary Good reliability –Six years of operations with minimal downtime Good flexibility –Same setup for different uses/workloads Scales to our needs Experience with NetApp at CERN IT/DB - 29

CERN IT Department CH-1211 Geneva 23 Switzerland t Q&A Thanks! Experience with NetApp at CERN IT/DB - 30