Pushing the Limits of Database clusters Jamie Shiers / CERN Werner Schueler / Intel.

Slides:



Advertisements
Similar presentations
NAS vs. SAN 10/2010 Palestinian Land Authority IT Department By Nahreen Ameen 1.
Advertisements

Transforming Business with Advanced Analytics: Introducing the New Intel® Xeon® Processor E7 v2 Family Seetha Rama Krishna Director, APAC HPC Solutions.
© Copyright 2012 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice. The future is mission.
Erhan Erdinç Pehlivan Computer Architecture Support for Database Applications.
Introduction to DBA.
Copyright © 2013, Oracle and/or its affiliates. All rights reserved.
2 June 2015 © Enterprise Storage Group, Inc. 1 The Case for File Server Consolidation using NAS Nancy Marrone Senior Analyst The Enterprise Storage Group,
Novell Server Linux vs. windows server 2008 By: Gabe Miller.
Symantec De-Duplication Solutions Complete Protection for your Information Driven Enterprise Richard Hobkirk Sr. Pre-Sales Consultant.
Copyright © 2006 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Technology Education Copyright © 2006 by The McGraw-Hill Companies,
Intel® 64-bit Platforms Platform Features. Agenda Introduction and Positioning of Intel® 64-bit Platforms Intel® 64-Bit Xeon™ Platforms Intel® Itanium®
® 1 Intel ® Server Platforms. ® *Other names and brands may be claimed as the property of others. 2 Building the enterprise business 8 years experience.
© 2009 Oracle Corporation. S : Slash Storage Costs with Oracle Automatic Storage Management Ara Vagharshakian ASM Product Manager – Oracle Product.
CERN IT Department CH-1211 Genève 23 Switzerland t Next generation of virtual infrastructure with Hyper-V Michal Kwiatek, Juraj Sucik, Rafal.
CERN/IT/DB Multi-PB Distributed Databases Jamie Shiers IT Division, DB Group, CERN, Geneva, Switzerland February 2001.
PROOF: the Parallel ROOT Facility Scheduling and Load-balancing ACAT 2007 Jan Iwaszkiewicz ¹ ² Gerardo Ganis ¹ Fons Rademakers ¹ ¹ CERN PH/SFT ² University.
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
Computer System Architectures Computer System Software
Bob Thome, Senior Director of Product Management, Oracle SIMPLIFYING YOUR HIGH AVAILABILITY DATABASE.
CERN-IT-DB Exabyte-Scale Data Management Using an Object-Relational Database: The LHC Project at CERN Jamie Shiers CERN, Switzerland
CERN - IT Department CH-1211 Genève 23 Switzerland t The High Performance Archiver for the LHC Experiments Manuel Gonzalez Berges CERN, Geneva.
Oracle Enterprise Grid Mark McGill Principal Sales Consultant Oracle EMEA Enterprise Technology Centre.
Database Edition for Sybase Sales Presentation. Market Drivers DBAs are facing immense time pressure in an environment with ever-increasing data Continuous.
Performance Concepts Mark A. Magumba. Introduction Research done on 1058 correspondents in 2006 found that 75% OF them would not return to a website that.
CERN - IT Department CH-1211 Genève 23 Switzerland t Tier0 database extensions and multi-core/64 bit studies Maria Girone, CERN IT-PSS LCG.
Managing Your Cross-Platform Data Keynote: The Power of 64 Speaker Name Here Date.
© Copyright 2012 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice. LogKV: Exploiting Key-Value.
High Performance Computing Processors Felix Noble Mirayma V. Rodriguez Agnes Velez Electric and Computer Engineer Department August 25, 2004.
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
Designing and Deploying a Scalable EPM Solution Ken Toole Platform Test Manager MS Project Microsoft.
Copyright © 2002, Intel Corporation. All rights reserved. *Other brands and names are the property of their respective owners
DATABASE MANAGEMENT SYSTEMS IN DATA INTENSIVE ENVIRONMENNTS Leon Guzenda Chief Technology Officer.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
AlphaServer UNIX Resource Consolidation.
Oracle RAC and Linux in the real enterprise October, 02 Mark Clark Director Merrill Lynch Europe PLC Global Database Technologies October, 02 Mark Clark.
Remote Direct Memory Access (RDMA) over IP PFLDNet 2003, Geneva Stephen Bailey, Sandburst Corp., Allyn Romanow, Cisco Systems,
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Achieving Scalability, Performance and Availability on Linux with Oracle 9iR2-RAC Grant McAlister Senior Database Engineer Amazon.com Paper
Copyright ©2003 Digitask Consultants Inc., All rights reserved Cluster Concepts Digitask Seminar November 29, 1999 Digitask Consultants, Inc.
CERN/IT/DB A Strawman Model for using Oracle for LHC Physics Data Jamie Shiers, IT-DB, CERN.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
US ATLAS Tier 1 Facility Rich Baker Brookhaven National Laboratory DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National Laboratory.
CERN-IT Oracle Database Physics Services Maria Girone, IT-DB 13 December 2004.
IDE disk servers at CERN Helge Meinhard / CERN-IT CERN OpenLab workshop 17 March 2003.
CERN Database Services for the LHC Computing Grid Maria Girone, CERN.
CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch.
Installation of Storage Foundation for Windows High Availability 5.1 SP2 1 Daniel Schnack Principle Technical Support Engineer.
Infrastructure for Data Warehouses. Basics Of Data Access Data Store Machine Memory Buffer Memory Cache Data Store Buffer Bus Structure.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
COMP381 by M. Hamdi 1 Clusters: Networks of WS/PC.
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 24 January 2005.
LCG LHC Computing Grid Project From the Web to the Grid 23 September 2003 Jamie Shiers, Database Group IT Division, CERN, Geneva, Switzerland
CERN/IT/DB DB US Visit Oracle Visit August 20 – [ plus related news ]
Maria Girone CERN - IT Tier0 plans and security and backup policy proposals Maria Girone, CERN IT-PSS.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 1 SIMULATION OF DAILY ACTIVITITIES AT REGIONAL CENTERS MONARC Collaboration Alexander Nazarenko.
Tackling I/O Issues 1 David Race 16 March 2010.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Background Computer System Architectures Computer System Software.
Thomas Baus Senior Sales Consultant Oracle/SAP Global Technology Center Mail: Phone:
Peter Idoine Managing Director Oracle New Zealand Limited.
IBM Systems and Technology Group © 2008 IBM Corporation Oracle Exadata Storage and the HP Oracle Database Machine Competitive Seller Podcast Mark Wulf.
Univa Grid Engine Makes Work Management Automatic and Efficient, Accelerates Deployment of Cloud Services with Power of Microsoft Azure MICROSOFT AZURE.
IT-DB Physics Services Planning for LHC start-up
Database Services at CERN Status Update
CLUSTER COMPUTING Presented By, Navaneeth.C.Mouly 1AY05IS037
Example of usage in Micron Italy (MIT)
Support for ”interactive batch”
SERVER INNOVATION ACCELERATES IT TRANSFORMATION
Microsoft Virtual Academy
Presentation transcript:

Pushing the Limits of Database clusters Jamie Shiers / CERN Werner Schueler / Intel

*Other trademarks and brands are the property of their respective owners 2 Agenda  Trend to Intel Clusters …  Introduction to CERN & Data Volumes  Current 9iRAC / IA32 status  Performance / Scalability / Reliability  Future tests & timeline  Plans for Oracle tests on IA64  Oracle9i RAC Performance  Oracle9i RAC on Itanium

*Other trademarks and brands are the property of their respective owners 3 “Scale Up” by Scaling Out InfoWorld – January 31, 2002 "It will be several years before the big machine dies, but inevitably the big machine will die.“ — Larry Ellison Top 10 Clustering for performance Source: tpc.org

*Other trademarks and brands are the property of their respective owners 4 Proprietary Solutions Lagging Source: IDC 8/01 Worldwide Operating Environment Installed Base Server/Host Environments Years Units(M) Windows Servers Linux Servers Proprietary UNIX Servers

*Other trademarks and brands are the property of their respective owners 5 CERN Large Hadron Collider

*Other trademarks and brands are the property of their respective owners 6 The Large Hadron Collider (LHC)

*Other trademarks and brands are the property of their respective owners 7 Inside The 27km Tunnel…

*Other trademarks and brands are the property of their respective owners 8 ATLAS Detector System for LHC Detector is the size of a 6-floor building!

*Other trademarks and brands are the property of their respective owners 9 LHC: A Multi-Petabyte Problem! Long Term Tape Storage Estimates LEPExperiments COMPASS LHCExperiments 0 2'000 4'000 6'000 8'000 10'000 12'000 14' Year TeraBytes

*Other trademarks and brands are the property of their respective owners 10 level 1 - special hardware 100 MHz (1000 TB/sec) level 2 - embedded processors level 3 - PCs 75 KHz (75 GB/sec) 5 KHz (5 GB/sec) 100 Hz (100 MB/sec) DB

*Other trademarks and brands are the property of their respective owners 11 LHC Data Volumes Data CategoryAnnual Total RAW1-3PB10-30PB Event Summary Data - ESD TB 1-5PB Analysis Object Data - AOD 10TB100TB TAG 1TB 10TB Total per experiment ~4PB~40PB Grand totals (10 years)~40PB~160PB Data CategoryAnnual Total RAW1-3PB10-30PB Event Summary Data - ESD TB 1-5PB Analysis Object Data - AOD 10TB100TB TAG 1TB 10TB Total per experiment ~4PB~40PB Grand totals (10 years)~40PB~160PB

*Other trademarks and brands are the property of their respective owners 12 LHC Summary  Multi-national research lab near Geneva  Building new accelerator: Large Hadron Collider  Will generate fantastic amounts of data: 1PB/second!  How can 9iRAC help?

*Other trademarks and brands are the property of their respective owners 13 LHC Computing Policy  Commodity solutions where-ever possible  Extensive use of Grid technologies  Intel / Linux for processing nodes –Farms of many K nodes: 200K in today’s terms –IA32 today moving to IA64 prior to LHC startup  9iRAC claims to extend commodity solutions to the database market  Does it live up to the promise?  DB needs: ~100PB total; few GB/s / PB; many thousand concurrent processes; distributed access (world-wide)

*Other trademarks and brands are the property of their respective owners 14 History and experience  Oracle Parallel Server since V7 –“Marketing clusters” – source Larry Ellison, OOW SFO 2001  OPS in production at CERN since 1996 –Mainly for high-availability  Tests of 9iRAC started Autumn 2001 –Servers: 9 dual Pentium® III Xeon Processor based servers, 512MB –Storage: single node as above –Suse 7.2, Oracle  Currently working with 9iR2 –Servers: 10 nodes as above –Storage: now 3TB via 2 Intel-based disk-servers

*Other trademarks and brands are the property of their respective owners 15 CERN Computer Centre Today… inside has

*Other trademarks and brands are the property of their respective owners 16 Benefits of 9iRAC  Scalability –Supports VLDBs using commodity h/w –Intel/Linux server nodes (target ~100TB / cluster)  Manageability Small number of RAC manageable Small number of RAC manageable  Tens / hundreds single instances a nightmare  Better Resource Utilization –Shared disk architecture avoids hot-spots and idle / overworked nodes –Shared cache improves performance for frequently accessed read-only data

*Other trademarks and brands are the property of their respective owners 17 9iRAC benefits ¥ € $ Cost –N x dual processors typically much much cheaper than single large multi-processor ¥ € $ Cost –Fewer DBAs ¥ € $ Cost –No need to oversize system for peak loads

*Other trademarks and brands are the property of their respective owners 18 Tests on Linux  Initial goals: Test that it works with commodity H/W + Linux Test that it works with commodity H/W + Linux Understand the configuration issues Understand the configuration issues –Check how it scales –Number of nodes –Network interconnect –CPU used for the cache coherency –Identify bottlenecks  Commodity? Server + interconnect ok Server + interconnect ok – Storage  outstanding question !!

*Other trademarks and brands are the property of their respective owners 19 Conventional Oracle Cluster Disks Database servers Clients (interactive, batch) e.g. Fibre channel based solution

*Other trademarks and brands are the property of their respective owners 20 Commodity Storage?  Critical issue for CERN –Massive amount of data –Extremely tight budget constraints  Long term (LHC: 2007) –network attached disks based on iSCSI?  Short/Medium term: cost effective disk servers –€7.5K for 1.5TB mirrored at > 60MB/s)

*Other trademarks and brands are the property of their respective owners 21 Commodity Oracle Cluster? Disks Database servers Clients (interactive, batch) 3 interconnects, e.g. GbitE, possibly different protocols General purpose network Intra-cluster communications I/O network

*Other trademarks and brands are the property of their respective owners 22 Test & Deployment Goals  Short-term (summer 2002): –Continue tests on multi-node 9iRAC up to ~3-5TB –Based on realistic data model & access patterns –Understand in-house, then test in Valbonne  Medium-term (Q1 2003): –Production 9iRAC with up to 25TB of data –Modest I/O rate; primarily read-only data  Long-term (LHC production phase): –Multiple multi-hundred TB RACs –Distributed in World-wide Grid

*Other trademarks and brands are the property of their respective owners 23 9iRAC Direction  Strong & visible commitment from Oracle –Repeated message at OracleWorld –New features in 9iR2 –e.g. cluster file system for Windows and Linux  Scalability depends to a certain extent on application –Our read-mostly data should be an excellent fit!  Multi-TB tests with “professional” storage –HP / COMPAQ centre in Valbonne, France  Target: 100TB per 9iRAC

*Other trademarks and brands are the property of their respective owners 24 Why 100TB?  Possible today –BT Enormous Proof of Concept: 37TB in 1999 –CERN ODBMS deployment: 3TB per node  Mainstream long before LHC –Winter 2000 VLDB survey: 100TB circa 2005  How does this match LHC need for 100PB? Analysis data: 100TB ok for ~10 years Analysis data: 100TB ok for ~10 years One 10 node 9iRAC per experiment One 10 node 9iRAC per experiment  Intermediate: 100TB ~1 year’s data – ~40 10 node 9iRACs  RAW data: 100TB = 1 month’s data –400 10node 9iRACs to handle all RAW data –10 RACs / year, 10 years, 4 experiments

*Other trademarks and brands are the property of their respective owners 25 LHC Data Volumes Revisited Data CategoryAnnual Total RAW1-3PB10-30PB Event Summary Data - ESD TB 1-5PB Analysis Object Data - AOD 10TB100TB TAG 1TB 10TB Total per experiment ~4PB~40PB Grand totals (15 years)~16PB~250PB Data CategoryAnnual Total RAW1-3PB10-30PB Event Summary Data - ESD TB 1-5PB Analysis Object Data - AOD 10TB100TB TAG 1TB 10TB Total per experiment ~4PB~40PB Grand totals (15 years)~16PB~250PB  

*Other trademarks and brands are the property of their respective owners 26 RAW & ESD: >> 100TB  RAW: –Access pattern: sequential –Access frequency: ~once per year –Use time partitioning + (offline tablespaces?) –100TB = 10 day time window –Current data (1 RAC) historic data (2 nd RAC)  ESD: –Expect RAC scalability to continue to increase –VLDB prediction for 2020: 1000,000,000 TB (YB)

*Other trademarks and brands are the property of their respective owners 27 RAWRAW ESDESD AODAOD TAG random seq. 1PB/yr (1PB/s prior to reduction!) 100TB/yr 10TB/yr 1TB/yr Data Users Tier0 Tier1

*Other trademarks and brands are the property of their respective owners 28 Oracle Tests on IA64  64 bit computing essential for LHC –Addressability: VLMs, 64 bit filesystems, VLDBs –Accuracy: need 64 bit precision to track sub- atomic particles over tens of metres  Migration IA32  IA64 prior to LHC startup

*Other trademarks and brands are the property of their respective owners 29 A solid history of Enterprise class processor development Intel’s technology innovations drive price/performance and scalability Time Performance RISC techniques for 2X i386™ performance Executes 2 instructions in parallel Multi-processor support Pentium ® processor Pentium ® II/III Xeon™ processors Pentium ® Pro processor Intel Xeon processor i486 ™ processor Intel ® Xeon™ processor MP Higher processing & data bandwidth for enterprise apps

*Other trademarks and brands are the property of their respective owners 30 Performance Via Technology Innovations  Balanced system performance through higher bandwidth and throughput –Intel ® NetBurst™ microarchitecture –Integrated multi-level cache architecture  Faster performance on business apps –Hyper-Threading Technology –up to 40% more efficient use of processor resources Processor Innovations for Increased Server Performance and Headroom

*Other trademarks and brands are the property of their respective owners 31 High Availability Back End ReliabilityAvailabilityReliabilityAvailability Mid-TierHigh-end General Purpose Scalability EPIC Architecture High Performance Front-end General Purpose BandwidthBandwidth ThroughputPerformance Matching Enterprise Requirements Itanium® Processor family Features System Requirements Enterprise Segments Features and flexibility to span the enterprise

*Other trademarks and brands are the property of their respective owners 32 Example: Calling circle OLTP model – Taken from a real world insurance example Best Performance… OLTP model –4 node x 4-way Pentium ® III Xeon ™ 700 MHz processor-based systems  128k TPM  Over 90% scalability TPM Intel-based Solution Outperforms 32-way Sun Solution by More than 2x

*Other trademarks and brands are the property of their respective owners 33 Best Performance… TPC/C  8 nodes * 4 way Database Servers Pentium III Xeon 900Mhz  16 load generating Application Servers Pentium III 1Ghz

*Other trademarks and brands are the property of their respective owners 34 Best Performance … TPC/C

*Other trademarks and brands are the property of their respective owners 35 Best Performance… Price/Performance  9iRAC on RedHat on e.g. Dell 69% faster and 85% less expensive than Oracle on RISC solutions

*Other trademarks and brands are the property of their respective owners 36 Itanium ® Processor Family Performance Itanium ® Processor Processor Itanium ® 2 Processor Processor Madison* / Deerfield* Deerfield* Montecito*Montecito* Introduce architecture Introduce architecture Deliver competitive performance Deliver competitive performance Focused target segments Focused target segments Build-out architecture/ platform Build-out architecture/ platform Establish world-class performance Establish world-class performance Significantly increase deployment Significantly increase deployment Extend performance leadership Extend performance leadership Broaden target applications Broaden target applications Common hardware * Indicate Intel processor codenames. All products, dates and figures are preliminary, for planning purposes only, and subject to change without notice. Software scales across generations

*Other trademarks and brands are the property of their respective owners 37 Itanium ® 2 Processor  On track for mid’02 releases from multiple OEMs and ISV  Substantial performance leadership vs. RISC Delivering on performance promise 1.00 Itanium ® processor 800MHz 4MB L3 SPECint2000 Using Itanium ® 2 optimizations Source: Intel Corporation SPECfp2000Stream OLTP ERP Linpack 10K CAE CPU/Bandwidth Enterprise Technical Computing ~2.0 ~2.0 ~1.7 ~1.7 ~2.1 ~1.9

*Other trademarks and brands are the property of their respective owners 38 Deployment Strategy Scale Out with fail-over clusters on 1 to2-way servers Scale Up on 4 and 8-way servers, then Scale Out on fail-over clusters Scale Up on 8-way and above servers ExamplesInktomi* Apache* Web Server Microsoft Exchange* Server Oracle* 9iRAC SAS Enterprise Miner* Oracle 9i* Positioned To Scale Right Intel Relevance MP Versatile Server Solutions For Scaling Right

*Other trademarks and brands are the property of their respective owners 39 Inflection point coming  Itanium2™ will have a 75%** price / performance lead over USIII at introduction in Q3’02 –Itanium2™ will outperform USIII by 40% –Itanium2™ will cost 20% less than USIII  Oracle and Intel working to make 9i on Itanium a success –Joint performance goal of 100k TPM-C on a single 4- way Itanium2™ server –13 Intel engineers onsite and an additional 24 at Intel working to optimize 9i on Itanium2™ –Intel supplying Oracle large numbers of Itanium2™ development systems * McKinley is next generation Itanium ™ processor ** Estimated Q3’02 figures

*Other trademarks and brands are the property of their respective owners 40 Summary  Existing Oracle technologies can be used to build 100TB databases  Familiar data warehousing techniques can be used to handle much larger volumes of historic data  Best Price and Performance through clusters vs. Risc  9iRAC makes this possible on commodity server platforms  Standard High Volume servers offer great performance today and promise a safe investment for the future

Thank you

Thank You