Presentation is loading. Please wait.

Presentation is loading. Please wait.

Pushing the Limits of Database clusters Jamie Shiers / CERN Werner Schueler / Intel.

Similar presentations


Presentation on theme: "Pushing the Limits of Database clusters Jamie Shiers / CERN Werner Schueler / Intel."— Presentation transcript:

1 Pushing the Limits of Database clusters Jamie Shiers / CERN Werner Schueler / Intel

2 *Other trademarks and brands are the property of their respective owners 2 Agenda  Trend to Intel Clusters …  Introduction to CERN & Data Volumes  Current 9iRAC / IA32 status  Performance / Scalability / Reliability  Future tests & timeline  Plans for Oracle tests on IA64  Oracle9i RAC Performance  Oracle9i RAC on Itanium

3 *Other trademarks and brands are the property of their respective owners 3 “Scale Up” by Scaling Out InfoWorld – January 31, 2002 "It will be several years before the big machine dies, but inevitably the big machine will die.“ — Larry Ellison Top 10 Clustering for performance Source: tpc.org

4 *Other trademarks and brands are the property of their respective owners 4 Proprietary Solutions Lagging Source: IDC 8/01 Worldwide Operating Environment Installed Base Server/Host Environments 2000-2005 0 4 8 12 16 200020012002200320042005 Years Units(M) Windows Servers Linux Servers Proprietary UNIX Servers

5 *Other trademarks and brands are the property of their respective owners 5 CERN Large Hadron Collider

6 *Other trademarks and brands are the property of their respective owners 6 The Large Hadron Collider (LHC)

7 *Other trademarks and brands are the property of their respective owners 7 Inside The 27km Tunnel…

8 *Other trademarks and brands are the property of their respective owners 8 ATLAS Detector System for LHC Detector is the size of a 6-floor building!

9 *Other trademarks and brands are the property of their respective owners 9 LHC: A Multi-Petabyte Problem! Long Term Tape Storage Estimates LEPExperiments COMPASS LHCExperiments 0 2'000 4'000 6'000 8'000 10'000 12'000 14'000 199519961997199819992000200120022003200420052006 Year TeraBytes

10 *Other trademarks and brands are the property of their respective owners 10 level 1 - special hardware 100 MHz (1000 TB/sec) level 2 - embedded processors level 3 - PCs 75 KHz (75 GB/sec) 5 KHz (5 GB/sec) 100 Hz (100 MB/sec) DB

11 *Other trademarks and brands are the property of their respective owners 11 LHC Data Volumes Data CategoryAnnual Total RAW1-3PB10-30PB Event Summary Data - ESD 100-500TB 1-5PB Analysis Object Data - AOD 10TB100TB TAG 1TB 10TB Total per experiment ~4PB~40PB Grand totals (10 years)~40PB~160PB Data CategoryAnnual Total RAW1-3PB10-30PB Event Summary Data - ESD 100-500TB 1-5PB Analysis Object Data - AOD 10TB100TB TAG 1TB 10TB Total per experiment ~4PB~40PB Grand totals (10 years)~40PB~160PB

12 *Other trademarks and brands are the property of their respective owners 12 LHC Summary  Multi-national research lab near Geneva  Building new accelerator: Large Hadron Collider  Will generate fantastic amounts of data: 1PB/second!  How can 9iRAC help?

13 *Other trademarks and brands are the property of their respective owners 13 LHC Computing Policy  Commodity solutions where-ever possible  Extensive use of Grid technologies  Intel / Linux for processing nodes –Farms of many K nodes: 200K in today’s terms –IA32 today moving to IA64 prior to LHC startup  9iRAC claims to extend commodity solutions to the database market  Does it live up to the promise?  DB needs: ~100PB total; few GB/s / PB; many thousand concurrent processes; distributed access (world-wide)

14 *Other trademarks and brands are the property of their respective owners 14 History and experience  Oracle Parallel Server since V7 –“Marketing clusters” – source Larry Ellison, OOW SFO 2001  OPS in production at CERN since 1996 –Mainly for high-availability  Tests of 9iRAC started Autumn 2001 –Servers: 9 dual Pentium® III Xeon Processor based servers, 512MB –Storage: single node as above –Suse 7.2, Oracle 9.0.1  Currently working with 9iR2 –Servers: 10 nodes as above –Storage: now 3TB via 2 Intel-based disk-servers

15 *Other trademarks and brands are the property of their respective owners 15 CERN Computer Centre Today… inside has

16 *Other trademarks and brands are the property of their respective owners 16 Benefits of 9iRAC  Scalability –Supports VLDBs using commodity h/w –Intel/Linux server nodes (target ~100TB / cluster)  Manageability Small number of RAC manageable Small number of RAC manageable  Tens / hundreds single instances a nightmare  Better Resource Utilization –Shared disk architecture avoids hot-spots and idle / overworked nodes –Shared cache improves performance for frequently accessed read-only data

17 *Other trademarks and brands are the property of their respective owners 17 9iRAC benefits ¥ € $ Cost –N x dual processors typically much much cheaper than single large multi-processor ¥ € $ Cost –Fewer DBAs ¥ € $ Cost –No need to oversize system for peak loads

18 *Other trademarks and brands are the property of their respective owners 18 Tests on Linux  Initial goals: Test that it works with commodity H/W + Linux Test that it works with commodity H/W + Linux Understand the configuration issues Understand the configuration issues –Check how it scales –Number of nodes –Network interconnect –CPU used for the cache coherency –Identify bottlenecks  Commodity? Server + interconnect ok Server + interconnect ok – Storage  outstanding question !!

19 *Other trademarks and brands are the property of their respective owners 19 Conventional Oracle Cluster Disks Database servers Clients (interactive, batch) e.g. Fibre channel based solution

20 *Other trademarks and brands are the property of their respective owners 20 Commodity Storage?  Critical issue for CERN –Massive amount of data –Extremely tight budget constraints  Long term (LHC: 2007) –network attached disks based on iSCSI?  Short/Medium term: cost effective disk servers –€7.5K for 1.5TB mirrored at > 60MB/s)

21 *Other trademarks and brands are the property of their respective owners 21 Commodity Oracle Cluster? Disks Database servers Clients (interactive, batch) 3 interconnects, e.g. GbitE, possibly different protocols General purpose network Intra-cluster communications I/O network

22 *Other trademarks and brands are the property of their respective owners 22 Test & Deployment Goals  Short-term (summer 2002): –Continue tests on multi-node 9iRAC up to ~3-5TB –Based on realistic data model & access patterns –Understand in-house, then test in Valbonne  Medium-term (Q1 2003): –Production 9iRAC with up to 25TB of data –Modest I/O rate; primarily read-only data  Long-term (LHC production phase): –Multiple multi-hundred TB RACs –Distributed in World-wide Grid

23 *Other trademarks and brands are the property of their respective owners 23 9iRAC Direction  Strong & visible commitment from Oracle –Repeated message at OracleWorld –New features in 9iR2 –e.g. cluster file system for Windows and Linux  Scalability depends to a certain extent on application –Our read-mostly data should be an excellent fit!  Multi-TB tests with “professional” storage –HP / COMPAQ centre in Valbonne, France  Target: 100TB per 9iRAC

24 *Other trademarks and brands are the property of their respective owners 24 Why 100TB?  Possible today –BT Enormous Proof of Concept: 37TB in 1999 –CERN ODBMS deployment: 3TB per node  Mainstream long before LHC –Winter 2000 VLDB survey: 100TB circa 2005  How does this match LHC need for 100PB? Analysis data: 100TB ok for ~10 years Analysis data: 100TB ok for ~10 years One 10 node 9iRAC per experiment One 10 node 9iRAC per experiment  Intermediate: 100TB ~1 year’s data – ~40 10 node 9iRACs  RAW data: 100TB = 1 month’s data –400 10node 9iRACs to handle all RAW data –10 RACs / year, 10 years, 4 experiments

25 *Other trademarks and brands are the property of their respective owners 25 LHC Data Volumes Revisited Data CategoryAnnual Total RAW1-3PB10-30PB Event Summary Data - ESD 100-500TB 1-5PB Analysis Object Data - AOD 10TB100TB TAG 1TB 10TB Total per experiment ~4PB~40PB Grand totals (15 years)~16PB~250PB Data CategoryAnnual Total RAW1-3PB10-30PB Event Summary Data - ESD 100-500TB 1-5PB Analysis Object Data - AOD 10TB100TB TAG 1TB 10TB Total per experiment ~4PB~40PB Grand totals (15 years)~16PB~250PB  

26 *Other trademarks and brands are the property of their respective owners 26 RAW & ESD: >> 100TB  RAW: –Access pattern: sequential –Access frequency: ~once per year –Use time partitioning + (offline tablespaces?) –100TB = 10 day time window –Current data (1 RAC) historic data (2 nd RAC)  ESD: –Expect RAC scalability to continue to increase –VLDB prediction for 2020: 1000,000,000 TB (YB)

27 *Other trademarks and brands are the property of their respective owners 27 RAWRAW ESDESD AODAOD TAG random seq. 1PB/yr (1PB/s prior to reduction!) 100TB/yr 10TB/yr 1TB/yr Data Users Tier0 Tier1

28 *Other trademarks and brands are the property of their respective owners 28 Oracle Tests on IA64  64 bit computing essential for LHC –Addressability: VLMs, 64 bit filesystems, VLDBs –Accuracy: need 64 bit precision to track sub- atomic particles over tens of metres  Migration IA32  IA64 prior to LHC startup

29 *Other trademarks and brands are the property of their respective owners 29 A solid history of Enterprise class processor development Intel’s technology innovations drive price/performance and scalability Time Performance RISC techniques for 2X i386™ performance Executes 2 instructions in parallel Multi-processor support Pentium ® processor Pentium ® II/III Xeon™ processors Pentium ® Pro processor Intel Xeon processor i486 ™ processor Intel ® Xeon™ processor MP Higher processing & data bandwidth for enterprise apps

30 *Other trademarks and brands are the property of their respective owners 30 Performance Via Technology Innovations  Balanced system performance through higher bandwidth and throughput –Intel ® NetBurst™ microarchitecture –Integrated multi-level cache architecture  Faster performance on business apps –Hyper-Threading Technology –up to 40% more efficient use of processor resources Processor Innovations for Increased Server Performance and Headroom

31 *Other trademarks and brands are the property of their respective owners 31 High Availability Back End ReliabilityAvailabilityReliabilityAvailability Mid-TierHigh-end General Purpose Scalability EPIC Architecture High Performance Front-end General Purpose BandwidthBandwidth ThroughputPerformance Matching Enterprise Requirements Itanium® Processor family Features System Requirements Enterprise Segments Features and flexibility to span the enterprise

32 *Other trademarks and brands are the property of their respective owners 32 Example: Calling circle OLTP model – Taken from a real world insurance example Best Performance… OLTP model –4 node x 4-way Pentium ® III Xeon ™ 700 MHz processor-based systems  128k TPM  Over 90% scalability TPM Intel-based Solution Outperforms 32-way Sun Solution by More than 2x

33 *Other trademarks and brands are the property of their respective owners 33 Best Performance… TPC/C  8 nodes * 4 way Database Servers Pentium III Xeon 900Mhz  16 load generating Application Servers Pentium III 1Ghz

34 *Other trademarks and brands are the property of their respective owners 34 Best Performance … TPC/C

35 *Other trademarks and brands are the property of their respective owners 35 Best Performance… Price/Performance  9iRAC on RedHat on e.g. Dell 69% faster and 85% less expensive than Oracle on RISC solutions

36 *Other trademarks and brands are the property of their respective owners 36 Itanium ® Processor Family Performance Itanium ® Processor Processor Itanium ® 2 Processor Processor Madison* / Deerfield* Deerfield* Montecito*Montecito* 2001 2002 2003 Introduce architecture Introduce architecture Deliver competitive performance Deliver competitive performance Focused target segments Focused target segments Build-out architecture/ platform Build-out architecture/ platform Establish world-class performance Establish world-class performance Significantly increase deployment Significantly increase deployment Extend performance leadership Extend performance leadership Broaden target applications Broaden target applications Common hardware * Indicate Intel processor codenames. All products, dates and figures are preliminary, for planning purposes only, and subject to change without notice. Software scales across generations

37 *Other trademarks and brands are the property of their respective owners 37 Itanium ® 2 Processor  On track for mid’02 releases from multiple OEMs and ISV  Substantial performance leadership vs. RISC Delivering on performance promise 1.00 Itanium ® processor 800MHz 4MB L3 SPECint2000 Using Itanium ® 2 optimizations Source: Intel Corporation SPECfp2000Stream OLTP ERP Linpack 10K CAE CPU/Bandwidth Enterprise Technical Computing ~2.0 ~2.0 ~1.7 ~1.7 ~2.1 ~1.9

38 *Other trademarks and brands are the property of their respective owners 38 Deployment Strategy Scale Out with fail-over clusters on 1 to2-way servers Scale Up on 4 and 8-way servers, then Scale Out on fail-over clusters Scale Up on 8-way and above servers ExamplesInktomi* Apache* Web Server Microsoft Exchange* Server Oracle* 9iRAC SAS Enterprise Miner* Oracle 9i* Positioned To Scale Right Intel Relevance MP Versatile Server Solutions For Scaling Right

39 *Other trademarks and brands are the property of their respective owners 39 Inflection point coming  Itanium2™ will have a 75%** price / performance lead over USIII at introduction in Q3’02 –Itanium2™ will outperform USIII by 40% –Itanium2™ will cost 20% less than USIII  Oracle and Intel working to make 9i on Itanium a success –Joint performance goal of 100k TPM-C on a single 4- way Itanium2™ server –13 Intel engineers onsite and an additional 24 at Intel working to optimize 9i on Itanium2™ –Intel supplying Oracle large numbers of Itanium2™ development systems * McKinley is next generation Itanium ™ processor ** Estimated Q3’02 figures

40 *Other trademarks and brands are the property of their respective owners 40 Summary  Existing Oracle technologies can be used to build 100TB databases  Familiar data warehousing techniques can be used to handle much larger volumes of historic data  Best Price and Performance through clusters vs. Risc  9iRAC makes this possible on commodity server platforms  Standard High Volume servers offer great performance today and promise a safe investment for the future

41 Thank you Jamie.Shiers@cern.chWerner.Schueler@Intel.com

42 Thank You


Download ppt "Pushing the Limits of Database clusters Jamie Shiers / CERN Werner Schueler / Intel."

Similar presentations


Ads by Google