Download presentation
Presentation is loading. Please wait.
Published byArlene White Modified over 9 years ago
1
Pushing the Limits of Database clusters Jamie Shiers / CERN Werner Schueler / Intel
2
*Other trademarks and brands are the property of their respective owners 2 Agenda Trend to Intel Clusters … Introduction to CERN & Data Volumes Current 9iRAC / IA32 status Performance / Scalability / Reliability Future tests & timeline Plans for Oracle tests on IA64 Oracle9i RAC Performance Oracle9i RAC on Itanium
3
*Other trademarks and brands are the property of their respective owners 3 “Scale Up” by Scaling Out InfoWorld – January 31, 2002 "It will be several years before the big machine dies, but inevitably the big machine will die.“ — Larry Ellison Top 10 Clustering for performance Source: tpc.org
4
*Other trademarks and brands are the property of their respective owners 4 Proprietary Solutions Lagging Source: IDC 8/01 Worldwide Operating Environment Installed Base Server/Host Environments 2000-2005 0 4 8 12 16 200020012002200320042005 Years Units(M) Windows Servers Linux Servers Proprietary UNIX Servers
5
*Other trademarks and brands are the property of their respective owners 5 CERN Large Hadron Collider
6
*Other trademarks and brands are the property of their respective owners 6 The Large Hadron Collider (LHC)
7
*Other trademarks and brands are the property of their respective owners 7 Inside The 27km Tunnel…
8
*Other trademarks and brands are the property of their respective owners 8 ATLAS Detector System for LHC Detector is the size of a 6-floor building!
9
*Other trademarks and brands are the property of their respective owners 9 LHC: A Multi-Petabyte Problem! Long Term Tape Storage Estimates LEPExperiments COMPASS LHCExperiments 0 2'000 4'000 6'000 8'000 10'000 12'000 14'000 199519961997199819992000200120022003200420052006 Year TeraBytes
10
*Other trademarks and brands are the property of their respective owners 10 level 1 - special hardware 100 MHz (1000 TB/sec) level 2 - embedded processors level 3 - PCs 75 KHz (75 GB/sec) 5 KHz (5 GB/sec) 100 Hz (100 MB/sec) DB
11
*Other trademarks and brands are the property of their respective owners 11 LHC Data Volumes Data CategoryAnnual Total RAW1-3PB10-30PB Event Summary Data - ESD 100-500TB 1-5PB Analysis Object Data - AOD 10TB100TB TAG 1TB 10TB Total per experiment ~4PB~40PB Grand totals (10 years)~40PB~160PB Data CategoryAnnual Total RAW1-3PB10-30PB Event Summary Data - ESD 100-500TB 1-5PB Analysis Object Data - AOD 10TB100TB TAG 1TB 10TB Total per experiment ~4PB~40PB Grand totals (10 years)~40PB~160PB
12
*Other trademarks and brands are the property of their respective owners 12 LHC Summary Multi-national research lab near Geneva Building new accelerator: Large Hadron Collider Will generate fantastic amounts of data: 1PB/second! How can 9iRAC help?
13
*Other trademarks and brands are the property of their respective owners 13 LHC Computing Policy Commodity solutions where-ever possible Extensive use of Grid technologies Intel / Linux for processing nodes –Farms of many K nodes: 200K in today’s terms –IA32 today moving to IA64 prior to LHC startup 9iRAC claims to extend commodity solutions to the database market Does it live up to the promise? DB needs: ~100PB total; few GB/s / PB; many thousand concurrent processes; distributed access (world-wide)
14
*Other trademarks and brands are the property of their respective owners 14 History and experience Oracle Parallel Server since V7 –“Marketing clusters” – source Larry Ellison, OOW SFO 2001 OPS in production at CERN since 1996 –Mainly for high-availability Tests of 9iRAC started Autumn 2001 –Servers: 9 dual Pentium® III Xeon Processor based servers, 512MB –Storage: single node as above –Suse 7.2, Oracle 9.0.1 Currently working with 9iR2 –Servers: 10 nodes as above –Storage: now 3TB via 2 Intel-based disk-servers
15
*Other trademarks and brands are the property of their respective owners 15 CERN Computer Centre Today… inside has
16
*Other trademarks and brands are the property of their respective owners 16 Benefits of 9iRAC Scalability –Supports VLDBs using commodity h/w –Intel/Linux server nodes (target ~100TB / cluster) Manageability Small number of RAC manageable Small number of RAC manageable Tens / hundreds single instances a nightmare Better Resource Utilization –Shared disk architecture avoids hot-spots and idle / overworked nodes –Shared cache improves performance for frequently accessed read-only data
17
*Other trademarks and brands are the property of their respective owners 17 9iRAC benefits ¥ € $ Cost –N x dual processors typically much much cheaper than single large multi-processor ¥ € $ Cost –Fewer DBAs ¥ € $ Cost –No need to oversize system for peak loads
18
*Other trademarks and brands are the property of their respective owners 18 Tests on Linux Initial goals: Test that it works with commodity H/W + Linux Test that it works with commodity H/W + Linux Understand the configuration issues Understand the configuration issues –Check how it scales –Number of nodes –Network interconnect –CPU used for the cache coherency –Identify bottlenecks Commodity? Server + interconnect ok Server + interconnect ok – Storage outstanding question !!
19
*Other trademarks and brands are the property of their respective owners 19 Conventional Oracle Cluster Disks Database servers Clients (interactive, batch) e.g. Fibre channel based solution
20
*Other trademarks and brands are the property of their respective owners 20 Commodity Storage? Critical issue for CERN –Massive amount of data –Extremely tight budget constraints Long term (LHC: 2007) –network attached disks based on iSCSI? Short/Medium term: cost effective disk servers –€7.5K for 1.5TB mirrored at > 60MB/s)
21
*Other trademarks and brands are the property of their respective owners 21 Commodity Oracle Cluster? Disks Database servers Clients (interactive, batch) 3 interconnects, e.g. GbitE, possibly different protocols General purpose network Intra-cluster communications I/O network
22
*Other trademarks and brands are the property of their respective owners 22 Test & Deployment Goals Short-term (summer 2002): –Continue tests on multi-node 9iRAC up to ~3-5TB –Based on realistic data model & access patterns –Understand in-house, then test in Valbonne Medium-term (Q1 2003): –Production 9iRAC with up to 25TB of data –Modest I/O rate; primarily read-only data Long-term (LHC production phase): –Multiple multi-hundred TB RACs –Distributed in World-wide Grid
23
*Other trademarks and brands are the property of their respective owners 23 9iRAC Direction Strong & visible commitment from Oracle –Repeated message at OracleWorld –New features in 9iR2 –e.g. cluster file system for Windows and Linux Scalability depends to a certain extent on application –Our read-mostly data should be an excellent fit! Multi-TB tests with “professional” storage –HP / COMPAQ centre in Valbonne, France Target: 100TB per 9iRAC
24
*Other trademarks and brands are the property of their respective owners 24 Why 100TB? Possible today –BT Enormous Proof of Concept: 37TB in 1999 –CERN ODBMS deployment: 3TB per node Mainstream long before LHC –Winter 2000 VLDB survey: 100TB circa 2005 How does this match LHC need for 100PB? Analysis data: 100TB ok for ~10 years Analysis data: 100TB ok for ~10 years One 10 node 9iRAC per experiment One 10 node 9iRAC per experiment Intermediate: 100TB ~1 year’s data – ~40 10 node 9iRACs RAW data: 100TB = 1 month’s data –400 10node 9iRACs to handle all RAW data –10 RACs / year, 10 years, 4 experiments
25
*Other trademarks and brands are the property of their respective owners 25 LHC Data Volumes Revisited Data CategoryAnnual Total RAW1-3PB10-30PB Event Summary Data - ESD 100-500TB 1-5PB Analysis Object Data - AOD 10TB100TB TAG 1TB 10TB Total per experiment ~4PB~40PB Grand totals (15 years)~16PB~250PB Data CategoryAnnual Total RAW1-3PB10-30PB Event Summary Data - ESD 100-500TB 1-5PB Analysis Object Data - AOD 10TB100TB TAG 1TB 10TB Total per experiment ~4PB~40PB Grand totals (15 years)~16PB~250PB
26
*Other trademarks and brands are the property of their respective owners 26 RAW & ESD: >> 100TB RAW: –Access pattern: sequential –Access frequency: ~once per year –Use time partitioning + (offline tablespaces?) –100TB = 10 day time window –Current data (1 RAC) historic data (2 nd RAC) ESD: –Expect RAC scalability to continue to increase –VLDB prediction for 2020: 1000,000,000 TB (YB)
27
*Other trademarks and brands are the property of their respective owners 27 RAWRAW ESDESD AODAOD TAG random seq. 1PB/yr (1PB/s prior to reduction!) 100TB/yr 10TB/yr 1TB/yr Data Users Tier0 Tier1
28
*Other trademarks and brands are the property of their respective owners 28 Oracle Tests on IA64 64 bit computing essential for LHC –Addressability: VLMs, 64 bit filesystems, VLDBs –Accuracy: need 64 bit precision to track sub- atomic particles over tens of metres Migration IA32 IA64 prior to LHC startup
29
*Other trademarks and brands are the property of their respective owners 29 A solid history of Enterprise class processor development Intel’s technology innovations drive price/performance and scalability Time Performance RISC techniques for 2X i386™ performance Executes 2 instructions in parallel Multi-processor support Pentium ® processor Pentium ® II/III Xeon™ processors Pentium ® Pro processor Intel Xeon processor i486 ™ processor Intel ® Xeon™ processor MP Higher processing & data bandwidth for enterprise apps
30
*Other trademarks and brands are the property of their respective owners 30 Performance Via Technology Innovations Balanced system performance through higher bandwidth and throughput –Intel ® NetBurst™ microarchitecture –Integrated multi-level cache architecture Faster performance on business apps –Hyper-Threading Technology –up to 40% more efficient use of processor resources Processor Innovations for Increased Server Performance and Headroom
31
*Other trademarks and brands are the property of their respective owners 31 High Availability Back End ReliabilityAvailabilityReliabilityAvailability Mid-TierHigh-end General Purpose Scalability EPIC Architecture High Performance Front-end General Purpose BandwidthBandwidth ThroughputPerformance Matching Enterprise Requirements Itanium® Processor family Features System Requirements Enterprise Segments Features and flexibility to span the enterprise
32
*Other trademarks and brands are the property of their respective owners 32 Example: Calling circle OLTP model – Taken from a real world insurance example Best Performance… OLTP model –4 node x 4-way Pentium ® III Xeon ™ 700 MHz processor-based systems 128k TPM Over 90% scalability TPM Intel-based Solution Outperforms 32-way Sun Solution by More than 2x
33
*Other trademarks and brands are the property of their respective owners 33 Best Performance… TPC/C 8 nodes * 4 way Database Servers Pentium III Xeon 900Mhz 16 load generating Application Servers Pentium III 1Ghz
34
*Other trademarks and brands are the property of their respective owners 34 Best Performance … TPC/C
35
*Other trademarks and brands are the property of their respective owners 35 Best Performance… Price/Performance 9iRAC on RedHat on e.g. Dell 69% faster and 85% less expensive than Oracle on RISC solutions
36
*Other trademarks and brands are the property of their respective owners 36 Itanium ® Processor Family Performance Itanium ® Processor Processor Itanium ® 2 Processor Processor Madison* / Deerfield* Deerfield* Montecito*Montecito* 2001 2002 2003 Introduce architecture Introduce architecture Deliver competitive performance Deliver competitive performance Focused target segments Focused target segments Build-out architecture/ platform Build-out architecture/ platform Establish world-class performance Establish world-class performance Significantly increase deployment Significantly increase deployment Extend performance leadership Extend performance leadership Broaden target applications Broaden target applications Common hardware * Indicate Intel processor codenames. All products, dates and figures are preliminary, for planning purposes only, and subject to change without notice. Software scales across generations
37
*Other trademarks and brands are the property of their respective owners 37 Itanium ® 2 Processor On track for mid’02 releases from multiple OEMs and ISV Substantial performance leadership vs. RISC Delivering on performance promise 1.00 Itanium ® processor 800MHz 4MB L3 SPECint2000 Using Itanium ® 2 optimizations Source: Intel Corporation SPECfp2000Stream OLTP ERP Linpack 10K CAE CPU/Bandwidth Enterprise Technical Computing ~2.0 ~2.0 ~1.7 ~1.7 ~2.1 ~1.9
38
*Other trademarks and brands are the property of their respective owners 38 Deployment Strategy Scale Out with fail-over clusters on 1 to2-way servers Scale Up on 4 and 8-way servers, then Scale Out on fail-over clusters Scale Up on 8-way and above servers ExamplesInktomi* Apache* Web Server Microsoft Exchange* Server Oracle* 9iRAC SAS Enterprise Miner* Oracle 9i* Positioned To Scale Right Intel Relevance MP Versatile Server Solutions For Scaling Right
39
*Other trademarks and brands are the property of their respective owners 39 Inflection point coming Itanium2™ will have a 75%** price / performance lead over USIII at introduction in Q3’02 –Itanium2™ will outperform USIII by 40% –Itanium2™ will cost 20% less than USIII Oracle and Intel working to make 9i on Itanium a success –Joint performance goal of 100k TPM-C on a single 4- way Itanium2™ server –13 Intel engineers onsite and an additional 24 at Intel working to optimize 9i on Itanium2™ –Intel supplying Oracle large numbers of Itanium2™ development systems * McKinley is next generation Itanium ™ processor ** Estimated Q3’02 figures
40
*Other trademarks and brands are the property of their respective owners 40 Summary Existing Oracle technologies can be used to build 100TB databases Familiar data warehousing techniques can be used to handle much larger volumes of historic data Best Price and Performance through clusters vs. Risc 9iRAC makes this possible on commodity server platforms Standard High Volume servers offer great performance today and promise a safe investment for the future
41
Thank you Jamie.Shiers@cern.chWerner.Schueler@Intel.com
42
Thank You
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.