PAE SuperCluster Peformance

Slides:



Advertisements
Similar presentations
Fluffy’s Safe Right? If you want to limit a user’s functionality, don’t make them an administrator.
Advertisements

Exadata Distinctives Brown Bag New features for tuning Oracle database applications.
Concurrent programming: From theory to practice Concurrent Algorithms 2014 Vasileios Trigonakis Georgios Chatzopoulos.
Oracle Exadata for SAP.
Copyright © 2012, Oracle and/or its affiliates. All rights reserved. 1.
Page 1 Dorado 400 Series Server Club Page 2 First member of the Dorado family based on the Next Generation architecture Employs Intel 64 Xeon Dual.
Automatic Storage Management The New Best Practice Steve Adams Ixora Rich Long Oracle Corporation Session id:
Copyright © 2012, Oracle and/or its affiliates. All rights reserved. 1 Oracle Active Data Guard Performance Joseph Meeks Director, Product Management Oracle.
SEDA: An Architecture for Well- Conditioned, Scalable Internet Services Matt Welsh, David Culler, and Eric Brewer Computer Science Division University.
SEDA: An Architecture for Well-Conditioned, Scalable Internet Services Matt Welsh, David Culler, and Eric Brewer Computer Science Division University of.
Everest: scaling down peak loads through I/O off-loading D. Narayanan, A. Donnelly, E. Thereska, S. Elnikety, A. Rowstron Microsoft Research Cambridge,
Performance Acceleration Module Tech ONTAP Live August 6 th, 2008.
© Hitachi Data Systems Corporation All rights reserved. 1 1 Det går pænt stærkt! Tony Franck Senior Solution Manager.
Solutions Road Show 2014 March 2014 | India Neeraj Matiyani Director Enterprise Storage Solutions Changing the Economics of Storage: Flash at the Price.
1 Some Context for This Session…  Performance historically a concern for virtualized applications  By 2009, VMware (through vSphere) and hardware vendors.
CS 149: Operating Systems April 21 Class Meeting
Hands-On Microsoft Windows Server 2008
How to Resolve Bottlenecks and Optimize your Virtual Environment Chris Chesley, Sr. Systems Engineer
SEDA: An Architecture for Well-Conditioned, Scalable Internet Services
Reliable Datagram Sockets and InfiniBand Hanan Hit NoCOUG Staff 2010.
© 2011 Cisco All rights reserved.Cisco Confidential 1 APP server Client library Memory (Managed Cache) Memory (Managed Cache) Queue to disk Disk NIC Replication.
Module – 4 Intelligent storage system
Improving Network I/O Virtualization for Cloud Computing.
Uncovering the Multicore Processor Bottlenecks Server Design Summit Shay Gal-On Director of Technology, EEMBC.
Designing and Deploying a Scalable EPM Solution Ken Toole Platform Test Manager MS Project Microsoft.
DBI313. MetricOLTPDWLog Read/Write mixMostly reads, smaller # of rows at a time Scan intensive, large portions of data at a time, bulk loading Mostly.
Achieving Scalability, Performance and Availability on Linux with Oracle 9iR2-RAC Grant McAlister Senior Database Engineer Amazon.com Paper
Srihari Makineni & Ravi Iyer Communications Technology Lab
Copyright © 2014 Oracle and/or its affiliates. All rights reserved. | Deploy Infrastructure Faster with Oracle’s Enterprise Cloud Infrastructure- ECI October.
Eric Burgener VP, Product Management A New Approach to Storage in Virtual Environments March 2012.
Consolidation and Optimization Best Practices: SQL Server 2008 and Hyper-V Dandy Weyn | Microsoft Corp. Antwerp, March
Hyper-V Performance, Scale & Architecture Changes Benjamin Armstrong Senior Program Manager Lead Microsoft Corporation VIR413.
Intel Research & Development ETA: Experience with an IA processor as a Packet Processing Engine HP Labs Computer Systems Colloquium August 2003 Greg Regnier.
Measuring the Capacity of a Web Server USENIX Sympo. on Internet Tech. and Sys. ‘ Koo-Min Ahn.
Full and Para Virtualization
6.894: Distributed Operating System Engineering Lecturers: Frans Kaashoek Robert Morris
An Efficient Threading Model to Boost Server Performance Anupam Chanda.
Jeff Mealiffe Sr. Program Manager Microsoft Corporation SESSION CODE: UNC301 Evan Morris Sr. Systems Engineer Hewlett-Packard.
Capacity Planning in a Virtual Environment Chris Chesley, Sr. Systems Engineer
LIOProf: Exposing Lustre File System Behavior for I/O Middleware
Introduction to Exadata X5 and X6 New Features
Copyright © 2013, Oracle and/or its affiliates. All rights reserved. 1.
Introduction to Performance Tuning Chia-heng Tu PAS Lab Summer Workshop 2009 June 30,
1© Copyright 2015 EMC Corporation. All rights reserved. NUMA(YEY) BY JACOB KUGLER.
Introduction to Operating Systems Concepts
The Docker Container Approach to Build Scalable and Performance Testing Environment Pankaj Rodge, VMware.
sponsored by HP Enterprise
Getting the Most out of Scientific Computing Resources
EonStor DS 2000.
Pro-Active Performance Engineering
Arrakis: The Operating System is the Control Plane
Getting the Most out of Scientific Computing Resources
Lecture 2: Performance Evaluation
Diskpool and cloud storage benchmarks used in IT-DSS
Oracle 11g Real Application Clusters Advanced Administration
HPE Persistent Memory Microsoft Ignite 2017
Installation and database instance essentials
Multi-PCIe socket network device
TCB 9 RK/MSFT Benchmark Results
Boost Linux Performance with Enhancements from Oracle
Upgrading to Microsoft SQL Server 2014
Migration Strategies – Business Desktop Deployment (BDD) Overview
VNX Storage Report Project: Sample VNX Report Project ID:
Unity Storage Array Profile
Predictive Performance
QNX Technology Overview
High-Performance Storage System for the LHCb Experiment
Windows Virtual PC / Hyper-V
Database System Architectures
Xing Pu21 Ling Liu1 Yiduo Mei31 Sankaran Sivathanu1 Younggyun Koh1
Presentation transcript:

PAE SuperCluster Peformance June 12, 2012

SuperCluster/Exadata Dashboard Focus Area Previous Status (05/23) Current Status (06/12) Comments Virtualization (scaling-within-a-box) G OLTP: Consistent performance across zones achieved by manually re-targeting ixgbe interrupt with SR-IOV VF. Currently testing on zones w/ SR-IOV VF IB support verifying that interrupt re-targeting is working as expected. DSS: CPU-intensive queries scaling close to linear from 1 to 8 zones (1 zone/core) on single-socket in split PCI-E LDOM. IO-intensive query doesn’t scale due to IO bottleneck. Exadata Expansion Rack Evaluation Avg IB latency of cells in Expansion Rack same as in SSC rack rds-stress peaks at 1.08M IOPS at 8 cells. Degrades to 0.9M at 21 cells. Orion 8k peak of 850K IOPS, avg 815K IOPS @ 6-8 cells. Throughput degrades slowly to 21 cells at 790K IOPS. Orion 8K IOPS: 24 cells scaling from 1 to 4 T4-4’s non-partitioned: 2.45M IOPS 1M Throughput: 24 cells: 1xT4-4=10.8GB/s, 4xT4-4=24GB/s IOPS issues due to CR 7169352, Throughput issues under investigation 2-socket 11gR2 LDOM vs T4-2 11GR2 LDom has 5% better remote latency than T4-2 (lmbench) iGenRAM & iGenCPU tests result in equivalent performance Storage Cells vs Cached Storage Array (log file sync) Flexcube log file sync issue due to I/O saturation of Storage Cells. Igen oltp log file sync improved from 6-9ms to 2ms w/ disabling nxge driver Storage Cell Flashlog Disk First investigation Niwot Controller Cache is faster than Aura flash for log writes: 97% of log writes satisfied by disk. If controller cache turned off 100% log writes satisfied by flash. With heavy concurrent read workload disk writes redo 100MB/s faster than flash. Exadata Flashcache with Aura2 R On hold: Exadata Storage Server software to support Aura2 not be available. RAC Scalability On Hold: Resources redirected to Storage Cells vs. Cached Storage Array project.

SuperCluster/Exadata – Highlights Performance issues reported by Field/non-DPA Orgs T4-2 vs 2-Socket 11gR2 LDOM CPU/Memory Performance Issue: T4-2 has better CPU and memory performance compared to 2-socket 11gR2 LDOM CPU/Memory Resolution: PAE repeated tests using iGenCPU and iGenRAM resulted in equivalent performance Exadata Storage Cell vs Cached Storage Array (log file sync issues) Issue: iGenOLTP on SSC had 35% better Response Time with Redo Logs on ZFSSA 7320 than on Exadata Storage Cells with FlashLog Resolution: Disabling nxge driver reduced Log File Sync from 6-9ms to 2ms. Response Time with Storage Cells is now 1.6x better than ZFSSA Issue: General Observation that with Exadata FlashLog only small percentage of redo writes are satisfied by flash. Resolution: Purpose of FlashLog is to eliminate outliers when controller cache is congested. Confirmed that Niwot Controller Cache on Exadata Storage is faster than Aura flash for log writes: 97% of log writes satisfied by disk. If controller cache turned off 100% log writes satisfied by flash. FlashLog behavior is correct. SuperCluster/Exadata – Highlights Confidential

SuperCluster/Exadata – Lowlights Performance issues reported by Field/non-DPA Orgs Issue: Flexcube team reported that performance for given batch step was 2.2x faster on T4-2 w/ ST6180 than on SSC 11gR2 Ldom w /3xCells due to redo log performance. Review of data showed I/O saturated on the storage cells. Also test conditions were not exactly the same. Highlights that SSC/Exadata performance work has concentrated on read performance only. Will expand scope to include write-intensive workloads. Evaluate Exadata Expansion Rack Target Goals for Orion I/O tests on Full Rack SSC connect to Full Exadata Expansion Rack (4xT4-4 and 24xExadata Storage Cells) 3M IOPS (random 8K reads) 48 GB/s (random 1M reads) Actual Results 2.45M IOPS (random 8k reads) 24GB/s (random 1M reads) Performance shortfall due mainly to 56 event queues/socket limitation which results in cpus sharing interrupts (CR 7169352) Exadata FlashCache with Aura2 Exadata Storage Server software to support Aura2 is still not available for testing. SuperCluster/Exadata – Lowlights Confidential

SuperCluster/EECS Performance Dashboard Focus Area Previous Status (02/12) Current Status (03/25) Comments MWM-E Met and exceeded December RR goals G Self-Tuning/worker thread optimization now works generally well based on MWM-E Lock-Less Resource Manager (LLRM) does not show overhead. LLRM is recent EECS optimization Improved OTD performance by 20%. Can reach IR 5900 using OTD or IR 6000 without OTD Virtualization (scaling-within-a-box) N/A WIP Characterize & Analyze VM, both Zones and SR-IOV LDOMs, scaling based on current SSC split PCIe LDoms configurations Current low-level SR-IOV IB preliminary performance status : Single Connection latency over VF to external client is 5-10% worse than baremetal to external client Request/response profile performance – IPoIB over VF performance is 30% less than as expected. Under investigation OTD Observe throughput drop with higher connections (IPoIB to WLS). Under Investigation Differentiator exploration – OTD + ILB: Solaris ILB is Layer 4 kernel Load Balancer. When conducting layer 4 LB, ILB consumes ¼ CPU of OTD for delivering the same throughput Found out ILB performance under IB UD mode is much better than under CM mode due to lack of soft-interrupts feature under IB CM mode. Prototyping in progress Infinibus/Coherence Proof of concept status: Unit tests passed, both C and Java portion, on Solaris x86 platform SPARC porting in progress TBD: likely need OFUV library optimization – from gcc to Sun Studio

OTD vs OTD + ILB Performance With Layer 4 LB Rules OTD Only With ILB (+OTD)

OTD Differentiator Next Step OTD + ILB + OpenFlow OS Kernel (ILB) Hardware To-Do:SSL offloading and acceleration Back-end servers Clients