WebSphere XD Compute Grid High Performance Architectures

Slides:



Advertisements
Similar presentations
Tableau Software Australia
Advertisements

Scheduling in Web Server Clusters CS 260 LECTURE 3 From: IBM Technical Report.
SSRS 2008 Architecture Improvements Scale-out SSRS 2008 Report Engine Scalability Improvements.
CERN LCG Overview & Scaling challenges David Smith For LCG Deployment Group CERN HEPiX 2003, Vancouver.
PowerVM Live Partitioned Mobility A feature of IBM Virtualization Presented by Group 3 Mayra Longoria Mehdi Jafry Ken Lancaster PowerVM Live Partitioned.
Bookshelf.EXE - BX A dynamic version of Bookshelf –Automatic submission of algorithm implementations, data and benchmarks into database Distributed computing.
A Grid Resource Broker Supporting Advance Reservations and Benchmark- Based Resource Selection Erik Elmroth and Johan Tordsson Reporter : S.Y.Chen.
Distributed Database Management Systems
Computer Organization and Architecture
1 Operating Systems Ch An Overview. Architecture of Computer Hardware and Systems Software Irv Englander, John Wiley, Bare Bones Computer.
Introduction to the new mainframe: Large-Scale Commercial Computing © Copyright IBM Corp., All rights reserved. Chapter 3: Scalability.
NPACI: National Partnership for Advanced Computational Infrastructure August 17-21, 1998 NPACI Parallel Computing Institute 1 Cluster Archtectures and.
Towards Autonomic Hosting of Multi-tier Internet Services Swaminathan Sivasubramanian, Guillaume Pierre and Maarten van Steen Vrije Universiteit, Amsterdam,
Understanding and Managing WebSphere V5
Overview SAP Basis Functions. SAP Technical Overview Learning Objectives What the Basis system is How does SAP handle a transaction request Differentiating.
Domino MailDomino AppsQuickPlace Sametime Domino WebHub / SMTP Top Ten Reasons to Consolidate Lotus  Workloads on IBM eServer  iSeries  and eServer.
Grid Computing Meets the Database Chris Smith Platform Computing Session #
Distributed Data Stores – Facebook Presented by Ben Gooding University of Arkansas – April 21, 2015.
Building Scalable.NET Applications Guy Nirpaz, EVP R&D, GigaSpaces Technologies.
Word Wide Cache Distributed Caching for the Distributed Enterprise.
Cloud Computing for the Enterprise November 18th, This work is licensed under a Creative Commons.
1 Copyright © 2004, Oracle. All rights reserved. Introduction to Oracle Forms Developer and Oracle Forms Services.
+ CS 325: CS Hardware and Software Organization and Architecture Cloud Architectures.
Oracle Coherence Product Overview Raanan Dagan / Coherence Team.
Principles of Scalable HPC System Design March 6, 2012 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
ENTERPRISE COMPUTING QUIZ By: Lean F. Torida
October 2008, Nice Using ProActive to bridge Grid and SOA The AGOS project Frédéric Linder Program Director, Service Oriented Architecture.
MACIASZEK, L.A. (2001): Requirements Analysis and System Design. Developing Information Systems with UML, Addison Wesley Chapter 6 - Tutorial Guided Tutorial.
Week 5 Lecture Distributed Database Management Systems Samuel ConnSamuel Conn, Asst Professor Suggestions for using the Lecture Slides.
Click to add text Introduction to the new mainframe: Large-Scale Commercial Computing © Copyright IBM Corp., All rights reserved. Chapter 2: Capacity.
Oracle 10g Database Administrator: Implementation and Administration Chapter 2 Tools and Architecture.
DataMigrator Data Analysis with WebFOCUS. 2 Metadata Data Lineage Data Profiling Data Transformation Administration Connectivity Portability DataMigrator.
An application architecture specifies the technologies to be used to implement one or more (and possibly all) information systems in terms of DATA, PROCESS,
Database Architectures Database System Architectures Considerations – Data storage: Where do the data and DBMS reside? – Processing: Where.
© 2008 IBM Corporation Snehal S. Antani WebSphere XD Technical Lead SOA Technology Practice, ISSW, IBM WebSphere XD Compute Grid Developing Tooling Story.
© 2008 IBM Corporation AIX Workload Partion Manger.
© 2008 IBM Corporation Swiss Re Case Study: Bringing Together Batch Processing and OLTP in a Unified Computing Environment Chris Vignola STSM, XD Compute.
Virtualization and Databases Ashraf Aboulnaga University of Waterloo.
Infrastructure for Data Warehouses. Basics Of Data Access Data Store Machine Memory Buffer Memory Cache Data Store Buffer Bus Structure.
© 2007 IBM Corporation Snehal S. Antani, WebSphere XD Technical Lead SOA Technology Practice IBM Software WebSphere.
© 2008 IBM Corporation Snehal S. Antani WebSphere XD Technical Lead SOA Technology Practice, ISSW, IBM TSM-1073 Enterprise Grid and Batch Computing with.
CLIENT SERVER COMPUTING. We have 2 types of n/w architectures – client server and peer to peer. In P2P, each system has equal capabilities and responsibilities.
® IBM Software Group © IBM Corporation 應用程式基礎設備虛擬化 — WebSphere Virtual Enterprise Lilian Wang( 王漪萍 ) WebSphere Technical Sales Support IBM Taiwan Software.
Tivoli Workload Scheduler for Applications PeopleSoft Integration
1 Copyright © 2008, Oracle. All rights reserved. Repository Basics.
Cofax Scalability Document Version Scaling Cofax in General The scalability of Cofax is directly related to the system software, hardware and network.
SysPlex -What’s the problem Problems are growing faster than uni-processor….1980’s Leads to SMP and loosely coupled Even faster than SMP and loosely coupled.
Md Baitul Al Sadi, Isaac J. Cushman, Lei Chen, Rami J. Haddad
Introduction to Oracle Forms Developer and Oracle Forms Services
Web Server Load Balancing/Scheduling
CIIT-Human Computer Interaction-CSC456-Fall-2015-Mr
Web Server Load Balancing/Scheduling
Flash Storage 101 Revolutionizing Databases
Measurement-based Design
Introduction to Oracle Forms Developer and Oracle Forms Services
IBM Tivoli Web Site Analyzer Training Document
Introduction to Oracle Forms Developer and Oracle Forms Services
VIRTUAL SERVERS Presented By: Ravi Joshi IV Year (IT)
IBM Linux Update for CAVMEN
What is the Azure SQL Datawarehouse?
20409A 7: Installing and Configuring System Center 2012 R2 Virtual Machine Manager Module 7 Installing and Configuring System Center 2012 R2 Virtual.
Tiers vs. Layers.
Design pattern for cloud Application
AWS Cloud Computing Masaki.
Web Application Server 2001/3/27 Kang, Seungwoo. Web Application Server A class of middleware Speeding application development Strategic platform for.
Cloud Computing Architecture
Wide Area Workload Management Work Package DATAGRID project
Performance And Scalability In Oracle9i And SQL Server 2000
Presentation transcript:

WebSphere XD Compute Grid High Performance Architectures Snehal S. Antani, antani@us.ibm.com WebSphere XD Technical Lead SOA Technology Practice IBM Software Services

Overview Key Components of WebSphere XD Compute Grid Job Scheduler [formerly named the Long Running Scheduler (LRS)] Parallel Job Manager (PJM) Grid Endpoints [formerly named Long Running Execution Environment (LREE)] High Performance Architectures and Considerations

XD Compute Grid Components Job Scheduler (JS) The job entry point to XD Compute grid Job life-cycle management (Submit, Stop, Cancel, etc) and monitoring Dispatches workload to either the PJM or LREE Hosts the Job Management Console (JMC) Parallel Job Manager (PJM)- Breaks large batch jobs into smaller partitions for parallel execution Provides job life-cycle management (Submit, Stop, Cancel, Restart) for the single logical job and each of its partitions Is *not* a required component in compute grid Grid Endpoints (GEE) Executes the actual business logic of the batch job

XD Compute Grid Components Load Balancer EJB Web Service JMS Command Line Job Console GEE WAS User JS WAS GEE WAS PJM WAS

Key Influencers for High Performance Compute Grids Proximity to the Data Bring the business logic to the data: co-locate on the same platform Bring the data to the business logic: in-memory databases, caching Affinity Routing Partitioned data with intelligent routing of work Divide and Conquer Highly parallel execution of workloads across the grid On-Demand Scalability

Proximity to the Data- Co-location of business logic with data Frame Job Scheduler WAS z/OS WAS z/OS Controller Controller GEE GEE GEE GEE GEE GEE Servant Servant Servant Servant Servant Servant DB2 on z/OS

Proximity to the Data- Bring data to the business logic with caching Frame CPU CPU Job Scheduler LPAR GEE GEE GEE Data Grid near-cache Data Grid near-cache Data Grid near-cache CPU CPU CPU CPU CPU CPU LPAR LPAR LPAR CPU CPU DG Server DG Server LPAR LPAR Database

Affinity Routing- Partitioned data with intelligent routing of work Frame Job Scheduler Records A-M Records N-Z WAS z/OS WAS z/OS Controller Controller A-D E-I J-M N-Q R-T W-Z GEE GEE GEE GEE GEE GEE Servant Servant Servant Servant Servant Servant DB2 Data Sharing Partition DB2 Data Sharing Partition Records A-M Records N-Z

Affinity Routing- Partitioned data with intelligent routing of work Frame CPU CPU Job Scheduler GEE GEE GEE Data Grid near-cache Data Grid near-cache Data Grid near-cache CPU CPU CPU CPU CPU CPU Records A-I Records J-R Records S-Z CPU CPU DG Server DG Server Records A-M Records N-Z Database

Divide and Conquer- Highly Parallel Grid Jobs Large Grid Job Frame CPU CPU PJM GEE GEE GEE Data Grid near-cache Data Grid near-cache Data Grid near-cache CPU CPU CPU CPU CPU CPU Records A-I Records J-R Records S-Z CPU CPU DG Server DG Server Records A-M Records N-Z Database

On-Demand Scalability- With WebSphere z/OS Frame Job Scheduler WAS z/OS WAS z/OS Controller Controller zWLM zWLM GEE GEE GEE GEE GEE GEE Servant Servant Servant Servant Servant Servant DB2 on z/OS

On-Demand Scalability- With XD Operations Optimization Frame CPU CPU CPU CPU Job Scheduler On-Demand Router LPAR LPAR GEE GEE GEE Data Grid near-cache Data Grid near-cache Data Grid near-cache CPU CPU CPU CPU CPU CPU LPAR LPAR LPAR CPU CPU DG Server DG Server LPAR LPAR Database

Backup

OG Server Miss (Access DB) Near-Cache Hit Probability of cache hit Time (ms) to retrieve data from cache Data Access Probability that data is in cache server OG Server Hit Time (ms) to retrieve data from cache server Probability of cache miss Near-Cache Miss Time (ms) to retrieve data from other storage Probability that data must be retrieved from database OG Server Miss (Access DB) Time (ms) to retrieve data from database Data Access time (ms) = (Probability of near- cache hit) * (Time to retrieve data from near-cache) + (Probability of near-cache miss) * (time to retrieve data from other storage); Time to retrieve data from other storage (ms) = (Probability that data is in cache server) * (Time to retrieve data from cache server) + (Probability that data must be retrieved from database) * (time to retrieve data from database);

OG Server Miss (Access DB) Near-Cache Hit P1 S1 Data Access OG Server Hit P3 S3 P2 Near-Cache Miss S2 P4 Data Access = (Near-Cache Hit) + (Near-Cache Miss) Near-Cache Hit = (P1)(S1) Near-Cache Miss = (P2) * [ (P3)(S3) + (P4)(S4) ] Improve data access time by: Increase P1: increase cache size (increase heap, etc) establish request affinity Decrease S1: Dynamically add more CPU Decrease S2 Increase P3 Increase size of cache server Establish query/data affinity Decrease S3 Decrease S4 Reduce network latency OG Server Miss (Access DB) S4

OG Server Miss (Access DB) Example calculation Near-Cache Hit 30% 1 ms Data Access OG Server Hit 47.2 ms 70% 10 ms 70% Near-Cache Miss 67 ms 30% OG Server Miss (Access DB) Data Access = (Near-Cache Hit) + (Near-Cache Miss) Near-Cache Hit = (P1)(S1) Near-Cache Miss = (P2) * [ (P3)(S3) + (P4)(S4) ] Near-cache miss = (.7)(10) + (.3)(200) = 7 + 60 = 67 ms Data Access = (.3)(1) + (.7)(67) = .3 + 46.9 = 47.2 ms 200 ms

OG Server Miss (Access DB) Example calculation- effects of increasing size of near-cache Near-Cache Hit 60% 1 ms Data Access OG Server Hit 27.4 ms 70% 10 ms 40% Near-Cache Miss 67 ms 30% OG Server Miss (Access DB) Data Access = (Near-Cache Hit) + (Near-Cache Miss) Near-Cache Hit = (P1)(S1) Near-Cache Miss = (P2) * [ (P3)(S3) + (P4)(S4) ] Near-cache miss = (.7)(10) + (.3)(200) = 7 + 60 = 67 ms Data Access = (.6)(1) + (.4)(67) = .6 + 26.8 = 27.4 ms (47.2 – 27.4) / 47.2 = 42% improvement in data access time 200 ms

OG Server Miss (Access DB) Example calculation- effects of adding more CPU and decreasing network latency to the DB Near-Cache Hit 30% 1 ms Data Access OG Server Hit 26.2 ms 70% 10 ms 70% Near-Cache Miss 37 ms 30% Data Access = (Near-Cache Hit) + (Near-Cache Miss) Near-Cache Hit = (P1)(S1) Near-Cache Miss = (P2) * [ (P3)(S3) + (P4)(S4) ] Near-cache miss = (.7)(10) + (.3)(100) = 7 + 30 = 37 ms Data Access = (.3)(1) + (.7)(37) = .3 + 25.9 = 26.2 ms (47.2 – 25.9) / 47.2 = 45% improvement in data access time OG Server Miss (Access DB) 100 ms

WebSphere XD Compute Grid Infrastructure Topology Considerations