IBM ATS Storage © 2013 IBM Corporation What are Consistency Groups ? The concept of grouping all system, middleware, and application volumes that are required.

Slides:



Advertisements
Similar presentations
Remus: High Availability via Asynchronous Virtual Machine Replication
Advertisements

Lectures on File Management
Chapter 16: Recovery System
© 2010 IBM Corporation ® Tivoli Storage Productivity Center for Replication Billy Olsen.
Business Continuity Section 3(chapter 8) BC:ISMDR:BEIT:VIII:chap8:Madhu N PIIT1.
1 CSIS 7102 Spring 2004 Lecture 8: Recovery (overview) Dr. King-Ip Lin.
© 2009 EMC Corporation. All rights reserved. Introduction to Business Continuity Module 3.1.
1 Disk Based Disaster Recovery & Data Replication Solutions Gavin Cole Storage Consultant SEE.
1 © Copyright 2010 EMC Corporation. All rights reserved. EMC RecoverPoint/Cluster Enabler for Microsoft Failover Cluster.
ICS (072)Database Recovery1 Database Recovery Concepts and Techniques Dr. Muhammad Shafique.
Keith Burns Microsoft UK Mission Critical Database.
CS-550 (M.Soneru): Recovery [SaS] 1 Recovery. CS-550 (M.Soneru): Recovery [SaS] 2 Recovery Computer system recovery: –Restore the system to a normal operational.
Chapter 12 File Management Systems
Module – 11 Local Replication
Module – 12 Remote Replication
1 I/O Management in Representative Operating Systems.
Section 3 : Business Continuity Lecture 29. After completing this chapter you will be able to:  Discuss local replication and the possible uses of local.
1© Copyright 2011 EMC Corporation. All rights reserved. EMC RECOVERPOINT/ CLUSTER ENABLER FOR MICROSOFT FAILOVER CLUSTER.
Module 14: Scalability and High Availability. Overview Key high availability features available in Oracle and SQL Server Key scalability features available.
Session 3 Windows Platform Dina Alkhoudari. Learning Objectives Understanding Server Storage Technologies Direct Attached Storage DAS Network-Attached.
Storwize V7000 IP Replication solution explained
Ravi Sankar Technology Evangelist | Microsoft Corporation
IBM TotalStorage ® IBM logo must not be moved, added to, or altered in any way. © 2007 IBM Corporation Break through with IBM TotalStorage Business Continuity.
Backup & Recovery 1.
Building Highly Available Systems with SQL Server™ 2005 Vineet Gupta Evangelist – Data and Integration Microsoft Corp.
Copyright © 2009 EMC Corporation. Do not Copy - All Rights Reserved.
Maintaining a Microsoft SQL Server 2008 Database SQLServer-Training.com.
Managing Multi-User Databases AIMS 3710 R. Nakatsu.
1 Chapter 12 File Management Systems. 2 Systems Architecture Chapter 12.
2Q2008 System z High Availability – Parallel Sysplex TGVL: System z Foundation 1 System z High Availability – Value of Parallel Sysplex IBM System z z10.
Running Work in Parallel. OS/390 is known for its strength and dependability in processing applications that solve large business problems. These are.
DB-2: OpenEdge® Replication: How to get Home in Time … Brian Bowman Sr. Solutions Engineer Sandy Caiado Sr. Solutions Engineer.
Continuous Access Overview Damian McNamara Consultant.
Storage 101: Bringing Up SAN Garry Moreau Senior Staff Alliance Consultant Ciena Communications (763)
Module 9 Planning a Disaster Recovery Solution. Module Overview Planning for Disaster Mitigation Planning Exchange Server Backup Planning Exchange Server.
Computer Emergency Notification System (CENS)
Recovery System By Dr.S.Sridhar, Ph.D.(JNUD), RACI(Paris, NICE), RMR(USA), RZFM(Germany) DIRECTOR ARUNAI ENGINEERING COLLEGE TIRUVANNAMALAI.
Copyright © 2009 EMC Corporation. Do not Copy - All Rights Reserved.
Chapter 16 Recovery Yonsei University 1 st Semester, 2015 Sanghyun Park.
Components of a Sysplex. A sysplex is not a single product that you install in your data center. Rather, a sysplex is a collection of products, both hardware.
Data Sharing. Data Sharing in a Sysplex Connecting a large number of systems together brings with it special considerations, such as how the large number.
Continuous Availability
Chapter 10 Recovery System. ACID Properties  Atomicity. Either all operations of the transaction are properly reflected in the database or none are.
High Availability in DB2 Nishant Sinha
Copyrighted material John Tullis 12/16/2015 page 1 04/08/00 MQ Series Middleware Presentation John Tullis DePaul Instructor
Peter Mattei HP Storage Consultant 16. May 2013
Chapter 20 Parallel Sysplex
Coupling Facility. The S/390 Coupling Facility (CF), the key component of the Parallel Sysplex cluster, enables multisystem coordination and datasharing.
© 2006 EMC Corporation. All rights reserved. The Host Environment Module 2.1.
Oracle Architecture - Structure. Oracle Architecture - Structure The Oracle Server architecture 1. Structures are well-defined objects that store the.
© 2009 IBM Corporation Statements of IBM future plans and directions are provided for information purposes only. Plans and direction are subject to change.
Praveen Srivatsa Director| AstrhaSoft Consulting blogs.asthrasoft.com/praveens |
© 2012 IBM Corporation IBM HyperSwap Options September 2012.
Database Recovery Zheng (Godric) Gu. Transaction Concept Storage Structure Failure Classification Log-Based Recovery Deferred Database Modification Immediate.
Advanced Technical Support (ATS) Americas © 2007 IBM Corporation What is FlashCopy? FlashCopy® is an “instant” T0 (Time 0) copy where the source and target.
IBM Advanced Technical Support (ATS) © 2010 IBM Corporation What is HyperSwap ? Ability to swap enterprise class System z™ data centers in a matter of.
SysPlex -What’s the problem Problems are growing faster than uni-processor….1980’s Leads to SMP and loosely coupled Even faster than SMP and loosely coupled.
Metro Mirror, Global Copy, and Global Mirror Quick Reference
Tivoli Storage Manager Product Family
Managing Multi-User Databases
High Availability Options with Storage
Maximum Availability Architecture Enterprise Technology Centre.
File Processing : Recovery
Chapter 9 – RPCs, Messaging & EAI
Outline Announcements Fault Tolerance.
DS6000/DS8000 FlashCopy in z/OS Environments Quick Reference
Recovery System.
Presentation transcript:

IBM ATS Storage © 2013 IBM Corporation What are Consistency Groups ? The concept of grouping all system, middleware, and application volumes that are required to be managed as a consistent entity has become known as a Consistency Group. Dependent write I/O sequences are illustrated in several examples in which the start of one write is time dependent on the completion of a previous write.  Database and log  Catalogs, volume table of contents  Index and data components Time sequences can be exposed whenever a copy of the data is being made at the physical record/block level. Consistency Groups enable customers to group ‘related volumes’ together representing a “business transaction” that may involve multiple transactions across a single or multiple databases / file systems located on a single or multiple heterogeneous server platform. Consistency Groups in a Nutshell This trifold covers the following topics:  What are Consistency Groups (CGs) ?  Dependent write I/O sequences  Why MUST I have Consistency Groups ?  Important for PiT Copy Solutions  Important for D/R (DB Restart instead of DBRecovery)  Important for data integrity  Scope can be Disk Storage Subsystem BUT generally requires Global Systems Level Monitoring.  What are the various ways for data replication technologies to create CGs?  Time based CGs  Data freeze CGs  Ordered I/O CGs  Quiescing / Stopping the application(s)  Specific application, database, or file system management For more information, Please contact your local IBM representative or IBM Business Partner. August 2013 It is imperative that these “dependent writes” be written to disk in the same sequence that they were issued by the application, as shown in points (1) through (3) above. Furthermore, continuing with the example above, there is no guarantee that the database log and the database itself reside on the same device or even in the same storage subsystem. Failure to execute the write sequence correctly may result in I/O operations (1) and (3) being executed, followed immediately by a system failure. When it came time to recover the database, the database log would indicate that the transaction had completed successfully when, in fact, that is not the case. The transaction would be lost, and the integrity of the database would be in question. Maintaining “dependent write I/O consistency”, also commonly called “Power Fail Consistency” is a critical component in all data replication products.. Dependent write I/O sequences Why must I have CGs ? With Data Replication Consistency Groups, cross- volume data integrity / data consistency dependent write I/O consistency is maintained on the physical level. However, data consistency at this level is very different from data consistency at a database, file system, or application level. For block / record-based data replication techniques, Consistency Groups insure that the alternate physical image of the data across the entire Consistency Group is maintained. Consistency Groups enable databases to perform a database RESTART function, which examines the DB logs and tables, backing out “in flight and In doubt ‘transactions,’ making the DB “transaction” consistent. Depending on the application, this may or may not be enough. A “business transaction” may involve a single or multiple transactions across a single or multiple applications. So, the recovery process at a “business transaction” level may require multiple levels of application(s) restarts in which each may involve a single or multiple database and/or file system restart(s). Again, the scope of “business transactions” may be across a single or multiple servers and operating system platforms. © International Business Machines Corporation, IBM, the IBM logo, GDPS, Global Dispersed Parallel Sysplex, HyperSwap, z/VM, z/OS, z.VSE, System z, are trademarks of International Business Machines Corporation in the United States, other countries or both. Microsoft, Windows, Windows NT, and the Windows logo are trademarks of Microsoft Corporation in the United States, other countries, or both. Other company, product and service names may be trademarks or service marks of others.

IBM ATS Storage © 2013 IBM Corporation Data Freeze CGs Automation code such as Geographically Dispersed Parallel Sysplex (GDPS®) and TPC-R exploit a storage subsystem feature known as CGROUP Freeze/Run. Software automation invokes these functions on hardware- or software-related errors to cause all synchronous mirrored volumes to be suspended while maintaining all dependent I/O. Numerous ‘trigger’ events are raised from the storage subsystem hardware as well as various messages / alerts raised by the software. The data freeze is used in functions such as HyperSwap, Consistent FlashCopy®, MM session management (GDPS, TPC- R, etc.) and Global Mirror (toggle process). Ordered I/O CGs IBM Storage Systems SAN Volume Controller Global Mirror (as well as many SAN-based data replication products) maintain cross-volume data consistency within the scope of all I/Os that go through that device by ordering the I/Os “first in, first out” (FIFO) for transport to the remote device. Maintaining write I/O order in a single queue for all devices that the data replication product manages ensures that dependent I/O operations are managed correctly. Time Based CGs XRC uses the IBM System z™ SYSPLEX Timer to timestamp all write I/O to XRC Active devices. The DFSMSsdm (System Data Mover) reads all updates from all primary storage subsystems and forms CGs based on time. Quiesce/Stop Application(s) CGs A final method of maintaining dependent I/O within or across multiple storage subsystem- attached devices, when those devices are involved in some form of data replication, is to stop all I/O activity. This can be done a number of different ways, including: Database quiesce Stopping the application(s) involved DB Command SET SUSPEND LOG Application quiesce On Open Systems, some customers like this method because it can flush writes completed in the host file system to the disk subsystem before a disk data replication “Sync” point is taken. XRC SDM uses common Sysplex Timer timestamp to sort the incoming data, form consistency groups 10's of times per second across large numbers of volumes, disk frames, and z/OS® images. XRC is able to back out in-flight incomplete write sequences because in event of outage, the XRC SDM does not write out incomplete data, thus what is on the disk is the most recent complete Consistency Group. Application Ordering Specific application, database, or file system knowledge is used to maintain cross-volume data consistency and integrity with many software-based replication schemes that are available with specific software databases, file systems and applications. XRC System Data Mover CGs 10's C ®. c INCREASING TIME z/OS Application 1 Server/Application 2 Server/Application 3 Server/Application 4 Server/Application 5 Server/Application 6 Server/Application 7 Server/Application 8 Server/Application 9 Host Disaster Recovery - Freeze Control Software Control Unit Control Unit Host Control Unit Control Unit Host Control Unit Control Unit Host Control Unit Control Unit During Control UnitControl UnitControl Unit 1. Both sites are suspend but in sync 2. Sites are not insync, however data to a site is consistent Control Unit Notification of suspending event CU holds I/Os to suspended device until SW can handle condition 1. GDPS or TPC-R SW issues freeze command to all Storage Subsystems a. suspends all PPRC pairs b. holds I/O until told to continue 2. Based on policy, SW will then allow host operation to resume or continue to hold them Control UnitControl UnitControl UnitControl Unit After, depending on policy: GDPS/PPRC Data Freeze Based Consistency Groups ERROR EVENT ANALYSIS Customer business policies Freeze & Go, Freeze & Stop, Freeze & Stop Conditional, Freeze & SWAP SITUATION MANAGEMENT Execute Freeze policy Remove failed site systems from Parallel Sysplex ® Switch DASD Acquire processing resources Switch Network Restart workload FREEZE AUTOMATION I/O consistent secondaries Across multiple CUs Multiple host systems Sysplex Triggers Disk subsystems Triggers Unit Checks SNMP Alerts VTS or TS7700 Tape Replication CF Triggers GDPS/PPRC Data Freeze-based Consistency Groups