Www.opendaylight.org 2-Node Clustering Active-Standby Deployment.

Slides:



Advertisements
Similar presentations
High Availability Deep Dive What’s New in vSphere 5 David Lane, Virtualization Engineer High Point Solutions.
Advertisements

© 2010 IBM Corporation ® Tivoli Storage Productivity Center for Replication Billy Olsen.
PostgreSQL Clustering Solutions Adam Brust Andrew Henle.
Mecanismos de alta disponibilidad con Microsoft SQL Server 2008 Por: ISC Lenin López Fernández de Lara.
ITIS 3110 Jason Watson. Replication methods o Primary/Backup o Master/Slave o Multi-master Load-balancing methods o DNS Round-Robin o Reverse Proxy.
© 2011 VMware Inc. All rights reserved High Availability Module 7.
Address Configuration in Adhoc Networks in IPv4. By Meenakshi Sundaram V.
High Availability 24 hours a day, 7 days a week, 365 days a year… Vik Nagjee Product Manager, Core Technologies InterSystems Corporation.
® IBM Software Group © 2006 IBM Corporation Rational Software France Object-Oriented Analysis and Design with UML2 and Rational Software Modeler 04. Other.
Active-Standby Deployment
Ethernet and switches selected topics 1. Agenda Scaling ethernet infrastructure VLANs 2.
June 23rd, 2009Inflectra Proprietary InformationPage: 1 SpiraTest/Plan/Team Deployment Considerations How to deploy for high-availability and strategies.
1 © Copyright 2010 EMC Corporation. All rights reserved. EMC RecoverPoint/Cluster Enabler for Microsoft Failover Cluster.
Lesson 1: Configuring Network Load Balancing
1© Copyright 2011 EMC Corporation. All rights reserved. EMC RECOVERPOINT/ CLUSTER ENABLER FOR MICROSOFT FAILOVER CLUSTER.
Virtualization Infrastructure Administration Cluster Jakub Yaghob.
National Manager Database Services
11 SERVER CLUSTERING Chapter 6. Chapter 6: SERVER CLUSTERING2 OVERVIEW  List the types of server clusters.  Determine which type of cluster to use for.
MCTS Guide to Microsoft Windows Server 2008 Applications Infrastructure Configuration (Exam # ) Chapter Ten Configuring Windows Server 2008 for High.
Ronen Gabbay Microsoft Regional Director Yside / Hi-Tech College
NDIS LBFO Miniports (Load Balancing And Failover) Larry Cleeton Program Manager Windows Networking And Communications Microsoft Corporation.
Failover Clustering & Hyper-V: Multisite Disaster Recovery
Hybrid Overlay Multicast Framework draft-irtf-sam-hybrid-overlay-framework-02.txt John Buford, Avaya Labs Research IETF 71.
DATABASE MIRRORING  Mirroring is mainly implemented for increasing the database availability.  Is configured on a Database level.  Mainly involves two.
Module 3: Designing IP Addressing. Module Overview Designing an IPv4 Addressing Scheme Designing DHCP Implementation Designing DHCP Configuration Options.
1 Moshe Shadmon ScaleDB Scaling MySQL in the Cloud.
Chapter 8: Virtual LAN (VLAN)
Site Power OutageNetwork Disconnect Node Shutdown for Patching Node Crash Quorum Witness Failure How do I make sure my Cluster stays up ??... Add/Evict.
11.1 © 2004 Pearson Education, Inc. Exam Designing a Microsoft ® Windows ® Server 2003 Active Directory and Network Infrastructure Lesson 11: Planning.
High-Availability MySQL DB based on DRBD-Heartbeat Ming Yue September 27, 2007 September 27, 2007.
Module 10: Maintaining High-Availability. Overview Introduction to Availability Increasing Availability Using Failover Clustering Standby Servers and.
Module 13 Implementing Business Continuity. Module Overview Protecting and Recovering Content Working with Backup and Restore for Disaster Recovery Implementing.
© Wiley Inc All Rights Reserved. MCSE: Windows Server 2003 Active Directory Planning, Implementation, and Maintenance Study Guide, Second Edition.
Course 301 – Secured Network Deployment and IPSec VPN
 Load balancing is the process of distributing a workload evenly throughout a group or cluster of computers to maximize throughput.  This means that.
1 © 2005 Cisco Systems, Inc. All rights reserved. 111 © 2004, Cisco Systems, Inc. All rights reserved. CNIT 221 Security 2 ver.2 Module 8 City College.
11 CLUSTERING AND AVAILABILITY Chapter 11. Chapter 11: CLUSTERING AND AVAILABILITY2 OVERVIEW  Describe the clustering capabilities of Microsoft Windows.
Geo-distributed Messaging with RabbitMQ
Peter Mattei HP Storage Consultant 16. May 2013
Improving Robustness in Distributed Systems Per Bergqvist Erlang User Conference 2001 (courtesy CellPoint Systems AB)
Network Virtualization in Multi-tenant Datacenters Author: VMware, UC Berkeley and ICSI Publisher: 11th USENIX Symposium on Networked Systems Design and.
Clustering in OpenDaylight
vSphere HA and vSphere Fault Tolerance
Deploy SDN-IP.
High-Availability MySQL with DR:BD and Heartbeat: MTV Japan mobile services ©2008 MTV Networks Japan K.K.
© 2015 Infoblox Inc. All Rights Reserved. Tom Coffeen, IPv6 Evangelist UKNOF January 2015 Tom Coffeen, IPv6 Evangelist UKNOF January 2015 DHCPv6 Operational.
Test and Performance Integration Group.
Atrium Router Project Proposal Subhas Mondal, Manoj Nair, Subhash Singh.
MGT 330 Week 5 Individual Management and Leadership Paper To purchase this material click on below link 330-Week-5-Individual-Management-and-Leadership-
Introduction to Clustering
REPLICATION & LOAD BALANCING
High Availability 24 hours a day, 7 days a week, 365 days a year…
High Availability Linux (HA Linux)
AlwaysOn Mirroring, Clustering
Configuring Cluster Communications
HP ArcSight ESM 6.8c HA Fail Over Illustrated
Cluster Communications
A Technical Overview of Microsoft® SQL Server™ 2005 High Availability Beta 2 Matthew Stephen IT Pro Evangelist (SQL Server)
System and Communication Faults
SDN Overview for UCAR IT meeting 19-March-2014
2018 Real Huawei H Exam Questions Killtest
Introduction to Clustering
Always On : Multi-site patterns
Always On : Multi-site patterns
SpiraTest/Plan/Team Deployment Considerations
Chapter 4: EtherChannel and HSRP
BGP Instability Jennifer Rexford
Chapter 4: EtherChannel and HSRP
Designing Database Solutions for SQL Server
Presentation transcript:

2-Node Clustering Active-Standby Deployment

Requirements Configuration of Primary controller in cluster (Must) Primary Controller services the NorthBound IP address, a Secondary takes over NB IP upon failover (Must) Configuration of whether on failover & recovery, configured Primary controller reasserts leadership (Must) Configuration of merge strategy on failover & recovery (Want) Primary controller is master of all devices and is leader of all shards (Must) Single node operation allowed (access to datastore on non-quorum) (Want) 2-Node Deployment Topology 2 Active-Standby Requirements

Failover Sequence 1. Secondary controller becomes master of all devices and leader of all shards Failure of Primary 3 Scenario 1: Master Stays Offline

Recovery Sequence 1. Controller A comes back online and its data is replaced by all of Controller B’s data 2. For Re-assert leadership configuration: 1. (ON) Controller A becomes master of all devices and leader of all shards 2. (OFF) Controller B stays master of all devices and maintains leadership of all shards Failure of Primary 4 Scenario 2: Primary Comes Back Online

Failover Sequence 1. Controller A becomes master of devices in its network segment and leader of all shards 2. Controller B becomes master of devices in its network segment and leader of all shards Network Partition 5 Scenario 1: During Network Partition

Recovery Sequence 1. Merge data according to pluggable merge strategy (Default: Secondary’s data replaced with Primary’s data.) 2. For Re-assert leadership configuration: 1. (ON) Controller A becomes master of all devices and leader of all shards again. 2. (OFF) Controller B becomes master of all devices and leader of all shards again Network Partition 6 Scenario 2: Network Partition Recovers

Scenarios 1. Secondary controller failure. 2. Any single link failure. 3. Secondary controller loses network connectivity (but device connections to Primary maintained) No-Op Failures 7 Failures That Do Not Result in Any Role Changes

Global 1. Cluster Leader (aka “Primary”) 1. Allow this to be changed on live system, e.g. maintenance. 2. Assigned (2-Node Case), Elected (Larger Cluster Case) 2. Cluster Leader Northbound IP 3. Reassert Leadership on Failover and Recovery 4. Network Partition Detection Alg. (pluggable) 5. Global Overrides of Per Device/Group and Per Shard items (below) Per Device / Group 1. Master / Slave Per Shard 1. Shard Leader (Shard Placement Strategy – pluggable) 2. Shard Data Merge (Shard Merge Strategy – pluggable) Cluster Configuration Options 8 Global & Granular Configuration

Can we Abstract Configurations to Admin-Defined Deployment Scenarios?  e.g. Admin Configures 2-Node (Active-Standby):  This means Primary controller is master of all devices and leader of all shards.  Conflicting configurations are overridden by deployment scenario. HA Deployment Scenarios 9 Simplified Global HA Settings

Clustering: 1. Refactoring of Raft Actor vs. 2-Node Raft Actor code. 2. Define Cluster Leader 3. Define Northbound Cluster Leader IP Alias OpenFlow Plugin: 1. OpenFlow Master/Slave Roles 2. Grouping of Master/Slave Roles (aka “Regions”) System: 1. Be Able to SUSPEND the Secondary controller to support Standby mode. Implementation Dependencies 10 Potential Changes to Other ODL Projects

TBD: 1. Is Master/Slave definition too tied to OpenFlow? (Generalize?) 1. Should device ownership/mastership be implemented by OF Plugin? 2. How to define Northbound Cluster Leader IP in a platform independent way? (Linux/Mac OSx: IP Alias, Windows: Possible) 1. Gratuitous ARP on Leader Change. 3. When both Controllers are active in Network Partition scenario which controller “owns” the Northbound Cluster Leader IP? 4. Define Controller-Wide SUSPEND behavior (how?) 5. On failure Primary controller should be elected (2-node case Secondary is only option to be elected) 6. How/Need to detect management plane failure? (Heartbeat timeout >> w.c. GC?) Open Issues 11 Follow-up Design Discussion Topics