How To : Recover the InnoDB Cluster from complete outage.

I’ve been playing with InnoDB cluster the whole weekend. on this post i’m trying to start my cluster from complete outage.

i don’t have idea which is the primary to bootstrap. using the manual way i know how to do it.
But now, i want to use MySQL Shell.

after shutdown all my 3 node then start it again. below is the current status. —>

How To : Creating an InnoDB Cluster From an Existing Group Replication Deployment

#MySQL #InnoDB cluster is a collection of products that work together to provide a complete High Availability solution for MySQL.
A group of MySQL servers can be configured to create a cluster using MySQL Shell.
In the default single-primary mode, the cluster of servers has a single read-write primary.
Multiple secondary servers are replicas of the primary.
Creating a cluster with at least three servers ensures a high availability cluster.
A client application is connected to the primary via MySQL Router.
If the primary fails, a secondary is automatically promoted to the role of primary,
and MySQL Router routes requests to the new primary.
Advanced users can also configure a cluster to have multiple-primaries.

On my previous post i already created my group replication, but to maximize the use feature of mysql router, we need a shell. —>

How To : Setup MySQL Group Replication

MySQL Group Replication is a MySQL Server plugin that enables you to create elastic, highly-available, fault-tolerant replication topologies.

Groups can operate in a single-primary mode with automatic primary election, where only one server accepts updates at a time. Alternatively, for more advanced users, groups can be deployed in multi-primary mode, where all servers can accept updates, even if they are issued concurrently.

There is a built-in group membership service that keeps the view of the group consistent and available for all servers at any given point in time. Servers can leave and join the group and the view is updated accordingly. Sometimes servers can leave the group unexpectedly, in which case the failure detection mechanism detects this and notifies the group that the view has changed. This is all automatic.

As i prepare to roll out a 6 Node Group Replication in production, we are currently in the process of setting up our UAT environment.
It will be running 3 Node cluster.