A node in MySQL 5.7 innoDB cluster is crashed and unable to re-join the crashed node to the clusterMysql master-master replicationMySQL delay in Relay Master log fileGot fatal error 1236 from master when reading data from binary log: 'Could not find first log file name in binary log index file'Mysql Replication Slave not getting connected “ERROR 1200 (HY000)”Master to master MySQL 5.7 replicationMySQL SSL replication error - Error_code: 2026MySQL replication fails when it replicate from a slave database serverMySQL node join group replication failed to due extra transactions after failoverMariaDB replication - slave lag due to gtid statements in relay logMySQL 8.0.13 Group Replication Recovery Error MY-002061

Why can't the Brexit deadlock in the UK parliament be solved with a plurality vote?

Did the UK lift the requirement for registering SIM cards?

It grows, but water kills it

Non-trope happy ending?

Make a Bowl of Alphabet Soup

Change the color of a single dot in `ddot` symbol

Has the laser at Magurele, Romania reached a tenth of the Sun's power?

Does the reader need to like the PoV character?

When were female captains banned from Starfleet?

Delete multiple columns using awk or sed

awk assign to multiple variables at once

Creating two special characters

Plot of a tornado shape like surface

What kind of floor tile is this?

Why is so much work done on numerical verification of the Riemann Hypothesis?

How could a planet have erratic days?

Shouldn’t conservatives embrace universal basic income?

15% tax on $7.5k earnings. Is that right?

Showing a sum is positive

How do you make your own symbol when Detexify fails?

Can you use Vicious Mockery to win an argument or gain favours?

Biological Blimps: Propulsion

What fields between the rationals and the reals allow a good notion of 2D distance?

Why do Radio Buttons not fill the entire outer circle?



A node in MySQL 5.7 innoDB cluster is crashed and unable to re-join the crashed node to the cluster


Mysql master-master replicationMySQL delay in Relay Master log fileGot fatal error 1236 from master when reading data from binary log: 'Could not find first log file name in binary log index file'Mysql Replication Slave not getting connected “ERROR 1200 (HY000)”Master to master MySQL 5.7 replicationMySQL SSL replication error - Error_code: 2026MySQL replication fails when it replicate from a slave database serverMySQL node join group replication failed to due extra transactions after failoverMariaDB replication - slave lag due to gtid statements in relay logMySQL 8.0.13 Group Replication Recovery Error MY-002061













0















We have a MySQL innodb cluster in one of our environments. One of the nodes in the cluster was crashed. Though, we were able to bring the crashed node online we were unable to join it to the cluster.



Can someone please help to recover/restore the node and join it to the cluster. We tried to use "dba.rebootClusterFromCompleteOutage()" but it didn't help.



Configuration: MySQL 5.7.24 Community Edition, CentOS 7, standard three node innodb cluster



Cluster Status:



MySQL NODE02:3306 ssl JS > var c=dba.getCluster()
MySQL NODE02:3306 ssl JS > c.status()

"clusterName": "QACluster",
"defaultReplicaSet":
"name": "default",
"primary": "NODE03:3306",
"ssl": "REQUIRED",
"status": "OK_NO_TOLERANCE",
"statusText": "Cluster is NOT tolerant to any failures. 1 member is not active",
"topology":
"NODE02:3306":
"address": "NODE02:3306",
"mode": "R/O",
"readReplicas": ,
"role": "HA",
"status": "ONLINE"
,
"NODE03:3306":
"address": "NODE03:3306",
"mode": "R/W",
"readReplicas": ,
"role": "HA",
"status": "ONLINE"
,
"NODE01:3306":
"address": "NODE01:3306",
"mode": "R/O",
"readReplicas": ,
"role": "HA",
"status": "(MISSING)"


,
"groupInformationSourceMember": "mysql://clusterAdmin@NODE03:3306"



Errors logged in mysql error log:



2019-03-04T23:49:36.970839Z 3624 [Note] Slave SQL thread for channel 'group_replication_recovery' initialized, starting replication in log 'FIRST' at position 0, relay log './NODE01-relay-bin-group_replication_recovery.000001' position: 4
2019-03-04T23:49:36.985336Z 3623 [Note] Slave I/O thread for channel 'group_replication_recovery': connected to master 'mysql_innodb_cluster_r0429584112@NODE02:3306',replication started in log 'FIRST' at position 4
2019-03-04T23:49:36.988164Z 3623 [ERROR] Error reading packet from server for channel 'group_replication_recovery': The slave is connecting using CHANGE MASTER TO MASTER_AUTO_POSITION = 1, but the master has purged binary logs containing GTIDs that the slave requires. (server_errno=1236)
2019-03-04T23:49:36.988213Z 3623 [ERROR] Slave I/O for channel 'group_replication_recovery': Got fatal error 1236 from master when reading data from binary log: 'The slave is connecting using CHANGE MASTER TO MASTER_AUTO_POSITION = 1, but the master has purged binary logs containing GTIDs that the slave requires.', Error_code: 1236
2019-03-04T23:49:36.988226Z 3623 [Note] Slave I/O thread exiting for channel 'group_replication_recovery', read up to log 'FIRST', position 4
2019-03-04T23:49:36.988286Z 41 [Note] Plugin group_replication reported: 'Terminating existing group replication donor connection and purging the corresponding logs.'
2019-03-04T23:49:36.988358Z 3624 [Note] Error reading relay log event for channel 'group_replication_recovery': slave SQL thread was killed
2019-03-04T23:49:36.988435Z 3624 [Note] Slave SQL thread for channel 'group_replication_recovery' exiting, replication stopped in log 'FIRST' at position 0
2019-03-04T23:49:37.016864Z 41 [Note] 'CHANGE MASTER TO FOR CHANNEL 'group_replication_recovery' executed'. Previous state master_host='NODE02', master_port= 3306, master_log_file='', master_log_pos= 4, master_bind=''. New state master_host='<NULL>', master_port= 0, master_log_file='', master_log_pos= 4, master_bind=''.
2019-03-04T23:49:37.030769Z 41 [ERROR] Plugin group_replication reported: 'Maximum number of retries when trying to connect to a donor reached. Aborting group replication recovery.'
2019-03-04T23:49:37.030798Z 41 [Note] Plugin group_replication reported: 'Terminating existing group replication donor connection and purging the corresponding logs.'
2019-03-04T23:49:37.051169Z 41 [Note] 'CHANGE MASTER TO FOR CHANNEL 'group_replication_recovery' executed'. Previous state master_host='<NULL>', master_port= 0, master_log_file='', master_log_pos= 4, master_bind=''. New state master_host='<NULL>', master_port= 0, master_log_file='', master_log_pos= 4, master_bind=''.
2019-03-04T23:49:37.069184Z 41 [ERROR] Plugin group_replication reported: 'Fatal error during the Recovery process of Group Replication. The server will leave the group.'
2019-03-04T23:49:37.069304Z 41 [Note] Plugin group_replication reported: 'Going to wait for view modification'
2019-03-04T23:49:40.336938Z 0 [Note] Plugin group_replication reported: 'Group membership changed: This member has left the group.'









share|improve this question




























    0















    We have a MySQL innodb cluster in one of our environments. One of the nodes in the cluster was crashed. Though, we were able to bring the crashed node online we were unable to join it to the cluster.



    Can someone please help to recover/restore the node and join it to the cluster. We tried to use "dba.rebootClusterFromCompleteOutage()" but it didn't help.



    Configuration: MySQL 5.7.24 Community Edition, CentOS 7, standard three node innodb cluster



    Cluster Status:



    MySQL NODE02:3306 ssl JS > var c=dba.getCluster()
    MySQL NODE02:3306 ssl JS > c.status()

    "clusterName": "QACluster",
    "defaultReplicaSet":
    "name": "default",
    "primary": "NODE03:3306",
    "ssl": "REQUIRED",
    "status": "OK_NO_TOLERANCE",
    "statusText": "Cluster is NOT tolerant to any failures. 1 member is not active",
    "topology":
    "NODE02:3306":
    "address": "NODE02:3306",
    "mode": "R/O",
    "readReplicas": ,
    "role": "HA",
    "status": "ONLINE"
    ,
    "NODE03:3306":
    "address": "NODE03:3306",
    "mode": "R/W",
    "readReplicas": ,
    "role": "HA",
    "status": "ONLINE"
    ,
    "NODE01:3306":
    "address": "NODE01:3306",
    "mode": "R/O",
    "readReplicas": ,
    "role": "HA",
    "status": "(MISSING)"


    ,
    "groupInformationSourceMember": "mysql://clusterAdmin@NODE03:3306"



    Errors logged in mysql error log:



    2019-03-04T23:49:36.970839Z 3624 [Note] Slave SQL thread for channel 'group_replication_recovery' initialized, starting replication in log 'FIRST' at position 0, relay log './NODE01-relay-bin-group_replication_recovery.000001' position: 4
    2019-03-04T23:49:36.985336Z 3623 [Note] Slave I/O thread for channel 'group_replication_recovery': connected to master 'mysql_innodb_cluster_r0429584112@NODE02:3306',replication started in log 'FIRST' at position 4
    2019-03-04T23:49:36.988164Z 3623 [ERROR] Error reading packet from server for channel 'group_replication_recovery': The slave is connecting using CHANGE MASTER TO MASTER_AUTO_POSITION = 1, but the master has purged binary logs containing GTIDs that the slave requires. (server_errno=1236)
    2019-03-04T23:49:36.988213Z 3623 [ERROR] Slave I/O for channel 'group_replication_recovery': Got fatal error 1236 from master when reading data from binary log: 'The slave is connecting using CHANGE MASTER TO MASTER_AUTO_POSITION = 1, but the master has purged binary logs containing GTIDs that the slave requires.', Error_code: 1236
    2019-03-04T23:49:36.988226Z 3623 [Note] Slave I/O thread exiting for channel 'group_replication_recovery', read up to log 'FIRST', position 4
    2019-03-04T23:49:36.988286Z 41 [Note] Plugin group_replication reported: 'Terminating existing group replication donor connection and purging the corresponding logs.'
    2019-03-04T23:49:36.988358Z 3624 [Note] Error reading relay log event for channel 'group_replication_recovery': slave SQL thread was killed
    2019-03-04T23:49:36.988435Z 3624 [Note] Slave SQL thread for channel 'group_replication_recovery' exiting, replication stopped in log 'FIRST' at position 0
    2019-03-04T23:49:37.016864Z 41 [Note] 'CHANGE MASTER TO FOR CHANNEL 'group_replication_recovery' executed'. Previous state master_host='NODE02', master_port= 3306, master_log_file='', master_log_pos= 4, master_bind=''. New state master_host='<NULL>', master_port= 0, master_log_file='', master_log_pos= 4, master_bind=''.
    2019-03-04T23:49:37.030769Z 41 [ERROR] Plugin group_replication reported: 'Maximum number of retries when trying to connect to a donor reached. Aborting group replication recovery.'
    2019-03-04T23:49:37.030798Z 41 [Note] Plugin group_replication reported: 'Terminating existing group replication donor connection and purging the corresponding logs.'
    2019-03-04T23:49:37.051169Z 41 [Note] 'CHANGE MASTER TO FOR CHANNEL 'group_replication_recovery' executed'. Previous state master_host='<NULL>', master_port= 0, master_log_file='', master_log_pos= 4, master_bind=''. New state master_host='<NULL>', master_port= 0, master_log_file='', master_log_pos= 4, master_bind=''.
    2019-03-04T23:49:37.069184Z 41 [ERROR] Plugin group_replication reported: 'Fatal error during the Recovery process of Group Replication. The server will leave the group.'
    2019-03-04T23:49:37.069304Z 41 [Note] Plugin group_replication reported: 'Going to wait for view modification'
    2019-03-04T23:49:40.336938Z 0 [Note] Plugin group_replication reported: 'Group membership changed: This member has left the group.'









    share|improve this question


























      0












      0








      0


      1






      We have a MySQL innodb cluster in one of our environments. One of the nodes in the cluster was crashed. Though, we were able to bring the crashed node online we were unable to join it to the cluster.



      Can someone please help to recover/restore the node and join it to the cluster. We tried to use "dba.rebootClusterFromCompleteOutage()" but it didn't help.



      Configuration: MySQL 5.7.24 Community Edition, CentOS 7, standard three node innodb cluster



      Cluster Status:



      MySQL NODE02:3306 ssl JS > var c=dba.getCluster()
      MySQL NODE02:3306 ssl JS > c.status()

      "clusterName": "QACluster",
      "defaultReplicaSet":
      "name": "default",
      "primary": "NODE03:3306",
      "ssl": "REQUIRED",
      "status": "OK_NO_TOLERANCE",
      "statusText": "Cluster is NOT tolerant to any failures. 1 member is not active",
      "topology":
      "NODE02:3306":
      "address": "NODE02:3306",
      "mode": "R/O",
      "readReplicas": ,
      "role": "HA",
      "status": "ONLINE"
      ,
      "NODE03:3306":
      "address": "NODE03:3306",
      "mode": "R/W",
      "readReplicas": ,
      "role": "HA",
      "status": "ONLINE"
      ,
      "NODE01:3306":
      "address": "NODE01:3306",
      "mode": "R/O",
      "readReplicas": ,
      "role": "HA",
      "status": "(MISSING)"


      ,
      "groupInformationSourceMember": "mysql://clusterAdmin@NODE03:3306"



      Errors logged in mysql error log:



      2019-03-04T23:49:36.970839Z 3624 [Note] Slave SQL thread for channel 'group_replication_recovery' initialized, starting replication in log 'FIRST' at position 0, relay log './NODE01-relay-bin-group_replication_recovery.000001' position: 4
      2019-03-04T23:49:36.985336Z 3623 [Note] Slave I/O thread for channel 'group_replication_recovery': connected to master 'mysql_innodb_cluster_r0429584112@NODE02:3306',replication started in log 'FIRST' at position 4
      2019-03-04T23:49:36.988164Z 3623 [ERROR] Error reading packet from server for channel 'group_replication_recovery': The slave is connecting using CHANGE MASTER TO MASTER_AUTO_POSITION = 1, but the master has purged binary logs containing GTIDs that the slave requires. (server_errno=1236)
      2019-03-04T23:49:36.988213Z 3623 [ERROR] Slave I/O for channel 'group_replication_recovery': Got fatal error 1236 from master when reading data from binary log: 'The slave is connecting using CHANGE MASTER TO MASTER_AUTO_POSITION = 1, but the master has purged binary logs containing GTIDs that the slave requires.', Error_code: 1236
      2019-03-04T23:49:36.988226Z 3623 [Note] Slave I/O thread exiting for channel 'group_replication_recovery', read up to log 'FIRST', position 4
      2019-03-04T23:49:36.988286Z 41 [Note] Plugin group_replication reported: 'Terminating existing group replication donor connection and purging the corresponding logs.'
      2019-03-04T23:49:36.988358Z 3624 [Note] Error reading relay log event for channel 'group_replication_recovery': slave SQL thread was killed
      2019-03-04T23:49:36.988435Z 3624 [Note] Slave SQL thread for channel 'group_replication_recovery' exiting, replication stopped in log 'FIRST' at position 0
      2019-03-04T23:49:37.016864Z 41 [Note] 'CHANGE MASTER TO FOR CHANNEL 'group_replication_recovery' executed'. Previous state master_host='NODE02', master_port= 3306, master_log_file='', master_log_pos= 4, master_bind=''. New state master_host='<NULL>', master_port= 0, master_log_file='', master_log_pos= 4, master_bind=''.
      2019-03-04T23:49:37.030769Z 41 [ERROR] Plugin group_replication reported: 'Maximum number of retries when trying to connect to a donor reached. Aborting group replication recovery.'
      2019-03-04T23:49:37.030798Z 41 [Note] Plugin group_replication reported: 'Terminating existing group replication donor connection and purging the corresponding logs.'
      2019-03-04T23:49:37.051169Z 41 [Note] 'CHANGE MASTER TO FOR CHANNEL 'group_replication_recovery' executed'. Previous state master_host='<NULL>', master_port= 0, master_log_file='', master_log_pos= 4, master_bind=''. New state master_host='<NULL>', master_port= 0, master_log_file='', master_log_pos= 4, master_bind=''.
      2019-03-04T23:49:37.069184Z 41 [ERROR] Plugin group_replication reported: 'Fatal error during the Recovery process of Group Replication. The server will leave the group.'
      2019-03-04T23:49:37.069304Z 41 [Note] Plugin group_replication reported: 'Going to wait for view modification'
      2019-03-04T23:49:40.336938Z 0 [Note] Plugin group_replication reported: 'Group membership changed: This member has left the group.'









      share|improve this question
















      We have a MySQL innodb cluster in one of our environments. One of the nodes in the cluster was crashed. Though, we were able to bring the crashed node online we were unable to join it to the cluster.



      Can someone please help to recover/restore the node and join it to the cluster. We tried to use "dba.rebootClusterFromCompleteOutage()" but it didn't help.



      Configuration: MySQL 5.7.24 Community Edition, CentOS 7, standard three node innodb cluster



      Cluster Status:



      MySQL NODE02:3306 ssl JS > var c=dba.getCluster()
      MySQL NODE02:3306 ssl JS > c.status()

      "clusterName": "QACluster",
      "defaultReplicaSet":
      "name": "default",
      "primary": "NODE03:3306",
      "ssl": "REQUIRED",
      "status": "OK_NO_TOLERANCE",
      "statusText": "Cluster is NOT tolerant to any failures. 1 member is not active",
      "topology":
      "NODE02:3306":
      "address": "NODE02:3306",
      "mode": "R/O",
      "readReplicas": ,
      "role": "HA",
      "status": "ONLINE"
      ,
      "NODE03:3306":
      "address": "NODE03:3306",
      "mode": "R/W",
      "readReplicas": ,
      "role": "HA",
      "status": "ONLINE"
      ,
      "NODE01:3306":
      "address": "NODE01:3306",
      "mode": "R/O",
      "readReplicas": ,
      "role": "HA",
      "status": "(MISSING)"


      ,
      "groupInformationSourceMember": "mysql://clusterAdmin@NODE03:3306"



      Errors logged in mysql error log:



      2019-03-04T23:49:36.970839Z 3624 [Note] Slave SQL thread for channel 'group_replication_recovery' initialized, starting replication in log 'FIRST' at position 0, relay log './NODE01-relay-bin-group_replication_recovery.000001' position: 4
      2019-03-04T23:49:36.985336Z 3623 [Note] Slave I/O thread for channel 'group_replication_recovery': connected to master 'mysql_innodb_cluster_r0429584112@NODE02:3306',replication started in log 'FIRST' at position 4
      2019-03-04T23:49:36.988164Z 3623 [ERROR] Error reading packet from server for channel 'group_replication_recovery': The slave is connecting using CHANGE MASTER TO MASTER_AUTO_POSITION = 1, but the master has purged binary logs containing GTIDs that the slave requires. (server_errno=1236)
      2019-03-04T23:49:36.988213Z 3623 [ERROR] Slave I/O for channel 'group_replication_recovery': Got fatal error 1236 from master when reading data from binary log: 'The slave is connecting using CHANGE MASTER TO MASTER_AUTO_POSITION = 1, but the master has purged binary logs containing GTIDs that the slave requires.', Error_code: 1236
      2019-03-04T23:49:36.988226Z 3623 [Note] Slave I/O thread exiting for channel 'group_replication_recovery', read up to log 'FIRST', position 4
      2019-03-04T23:49:36.988286Z 41 [Note] Plugin group_replication reported: 'Terminating existing group replication donor connection and purging the corresponding logs.'
      2019-03-04T23:49:36.988358Z 3624 [Note] Error reading relay log event for channel 'group_replication_recovery': slave SQL thread was killed
      2019-03-04T23:49:36.988435Z 3624 [Note] Slave SQL thread for channel 'group_replication_recovery' exiting, replication stopped in log 'FIRST' at position 0
      2019-03-04T23:49:37.016864Z 41 [Note] 'CHANGE MASTER TO FOR CHANNEL 'group_replication_recovery' executed'. Previous state master_host='NODE02', master_port= 3306, master_log_file='', master_log_pos= 4, master_bind=''. New state master_host='<NULL>', master_port= 0, master_log_file='', master_log_pos= 4, master_bind=''.
      2019-03-04T23:49:37.030769Z 41 [ERROR] Plugin group_replication reported: 'Maximum number of retries when trying to connect to a donor reached. Aborting group replication recovery.'
      2019-03-04T23:49:37.030798Z 41 [Note] Plugin group_replication reported: 'Terminating existing group replication donor connection and purging the corresponding logs.'
      2019-03-04T23:49:37.051169Z 41 [Note] 'CHANGE MASTER TO FOR CHANNEL 'group_replication_recovery' executed'. Previous state master_host='<NULL>', master_port= 0, master_log_file='', master_log_pos= 4, master_bind=''. New state master_host='<NULL>', master_port= 0, master_log_file='', master_log_pos= 4, master_bind=''.
      2019-03-04T23:49:37.069184Z 41 [ERROR] Plugin group_replication reported: 'Fatal error during the Recovery process of Group Replication. The server will leave the group.'
      2019-03-04T23:49:37.069304Z 41 [Note] Plugin group_replication reported: 'Going to wait for view modification'
      2019-03-04T23:49:40.336938Z 0 [Note] Plugin group_replication reported: 'Group membership changed: This member has left the group.'






      mysql innodb mysql-cluster mysql-group-replication






      share|improve this question















      share|improve this question













      share|improve this question




      share|improve this question








      edited Mar 15 at 16:31







      sqlcheckpoint

















      asked Mar 7 at 4:45









      sqlcheckpointsqlcheckpoint

      351519




      351519






















          2 Answers
          2






          active

          oldest

          votes


















          1















          2019-03-04T23:49:36.988213Z 3623 [ERROR] Slave I/O for channel
          'group_replication_recovery': Got fatal error 1236 from master when
          reading data from binary log: 'The slave is connecting using CHANGE
          MASTER TO MASTER_AUTO_POSITION = 1, but the master has purged binary
          logs containing GTIDs that the slave requires.', Error_code: 1236




          Lets put it on simple terms:
          What this means is that you had 2 servers (S1,S2), one crashed (S2). While S2 was crashed you did some operations on S1 (transactions T1,T2) but at some point in time the binlogs on S1 were purged and they included info for T1.



          Bottom line, S2 can't join the group because it knows it is missing T1, but there is no binlog file with that information available to transmit that info during recovery.



          Sorry but there is no magic command here, that is just bad luck.
          For now your best option is to provision S2 with data from S1 using some tool like dump or another in that group.



          For the future you should recheck your binlog purge policies and make if possible the recovery of crashed machines a bit more fast.






          share|improve this answer























          • Thank you. Any specific instructions to follow in restoring or recovering the failed node from the master backup?

            – sqlcheckpoint
            Mar 7 at 17:08











          • Not an expert on DBA operations but if you search restore slave from backup you will find various answers. You need a tool to take the data from the running member and then replay that on the crashed member. Just a note that if your tool does not handle GTIDs you need to set GTID_purged on the recovering node after you restore the data.

            – Pedro Gomes
            Mar 7 at 18:23











          • sure, thank you. I'm working on it to fix. If I'm able to I will publish possible solution.

            – sqlcheckpoint
            Mar 8 at 1:25


















          0














          I did the following to restore the failed node from backup and able to recover the cluster state.



          1)Below is the status of the cluster when one of the nodes failed (NODE01).



           MySQL NODE02:3306 ssl JS > var c=dba.getCluster()
          MySQL NODE02:3306 ssl JS > c.status()

          "clusterName": "QACluster",
          "defaultReplicaSet":
          "name": "default",
          "primary": "NODE03:3306",
          "ssl": "REQUIRED",
          "status": "OK_NO_TOLERANCE",
          "statusText": "Cluster is NOT tolerant to any failures. 1 member is not active",
          "topology":
          "NODE02:3306":
          "address": "NODE02:3306",
          "mode": "R/O",
          "readReplicas": ,
          "role": "HA",
          "status": "ONLINE"
          ,
          "NODE03:3306":
          "address": "NODE03:3306",
          "mode": "R/W",
          "readReplicas": ,
          "role": "HA",
          "status": "ONLINE"
          ,
          "NODE01:3306":
          "address": "NODE01:3306",
          "mode": "R/O",
          "readReplicas": ,
          "role": "HA",
          "status": "(MISSING)"


          ,
          "groupInformationSourceMember": "mysql://clusterAdmin@NODE03:3306"



          2) Take mysqldump from the master node (healthy node) using the following command.



          [root@NODE03 db_backup]# mysqldump --all-databases --add-drop-database --single-transaction --triggers --routines --port=mysql_port --user=root -p > /db_backup/mysql_dump_03062019.sql
          Enter password:
          Warning: A partial dump from a server that has GTIDs will by default include the GTIDs of all transactions, even those that changed suppressed parts of the database. If you don't want to restore GTIDs, pass --set-gtid-purged=OFF. To make a complete dump, pass --all-databases --triggers --routines --events.


          3) Execute below step to remove the failed node from the cluster.



           MySQL NODE03:3306 ssl JS > var c=dba.getCluster()
          MySQL NODE03:3306 ssl JS > c.rescan()
          Rescanning the cluster...

          Result of the rescanning operation:

          "defaultReplicaSet":
          "name": "default",
          "newlyDiscoveredInstances": [],
          "unavailableInstances": [

          "host": "NODE01:3306",
          "label": "NODE01:3306",
          "member_id": "e2aa897d-1828-11e9-85b3-00505692188c"

          ]



          The instance 'NODE01:3306' is no longer part of the HA setup. It is either offline or left the HA group.
          You can try to add it to the cluster again with the cluster.rejoinInstance('NODE01:3306') command or you can remove it from the cluster configuration.
          Would you like to remove it from the cluster metadata? [Y/n]: Y
          Removing instance from the cluster metadata...

          The instance 'NODE01:3306' was successfully removed from the cluster metadata.

          MySQL NODE03:3306 ssl JS > c.status()

          "clusterName": "QACluster",
          "defaultReplicaSet":
          "name": "default",
          "primary": "NODE03:3306",
          "ssl": "REQUIRED",
          "status": "OK_NO_TOLERANCE",
          "statusText": "Cluster is NOT tolerant to any failures.",
          "topology":
          "NODE02:3306":
          "address": "NODE02:3306",
          "mode": "R/O",
          "readReplicas": ,
          "role": "HA",
          "status": "ONLINE"
          ,
          "NODE03:3306":
          "address": "NODE03:3306",
          "mode": "R/W",
          "readReplicas": ,
          "role": "HA",
          "status": "ONLINE"


          ,
          "groupInformationSourceMember": "mysql://clusterAdmin@NODE03:3306"



          4) Stop group replication if it is still running on failed node.



          mysql> STOP GROUP_REPLICATION;
          Query OK, 0 rows affected (1.01 sec)


          5) Reset "gtid_executed" on the failed node.



          mysql> show global variables like 'GTID_EXECUTED';
          +---------------+--------------------------------------------------------------------------------------------+
          | Variable_name | Value |
          +---------------+--------------------------------------------------------------------------------------------+
          | gtid_executed | 01f27b9c-182a-11e9-a199-00505692188c:1-14134172,
          e2aa897d-1828-11e9-85b3-00505692188c:1-12 |
          +---------------+--------------------------------------------------------------------------------------------+
          1 row in set (0.01 sec)

          mysql> reset master;
          Query OK, 0 rows affected (0.02 sec)

          mysql> reset slave;
          Query OK, 0 rows affected (0.02 sec)

          mysql> show global variables like 'GTID_EXECUTED';
          +---------------+-------+
          | Variable_name | Value |
          +---------------+-------+
          | gtid_executed | |
          +---------------+-------+
          1 row in set (0.00 sec)


          6) Disable "super_readonly_flag" on the failed node.



          mysql> SELECT @@global.read_only, @@global.super_read_only;
          +--------------------+--------------------------+
          | @@global.read_only | @@global.super_read_only |
          +--------------------+--------------------------+
          | 1 | 1 |
          +--------------------+--------------------------+
          1 row in set (0.00 sec)

          mysql> SET GLOBAL super_read_only = 0;
          Query OK, 0 rows affected (0.00 sec)

          mysql> SELECT @@global.read_only, @@global.super_read_only;
          +--------------------+--------------------------+
          | @@global.read_only | @@global.super_read_only |
          +--------------------+--------------------------+
          | 1 | 0 |
          +--------------------+--------------------------+
          1 row in set (0.00 sec)


          7) Restore the mysqldump from master on to the failed node.



          [root@E2LXQA1ALFDB01 db_backup]# mysql -uroot -p < mysql_dump_03062019.sql


          8) Once restore is completed enable "super_readonly_flag" on the failed node.



          mysql> SELECT @@global.read_only, @@global.super_read_only;
          +--------------------+--------------------------+
          | @@global.read_only | @@global.super_read_only |
          +--------------------+--------------------------+
          | 1 | 0 |
          +--------------------+--------------------------+
          1 row in set (0.00 sec)

          mysql> SET GLOBAL super_read_only = 1;
          Query OK, 0 rows affected (0.00 sec)


          mysql> SELECT @@global.read_only, @@global.super_read_only;
          +--------------------+--------------------------+
          | @@global.read_only | @@global.super_read_only |
          +--------------------+--------------------------+
          | 1 | 1 |
          +--------------------+--------------------------+
          1 row in set (0.00 sec)


          9) Finally add the failed node back to the innodb cluster.



          MySQL NODE03:3306 ssl JS > c.addInstance('clusterAdmin@NODE01:3306');
          A new instance will be added to the InnoDB cluster. Depending on the amount of
          data on the cluster this might take from a few seconds to several hours.

          Adding instance to the cluster ...

          Please provide the password for 'clusterAdmin@NODE01:3306': *******************
          Save password for 'clusterAdmin@NODE01:3306'? [Y]es/[N]o/Ne[v]er (default No):
          Validating instance at NODE01:3306...

          This instance reports its own address as NODE01
          WARNING: The following tables do not have a Primary Key or equivalent column:
          ephesoft.dlf, report.correction_type, report.field_details_ag, report_archive.correction_type, report_archive.field_details_ag, report_archive.global_data_ag

          Group Replication requires tables to use InnoDB and have a PRIMARY KEY or PRIMARY KEY Equivalent (non-null unique key). Tables that do not follow these requirements will be readable but not updateable when used with Group Replication. If your applications make updates (INSERT, UPDATE or DELETE) to these tables, ensure they use the InnoDB storage engine and have a PRIMARY KEY or PRIMARY KEY Equivalent.

          Instance configuration is suitable.
          WARNING: On instance 'NODE01:3306' membership change cannot be persisted since MySQL version 5.7.24 does not support the SET PERSIST command (MySQL version >= 8.0.11 required). Please use the .configureLocalInstance command locally to persist the changes.
          WARNING: On instance 'NODE02:3306' membership change cannot be persisted since MySQL version 5.7.24 does not support the SET PERSIST command (MySQL version >= 8.0.11 required). Please use the .configureLocalInstance command locally to persist the changes.
          WARNING: On instance 'NODE03:3306' membership change cannot be persisted since MySQL version 5.7.24 does not support the SET PERSIST command (MySQL version >= 8.0.11 required). Please use the .configureLocalInstance command locally to persist the changes.
          The instance 'clusterAdmin@NODE01:3306' was successfully added to the cluster.


          MySQL NODE03:3306 ssl JS > c.status()

          "clusterName": "QACluster",
          "defaultReplicaSet":
          "name": "default",
          "primary": "NODE03:3306",
          "ssl": "REQUIRED",
          "status": "OK",
          "statusText": "Cluster is ONLINE and can tolerate up to ONE failure.",
          "topology":
          "NODE01:3306":
          "address": "NODE01:3306",
          "mode": "R/O",
          "readReplicas": ,
          "role": "HA",
          "status": "ONLINE"
          ,
          "NODE02:3306":
          "address": "NODE02:3306",
          "mode": "R/O",
          "readReplicas": ,
          "role": "HA",
          "status": "ONLINE"
          ,
          "NODE03:3306":
          "address": "NODE03:3306",
          "mode": "R/W",
          "readReplicas": ,
          "role": "HA",
          "status": "ONLINE"


          ,
          "groupInformationSourceMember": "mysql://clusterAdmin@NODE03:3306"






          share|improve this answer
























            Your Answer






            StackExchange.ifUsing("editor", function ()
            StackExchange.using("externalEditor", function ()
            StackExchange.using("snippets", function ()
            StackExchange.snippets.init();
            );
            );
            , "code-snippets");

            StackExchange.ready(function()
            var channelOptions =
            tags: "".split(" "),
            id: "1"
            ;
            initTagRenderer("".split(" "), "".split(" "), channelOptions);

            StackExchange.using("externalEditor", function()
            // Have to fire editor after snippets, if snippets enabled
            if (StackExchange.settings.snippets.snippetsEnabled)
            StackExchange.using("snippets", function()
            createEditor();
            );

            else
            createEditor();

            );

            function createEditor()
            StackExchange.prepareEditor(
            heartbeatType: 'answer',
            autoActivateHeartbeat: false,
            convertImagesToLinks: true,
            noModals: true,
            showLowRepImageUploadWarning: true,
            reputationToPostImages: 10,
            bindNavPrevention: true,
            postfix: "",
            imageUploader:
            brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
            contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
            allowUrls: true
            ,
            onDemand: true,
            discardSelector: ".discard-answer"
            ,immediatelyShowMarkdownHelp:true
            );



            );













            draft saved

            draft discarded


















            StackExchange.ready(
            function ()
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f55036255%2fa-node-in-mysql-5-7-innodb-cluster-is-crashed-and-unable-to-re-join-the-crashed%23new-answer', 'question_page');

            );

            Post as a guest















            Required, but never shown

























            2 Answers
            2






            active

            oldest

            votes








            2 Answers
            2






            active

            oldest

            votes









            active

            oldest

            votes






            active

            oldest

            votes









            1















            2019-03-04T23:49:36.988213Z 3623 [ERROR] Slave I/O for channel
            'group_replication_recovery': Got fatal error 1236 from master when
            reading data from binary log: 'The slave is connecting using CHANGE
            MASTER TO MASTER_AUTO_POSITION = 1, but the master has purged binary
            logs containing GTIDs that the slave requires.', Error_code: 1236




            Lets put it on simple terms:
            What this means is that you had 2 servers (S1,S2), one crashed (S2). While S2 was crashed you did some operations on S1 (transactions T1,T2) but at some point in time the binlogs on S1 were purged and they included info for T1.



            Bottom line, S2 can't join the group because it knows it is missing T1, but there is no binlog file with that information available to transmit that info during recovery.



            Sorry but there is no magic command here, that is just bad luck.
            For now your best option is to provision S2 with data from S1 using some tool like dump or another in that group.



            For the future you should recheck your binlog purge policies and make if possible the recovery of crashed machines a bit more fast.






            share|improve this answer























            • Thank you. Any specific instructions to follow in restoring or recovering the failed node from the master backup?

              – sqlcheckpoint
              Mar 7 at 17:08











            • Not an expert on DBA operations but if you search restore slave from backup you will find various answers. You need a tool to take the data from the running member and then replay that on the crashed member. Just a note that if your tool does not handle GTIDs you need to set GTID_purged on the recovering node after you restore the data.

              – Pedro Gomes
              Mar 7 at 18:23











            • sure, thank you. I'm working on it to fix. If I'm able to I will publish possible solution.

              – sqlcheckpoint
              Mar 8 at 1:25















            1















            2019-03-04T23:49:36.988213Z 3623 [ERROR] Slave I/O for channel
            'group_replication_recovery': Got fatal error 1236 from master when
            reading data from binary log: 'The slave is connecting using CHANGE
            MASTER TO MASTER_AUTO_POSITION = 1, but the master has purged binary
            logs containing GTIDs that the slave requires.', Error_code: 1236




            Lets put it on simple terms:
            What this means is that you had 2 servers (S1,S2), one crashed (S2). While S2 was crashed you did some operations on S1 (transactions T1,T2) but at some point in time the binlogs on S1 were purged and they included info for T1.



            Bottom line, S2 can't join the group because it knows it is missing T1, but there is no binlog file with that information available to transmit that info during recovery.



            Sorry but there is no magic command here, that is just bad luck.
            For now your best option is to provision S2 with data from S1 using some tool like dump or another in that group.



            For the future you should recheck your binlog purge policies and make if possible the recovery of crashed machines a bit more fast.






            share|improve this answer























            • Thank you. Any specific instructions to follow in restoring or recovering the failed node from the master backup?

              – sqlcheckpoint
              Mar 7 at 17:08











            • Not an expert on DBA operations but if you search restore slave from backup you will find various answers. You need a tool to take the data from the running member and then replay that on the crashed member. Just a note that if your tool does not handle GTIDs you need to set GTID_purged on the recovering node after you restore the data.

              – Pedro Gomes
              Mar 7 at 18:23











            • sure, thank you. I'm working on it to fix. If I'm able to I will publish possible solution.

              – sqlcheckpoint
              Mar 8 at 1:25













            1












            1








            1








            2019-03-04T23:49:36.988213Z 3623 [ERROR] Slave I/O for channel
            'group_replication_recovery': Got fatal error 1236 from master when
            reading data from binary log: 'The slave is connecting using CHANGE
            MASTER TO MASTER_AUTO_POSITION = 1, but the master has purged binary
            logs containing GTIDs that the slave requires.', Error_code: 1236




            Lets put it on simple terms:
            What this means is that you had 2 servers (S1,S2), one crashed (S2). While S2 was crashed you did some operations on S1 (transactions T1,T2) but at some point in time the binlogs on S1 were purged and they included info for T1.



            Bottom line, S2 can't join the group because it knows it is missing T1, but there is no binlog file with that information available to transmit that info during recovery.



            Sorry but there is no magic command here, that is just bad luck.
            For now your best option is to provision S2 with data from S1 using some tool like dump or another in that group.



            For the future you should recheck your binlog purge policies and make if possible the recovery of crashed machines a bit more fast.






            share|improve this answer














            2019-03-04T23:49:36.988213Z 3623 [ERROR] Slave I/O for channel
            'group_replication_recovery': Got fatal error 1236 from master when
            reading data from binary log: 'The slave is connecting using CHANGE
            MASTER TO MASTER_AUTO_POSITION = 1, but the master has purged binary
            logs containing GTIDs that the slave requires.', Error_code: 1236




            Lets put it on simple terms:
            What this means is that you had 2 servers (S1,S2), one crashed (S2). While S2 was crashed you did some operations on S1 (transactions T1,T2) but at some point in time the binlogs on S1 were purged and they included info for T1.



            Bottom line, S2 can't join the group because it knows it is missing T1, but there is no binlog file with that information available to transmit that info during recovery.



            Sorry but there is no magic command here, that is just bad luck.
            For now your best option is to provision S2 with data from S1 using some tool like dump or another in that group.



            For the future you should recheck your binlog purge policies and make if possible the recovery of crashed machines a bit more fast.







            share|improve this answer












            share|improve this answer



            share|improve this answer










            answered Mar 7 at 11:22









            Pedro GomesPedro Gomes

            3614




            3614












            • Thank you. Any specific instructions to follow in restoring or recovering the failed node from the master backup?

              – sqlcheckpoint
              Mar 7 at 17:08











            • Not an expert on DBA operations but if you search restore slave from backup you will find various answers. You need a tool to take the data from the running member and then replay that on the crashed member. Just a note that if your tool does not handle GTIDs you need to set GTID_purged on the recovering node after you restore the data.

              – Pedro Gomes
              Mar 7 at 18:23











            • sure, thank you. I'm working on it to fix. If I'm able to I will publish possible solution.

              – sqlcheckpoint
              Mar 8 at 1:25

















            • Thank you. Any specific instructions to follow in restoring or recovering the failed node from the master backup?

              – sqlcheckpoint
              Mar 7 at 17:08











            • Not an expert on DBA operations but if you search restore slave from backup you will find various answers. You need a tool to take the data from the running member and then replay that on the crashed member. Just a note that if your tool does not handle GTIDs you need to set GTID_purged on the recovering node after you restore the data.

              – Pedro Gomes
              Mar 7 at 18:23











            • sure, thank you. I'm working on it to fix. If I'm able to I will publish possible solution.

              – sqlcheckpoint
              Mar 8 at 1:25
















            Thank you. Any specific instructions to follow in restoring or recovering the failed node from the master backup?

            – sqlcheckpoint
            Mar 7 at 17:08





            Thank you. Any specific instructions to follow in restoring or recovering the failed node from the master backup?

            – sqlcheckpoint
            Mar 7 at 17:08













            Not an expert on DBA operations but if you search restore slave from backup you will find various answers. You need a tool to take the data from the running member and then replay that on the crashed member. Just a note that if your tool does not handle GTIDs you need to set GTID_purged on the recovering node after you restore the data.

            – Pedro Gomes
            Mar 7 at 18:23





            Not an expert on DBA operations but if you search restore slave from backup you will find various answers. You need a tool to take the data from the running member and then replay that on the crashed member. Just a note that if your tool does not handle GTIDs you need to set GTID_purged on the recovering node after you restore the data.

            – Pedro Gomes
            Mar 7 at 18:23













            sure, thank you. I'm working on it to fix. If I'm able to I will publish possible solution.

            – sqlcheckpoint
            Mar 8 at 1:25





            sure, thank you. I'm working on it to fix. If I'm able to I will publish possible solution.

            – sqlcheckpoint
            Mar 8 at 1:25













            0














            I did the following to restore the failed node from backup and able to recover the cluster state.



            1)Below is the status of the cluster when one of the nodes failed (NODE01).



             MySQL NODE02:3306 ssl JS > var c=dba.getCluster()
            MySQL NODE02:3306 ssl JS > c.status()

            "clusterName": "QACluster",
            "defaultReplicaSet":
            "name": "default",
            "primary": "NODE03:3306",
            "ssl": "REQUIRED",
            "status": "OK_NO_TOLERANCE",
            "statusText": "Cluster is NOT tolerant to any failures. 1 member is not active",
            "topology":
            "NODE02:3306":
            "address": "NODE02:3306",
            "mode": "R/O",
            "readReplicas": ,
            "role": "HA",
            "status": "ONLINE"
            ,
            "NODE03:3306":
            "address": "NODE03:3306",
            "mode": "R/W",
            "readReplicas": ,
            "role": "HA",
            "status": "ONLINE"
            ,
            "NODE01:3306":
            "address": "NODE01:3306",
            "mode": "R/O",
            "readReplicas": ,
            "role": "HA",
            "status": "(MISSING)"


            ,
            "groupInformationSourceMember": "mysql://clusterAdmin@NODE03:3306"



            2) Take mysqldump from the master node (healthy node) using the following command.



            [root@NODE03 db_backup]# mysqldump --all-databases --add-drop-database --single-transaction --triggers --routines --port=mysql_port --user=root -p > /db_backup/mysql_dump_03062019.sql
            Enter password:
            Warning: A partial dump from a server that has GTIDs will by default include the GTIDs of all transactions, even those that changed suppressed parts of the database. If you don't want to restore GTIDs, pass --set-gtid-purged=OFF. To make a complete dump, pass --all-databases --triggers --routines --events.


            3) Execute below step to remove the failed node from the cluster.



             MySQL NODE03:3306 ssl JS > var c=dba.getCluster()
            MySQL NODE03:3306 ssl JS > c.rescan()
            Rescanning the cluster...

            Result of the rescanning operation:

            "defaultReplicaSet":
            "name": "default",
            "newlyDiscoveredInstances": [],
            "unavailableInstances": [

            "host": "NODE01:3306",
            "label": "NODE01:3306",
            "member_id": "e2aa897d-1828-11e9-85b3-00505692188c"

            ]



            The instance 'NODE01:3306' is no longer part of the HA setup. It is either offline or left the HA group.
            You can try to add it to the cluster again with the cluster.rejoinInstance('NODE01:3306') command or you can remove it from the cluster configuration.
            Would you like to remove it from the cluster metadata? [Y/n]: Y
            Removing instance from the cluster metadata...

            The instance 'NODE01:3306' was successfully removed from the cluster metadata.

            MySQL NODE03:3306 ssl JS > c.status()

            "clusterName": "QACluster",
            "defaultReplicaSet":
            "name": "default",
            "primary": "NODE03:3306",
            "ssl": "REQUIRED",
            "status": "OK_NO_TOLERANCE",
            "statusText": "Cluster is NOT tolerant to any failures.",
            "topology":
            "NODE02:3306":
            "address": "NODE02:3306",
            "mode": "R/O",
            "readReplicas": ,
            "role": "HA",
            "status": "ONLINE"
            ,
            "NODE03:3306":
            "address": "NODE03:3306",
            "mode": "R/W",
            "readReplicas": ,
            "role": "HA",
            "status": "ONLINE"


            ,
            "groupInformationSourceMember": "mysql://clusterAdmin@NODE03:3306"



            4) Stop group replication if it is still running on failed node.



            mysql> STOP GROUP_REPLICATION;
            Query OK, 0 rows affected (1.01 sec)


            5) Reset "gtid_executed" on the failed node.



            mysql> show global variables like 'GTID_EXECUTED';
            +---------------+--------------------------------------------------------------------------------------------+
            | Variable_name | Value |
            +---------------+--------------------------------------------------------------------------------------------+
            | gtid_executed | 01f27b9c-182a-11e9-a199-00505692188c:1-14134172,
            e2aa897d-1828-11e9-85b3-00505692188c:1-12 |
            +---------------+--------------------------------------------------------------------------------------------+
            1 row in set (0.01 sec)

            mysql> reset master;
            Query OK, 0 rows affected (0.02 sec)

            mysql> reset slave;
            Query OK, 0 rows affected (0.02 sec)

            mysql> show global variables like 'GTID_EXECUTED';
            +---------------+-------+
            | Variable_name | Value |
            +---------------+-------+
            | gtid_executed | |
            +---------------+-------+
            1 row in set (0.00 sec)


            6) Disable "super_readonly_flag" on the failed node.



            mysql> SELECT @@global.read_only, @@global.super_read_only;
            +--------------------+--------------------------+
            | @@global.read_only | @@global.super_read_only |
            +--------------------+--------------------------+
            | 1 | 1 |
            +--------------------+--------------------------+
            1 row in set (0.00 sec)

            mysql> SET GLOBAL super_read_only = 0;
            Query OK, 0 rows affected (0.00 sec)

            mysql> SELECT @@global.read_only, @@global.super_read_only;
            +--------------------+--------------------------+
            | @@global.read_only | @@global.super_read_only |
            +--------------------+--------------------------+
            | 1 | 0 |
            +--------------------+--------------------------+
            1 row in set (0.00 sec)


            7) Restore the mysqldump from master on to the failed node.



            [root@E2LXQA1ALFDB01 db_backup]# mysql -uroot -p < mysql_dump_03062019.sql


            8) Once restore is completed enable "super_readonly_flag" on the failed node.



            mysql> SELECT @@global.read_only, @@global.super_read_only;
            +--------------------+--------------------------+
            | @@global.read_only | @@global.super_read_only |
            +--------------------+--------------------------+
            | 1 | 0 |
            +--------------------+--------------------------+
            1 row in set (0.00 sec)

            mysql> SET GLOBAL super_read_only = 1;
            Query OK, 0 rows affected (0.00 sec)


            mysql> SELECT @@global.read_only, @@global.super_read_only;
            +--------------------+--------------------------+
            | @@global.read_only | @@global.super_read_only |
            +--------------------+--------------------------+
            | 1 | 1 |
            +--------------------+--------------------------+
            1 row in set (0.00 sec)


            9) Finally add the failed node back to the innodb cluster.



            MySQL NODE03:3306 ssl JS > c.addInstance('clusterAdmin@NODE01:3306');
            A new instance will be added to the InnoDB cluster. Depending on the amount of
            data on the cluster this might take from a few seconds to several hours.

            Adding instance to the cluster ...

            Please provide the password for 'clusterAdmin@NODE01:3306': *******************
            Save password for 'clusterAdmin@NODE01:3306'? [Y]es/[N]o/Ne[v]er (default No):
            Validating instance at NODE01:3306...

            This instance reports its own address as NODE01
            WARNING: The following tables do not have a Primary Key or equivalent column:
            ephesoft.dlf, report.correction_type, report.field_details_ag, report_archive.correction_type, report_archive.field_details_ag, report_archive.global_data_ag

            Group Replication requires tables to use InnoDB and have a PRIMARY KEY or PRIMARY KEY Equivalent (non-null unique key). Tables that do not follow these requirements will be readable but not updateable when used with Group Replication. If your applications make updates (INSERT, UPDATE or DELETE) to these tables, ensure they use the InnoDB storage engine and have a PRIMARY KEY or PRIMARY KEY Equivalent.

            Instance configuration is suitable.
            WARNING: On instance 'NODE01:3306' membership change cannot be persisted since MySQL version 5.7.24 does not support the SET PERSIST command (MySQL version >= 8.0.11 required). Please use the .configureLocalInstance command locally to persist the changes.
            WARNING: On instance 'NODE02:3306' membership change cannot be persisted since MySQL version 5.7.24 does not support the SET PERSIST command (MySQL version >= 8.0.11 required). Please use the .configureLocalInstance command locally to persist the changes.
            WARNING: On instance 'NODE03:3306' membership change cannot be persisted since MySQL version 5.7.24 does not support the SET PERSIST command (MySQL version >= 8.0.11 required). Please use the .configureLocalInstance command locally to persist the changes.
            The instance 'clusterAdmin@NODE01:3306' was successfully added to the cluster.


            MySQL NODE03:3306 ssl JS > c.status()

            "clusterName": "QACluster",
            "defaultReplicaSet":
            "name": "default",
            "primary": "NODE03:3306",
            "ssl": "REQUIRED",
            "status": "OK",
            "statusText": "Cluster is ONLINE and can tolerate up to ONE failure.",
            "topology":
            "NODE01:3306":
            "address": "NODE01:3306",
            "mode": "R/O",
            "readReplicas": ,
            "role": "HA",
            "status": "ONLINE"
            ,
            "NODE02:3306":
            "address": "NODE02:3306",
            "mode": "R/O",
            "readReplicas": ,
            "role": "HA",
            "status": "ONLINE"
            ,
            "NODE03:3306":
            "address": "NODE03:3306",
            "mode": "R/W",
            "readReplicas": ,
            "role": "HA",
            "status": "ONLINE"


            ,
            "groupInformationSourceMember": "mysql://clusterAdmin@NODE03:3306"






            share|improve this answer





























              0














              I did the following to restore the failed node from backup and able to recover the cluster state.



              1)Below is the status of the cluster when one of the nodes failed (NODE01).



               MySQL NODE02:3306 ssl JS > var c=dba.getCluster()
              MySQL NODE02:3306 ssl JS > c.status()

              "clusterName": "QACluster",
              "defaultReplicaSet":
              "name": "default",
              "primary": "NODE03:3306",
              "ssl": "REQUIRED",
              "status": "OK_NO_TOLERANCE",
              "statusText": "Cluster is NOT tolerant to any failures. 1 member is not active",
              "topology":
              "NODE02:3306":
              "address": "NODE02:3306",
              "mode": "R/O",
              "readReplicas": ,
              "role": "HA",
              "status": "ONLINE"
              ,
              "NODE03:3306":
              "address": "NODE03:3306",
              "mode": "R/W",
              "readReplicas": ,
              "role": "HA",
              "status": "ONLINE"
              ,
              "NODE01:3306":
              "address": "NODE01:3306",
              "mode": "R/O",
              "readReplicas": ,
              "role": "HA",
              "status": "(MISSING)"


              ,
              "groupInformationSourceMember": "mysql://clusterAdmin@NODE03:3306"



              2) Take mysqldump from the master node (healthy node) using the following command.



              [root@NODE03 db_backup]# mysqldump --all-databases --add-drop-database --single-transaction --triggers --routines --port=mysql_port --user=root -p > /db_backup/mysql_dump_03062019.sql
              Enter password:
              Warning: A partial dump from a server that has GTIDs will by default include the GTIDs of all transactions, even those that changed suppressed parts of the database. If you don't want to restore GTIDs, pass --set-gtid-purged=OFF. To make a complete dump, pass --all-databases --triggers --routines --events.


              3) Execute below step to remove the failed node from the cluster.



               MySQL NODE03:3306 ssl JS > var c=dba.getCluster()
              MySQL NODE03:3306 ssl JS > c.rescan()
              Rescanning the cluster...

              Result of the rescanning operation:

              "defaultReplicaSet":
              "name": "default",
              "newlyDiscoveredInstances": [],
              "unavailableInstances": [

              "host": "NODE01:3306",
              "label": "NODE01:3306",
              "member_id": "e2aa897d-1828-11e9-85b3-00505692188c"

              ]



              The instance 'NODE01:3306' is no longer part of the HA setup. It is either offline or left the HA group.
              You can try to add it to the cluster again with the cluster.rejoinInstance('NODE01:3306') command or you can remove it from the cluster configuration.
              Would you like to remove it from the cluster metadata? [Y/n]: Y
              Removing instance from the cluster metadata...

              The instance 'NODE01:3306' was successfully removed from the cluster metadata.

              MySQL NODE03:3306 ssl JS > c.status()

              "clusterName": "QACluster",
              "defaultReplicaSet":
              "name": "default",
              "primary": "NODE03:3306",
              "ssl": "REQUIRED",
              "status": "OK_NO_TOLERANCE",
              "statusText": "Cluster is NOT tolerant to any failures.",
              "topology":
              "NODE02:3306":
              "address": "NODE02:3306",
              "mode": "R/O",
              "readReplicas": ,
              "role": "HA",
              "status": "ONLINE"
              ,
              "NODE03:3306":
              "address": "NODE03:3306",
              "mode": "R/W",
              "readReplicas": ,
              "role": "HA",
              "status": "ONLINE"


              ,
              "groupInformationSourceMember": "mysql://clusterAdmin@NODE03:3306"



              4) Stop group replication if it is still running on failed node.



              mysql> STOP GROUP_REPLICATION;
              Query OK, 0 rows affected (1.01 sec)


              5) Reset "gtid_executed" on the failed node.



              mysql> show global variables like 'GTID_EXECUTED';
              +---------------+--------------------------------------------------------------------------------------------+
              | Variable_name | Value |
              +---------------+--------------------------------------------------------------------------------------------+
              | gtid_executed | 01f27b9c-182a-11e9-a199-00505692188c:1-14134172,
              e2aa897d-1828-11e9-85b3-00505692188c:1-12 |
              +---------------+--------------------------------------------------------------------------------------------+
              1 row in set (0.01 sec)

              mysql> reset master;
              Query OK, 0 rows affected (0.02 sec)

              mysql> reset slave;
              Query OK, 0 rows affected (0.02 sec)

              mysql> show global variables like 'GTID_EXECUTED';
              +---------------+-------+
              | Variable_name | Value |
              +---------------+-------+
              | gtid_executed | |
              +---------------+-------+
              1 row in set (0.00 sec)


              6) Disable "super_readonly_flag" on the failed node.



              mysql> SELECT @@global.read_only, @@global.super_read_only;
              +--------------------+--------------------------+
              | @@global.read_only | @@global.super_read_only |
              +--------------------+--------------------------+
              | 1 | 1 |
              +--------------------+--------------------------+
              1 row in set (0.00 sec)

              mysql> SET GLOBAL super_read_only = 0;
              Query OK, 0 rows affected (0.00 sec)

              mysql> SELECT @@global.read_only, @@global.super_read_only;
              +--------------------+--------------------------+
              | @@global.read_only | @@global.super_read_only |
              +--------------------+--------------------------+
              | 1 | 0 |
              +--------------------+--------------------------+
              1 row in set (0.00 sec)


              7) Restore the mysqldump from master on to the failed node.



              [root@E2LXQA1ALFDB01 db_backup]# mysql -uroot -p < mysql_dump_03062019.sql


              8) Once restore is completed enable "super_readonly_flag" on the failed node.



              mysql> SELECT @@global.read_only, @@global.super_read_only;
              +--------------------+--------------------------+
              | @@global.read_only | @@global.super_read_only |
              +--------------------+--------------------------+
              | 1 | 0 |
              +--------------------+--------------------------+
              1 row in set (0.00 sec)

              mysql> SET GLOBAL super_read_only = 1;
              Query OK, 0 rows affected (0.00 sec)


              mysql> SELECT @@global.read_only, @@global.super_read_only;
              +--------------------+--------------------------+
              | @@global.read_only | @@global.super_read_only |
              +--------------------+--------------------------+
              | 1 | 1 |
              +--------------------+--------------------------+
              1 row in set (0.00 sec)


              9) Finally add the failed node back to the innodb cluster.



              MySQL NODE03:3306 ssl JS > c.addInstance('clusterAdmin@NODE01:3306');
              A new instance will be added to the InnoDB cluster. Depending on the amount of
              data on the cluster this might take from a few seconds to several hours.

              Adding instance to the cluster ...

              Please provide the password for 'clusterAdmin@NODE01:3306': *******************
              Save password for 'clusterAdmin@NODE01:3306'? [Y]es/[N]o/Ne[v]er (default No):
              Validating instance at NODE01:3306...

              This instance reports its own address as NODE01
              WARNING: The following tables do not have a Primary Key or equivalent column:
              ephesoft.dlf, report.correction_type, report.field_details_ag, report_archive.correction_type, report_archive.field_details_ag, report_archive.global_data_ag

              Group Replication requires tables to use InnoDB and have a PRIMARY KEY or PRIMARY KEY Equivalent (non-null unique key). Tables that do not follow these requirements will be readable but not updateable when used with Group Replication. If your applications make updates (INSERT, UPDATE or DELETE) to these tables, ensure they use the InnoDB storage engine and have a PRIMARY KEY or PRIMARY KEY Equivalent.

              Instance configuration is suitable.
              WARNING: On instance 'NODE01:3306' membership change cannot be persisted since MySQL version 5.7.24 does not support the SET PERSIST command (MySQL version >= 8.0.11 required). Please use the .configureLocalInstance command locally to persist the changes.
              WARNING: On instance 'NODE02:3306' membership change cannot be persisted since MySQL version 5.7.24 does not support the SET PERSIST command (MySQL version >= 8.0.11 required). Please use the .configureLocalInstance command locally to persist the changes.
              WARNING: On instance 'NODE03:3306' membership change cannot be persisted since MySQL version 5.7.24 does not support the SET PERSIST command (MySQL version >= 8.0.11 required). Please use the .configureLocalInstance command locally to persist the changes.
              The instance 'clusterAdmin@NODE01:3306' was successfully added to the cluster.


              MySQL NODE03:3306 ssl JS > c.status()

              "clusterName": "QACluster",
              "defaultReplicaSet":
              "name": "default",
              "primary": "NODE03:3306",
              "ssl": "REQUIRED",
              "status": "OK",
              "statusText": "Cluster is ONLINE and can tolerate up to ONE failure.",
              "topology":
              "NODE01:3306":
              "address": "NODE01:3306",
              "mode": "R/O",
              "readReplicas": ,
              "role": "HA",
              "status": "ONLINE"
              ,
              "NODE02:3306":
              "address": "NODE02:3306",
              "mode": "R/O",
              "readReplicas": ,
              "role": "HA",
              "status": "ONLINE"
              ,
              "NODE03:3306":
              "address": "NODE03:3306",
              "mode": "R/W",
              "readReplicas": ,
              "role": "HA",
              "status": "ONLINE"


              ,
              "groupInformationSourceMember": "mysql://clusterAdmin@NODE03:3306"






              share|improve this answer



























                0












                0








                0







                I did the following to restore the failed node from backup and able to recover the cluster state.



                1)Below is the status of the cluster when one of the nodes failed (NODE01).



                 MySQL NODE02:3306 ssl JS > var c=dba.getCluster()
                MySQL NODE02:3306 ssl JS > c.status()

                "clusterName": "QACluster",
                "defaultReplicaSet":
                "name": "default",
                "primary": "NODE03:3306",
                "ssl": "REQUIRED",
                "status": "OK_NO_TOLERANCE",
                "statusText": "Cluster is NOT tolerant to any failures. 1 member is not active",
                "topology":
                "NODE02:3306":
                "address": "NODE02:3306",
                "mode": "R/O",
                "readReplicas": ,
                "role": "HA",
                "status": "ONLINE"
                ,
                "NODE03:3306":
                "address": "NODE03:3306",
                "mode": "R/W",
                "readReplicas": ,
                "role": "HA",
                "status": "ONLINE"
                ,
                "NODE01:3306":
                "address": "NODE01:3306",
                "mode": "R/O",
                "readReplicas": ,
                "role": "HA",
                "status": "(MISSING)"


                ,
                "groupInformationSourceMember": "mysql://clusterAdmin@NODE03:3306"



                2) Take mysqldump from the master node (healthy node) using the following command.



                [root@NODE03 db_backup]# mysqldump --all-databases --add-drop-database --single-transaction --triggers --routines --port=mysql_port --user=root -p > /db_backup/mysql_dump_03062019.sql
                Enter password:
                Warning: A partial dump from a server that has GTIDs will by default include the GTIDs of all transactions, even those that changed suppressed parts of the database. If you don't want to restore GTIDs, pass --set-gtid-purged=OFF. To make a complete dump, pass --all-databases --triggers --routines --events.


                3) Execute below step to remove the failed node from the cluster.



                 MySQL NODE03:3306 ssl JS > var c=dba.getCluster()
                MySQL NODE03:3306 ssl JS > c.rescan()
                Rescanning the cluster...

                Result of the rescanning operation:

                "defaultReplicaSet":
                "name": "default",
                "newlyDiscoveredInstances": [],
                "unavailableInstances": [

                "host": "NODE01:3306",
                "label": "NODE01:3306",
                "member_id": "e2aa897d-1828-11e9-85b3-00505692188c"

                ]



                The instance 'NODE01:3306' is no longer part of the HA setup. It is either offline or left the HA group.
                You can try to add it to the cluster again with the cluster.rejoinInstance('NODE01:3306') command or you can remove it from the cluster configuration.
                Would you like to remove it from the cluster metadata? [Y/n]: Y
                Removing instance from the cluster metadata...

                The instance 'NODE01:3306' was successfully removed from the cluster metadata.

                MySQL NODE03:3306 ssl JS > c.status()

                "clusterName": "QACluster",
                "defaultReplicaSet":
                "name": "default",
                "primary": "NODE03:3306",
                "ssl": "REQUIRED",
                "status": "OK_NO_TOLERANCE",
                "statusText": "Cluster is NOT tolerant to any failures.",
                "topology":
                "NODE02:3306":
                "address": "NODE02:3306",
                "mode": "R/O",
                "readReplicas": ,
                "role": "HA",
                "status": "ONLINE"
                ,
                "NODE03:3306":
                "address": "NODE03:3306",
                "mode": "R/W",
                "readReplicas": ,
                "role": "HA",
                "status": "ONLINE"


                ,
                "groupInformationSourceMember": "mysql://clusterAdmin@NODE03:3306"



                4) Stop group replication if it is still running on failed node.



                mysql> STOP GROUP_REPLICATION;
                Query OK, 0 rows affected (1.01 sec)


                5) Reset "gtid_executed" on the failed node.



                mysql> show global variables like 'GTID_EXECUTED';
                +---------------+--------------------------------------------------------------------------------------------+
                | Variable_name | Value |
                +---------------+--------------------------------------------------------------------------------------------+
                | gtid_executed | 01f27b9c-182a-11e9-a199-00505692188c:1-14134172,
                e2aa897d-1828-11e9-85b3-00505692188c:1-12 |
                +---------------+--------------------------------------------------------------------------------------------+
                1 row in set (0.01 sec)

                mysql> reset master;
                Query OK, 0 rows affected (0.02 sec)

                mysql> reset slave;
                Query OK, 0 rows affected (0.02 sec)

                mysql> show global variables like 'GTID_EXECUTED';
                +---------------+-------+
                | Variable_name | Value |
                +---------------+-------+
                | gtid_executed | |
                +---------------+-------+
                1 row in set (0.00 sec)


                6) Disable "super_readonly_flag" on the failed node.



                mysql> SELECT @@global.read_only, @@global.super_read_only;
                +--------------------+--------------------------+
                | @@global.read_only | @@global.super_read_only |
                +--------------------+--------------------------+
                | 1 | 1 |
                +--------------------+--------------------------+
                1 row in set (0.00 sec)

                mysql> SET GLOBAL super_read_only = 0;
                Query OK, 0 rows affected (0.00 sec)

                mysql> SELECT @@global.read_only, @@global.super_read_only;
                +--------------------+--------------------------+
                | @@global.read_only | @@global.super_read_only |
                +--------------------+--------------------------+
                | 1 | 0 |
                +--------------------+--------------------------+
                1 row in set (0.00 sec)


                7) Restore the mysqldump from master on to the failed node.



                [root@E2LXQA1ALFDB01 db_backup]# mysql -uroot -p < mysql_dump_03062019.sql


                8) Once restore is completed enable "super_readonly_flag" on the failed node.



                mysql> SELECT @@global.read_only, @@global.super_read_only;
                +--------------------+--------------------------+
                | @@global.read_only | @@global.super_read_only |
                +--------------------+--------------------------+
                | 1 | 0 |
                +--------------------+--------------------------+
                1 row in set (0.00 sec)

                mysql> SET GLOBAL super_read_only = 1;
                Query OK, 0 rows affected (0.00 sec)


                mysql> SELECT @@global.read_only, @@global.super_read_only;
                +--------------------+--------------------------+
                | @@global.read_only | @@global.super_read_only |
                +--------------------+--------------------------+
                | 1 | 1 |
                +--------------------+--------------------------+
                1 row in set (0.00 sec)


                9) Finally add the failed node back to the innodb cluster.



                MySQL NODE03:3306 ssl JS > c.addInstance('clusterAdmin@NODE01:3306');
                A new instance will be added to the InnoDB cluster. Depending on the amount of
                data on the cluster this might take from a few seconds to several hours.

                Adding instance to the cluster ...

                Please provide the password for 'clusterAdmin@NODE01:3306': *******************
                Save password for 'clusterAdmin@NODE01:3306'? [Y]es/[N]o/Ne[v]er (default No):
                Validating instance at NODE01:3306...

                This instance reports its own address as NODE01
                WARNING: The following tables do not have a Primary Key or equivalent column:
                ephesoft.dlf, report.correction_type, report.field_details_ag, report_archive.correction_type, report_archive.field_details_ag, report_archive.global_data_ag

                Group Replication requires tables to use InnoDB and have a PRIMARY KEY or PRIMARY KEY Equivalent (non-null unique key). Tables that do not follow these requirements will be readable but not updateable when used with Group Replication. If your applications make updates (INSERT, UPDATE or DELETE) to these tables, ensure they use the InnoDB storage engine and have a PRIMARY KEY or PRIMARY KEY Equivalent.

                Instance configuration is suitable.
                WARNING: On instance 'NODE01:3306' membership change cannot be persisted since MySQL version 5.7.24 does not support the SET PERSIST command (MySQL version >= 8.0.11 required). Please use the .configureLocalInstance command locally to persist the changes.
                WARNING: On instance 'NODE02:3306' membership change cannot be persisted since MySQL version 5.7.24 does not support the SET PERSIST command (MySQL version >= 8.0.11 required). Please use the .configureLocalInstance command locally to persist the changes.
                WARNING: On instance 'NODE03:3306' membership change cannot be persisted since MySQL version 5.7.24 does not support the SET PERSIST command (MySQL version >= 8.0.11 required). Please use the .configureLocalInstance command locally to persist the changes.
                The instance 'clusterAdmin@NODE01:3306' was successfully added to the cluster.


                MySQL NODE03:3306 ssl JS > c.status()

                "clusterName": "QACluster",
                "defaultReplicaSet":
                "name": "default",
                "primary": "NODE03:3306",
                "ssl": "REQUIRED",
                "status": "OK",
                "statusText": "Cluster is ONLINE and can tolerate up to ONE failure.",
                "topology":
                "NODE01:3306":
                "address": "NODE01:3306",
                "mode": "R/O",
                "readReplicas": ,
                "role": "HA",
                "status": "ONLINE"
                ,
                "NODE02:3306":
                "address": "NODE02:3306",
                "mode": "R/O",
                "readReplicas": ,
                "role": "HA",
                "status": "ONLINE"
                ,
                "NODE03:3306":
                "address": "NODE03:3306",
                "mode": "R/W",
                "readReplicas": ,
                "role": "HA",
                "status": "ONLINE"


                ,
                "groupInformationSourceMember": "mysql://clusterAdmin@NODE03:3306"






                share|improve this answer















                I did the following to restore the failed node from backup and able to recover the cluster state.



                1)Below is the status of the cluster when one of the nodes failed (NODE01).



                 MySQL NODE02:3306 ssl JS > var c=dba.getCluster()
                MySQL NODE02:3306 ssl JS > c.status()

                "clusterName": "QACluster",
                "defaultReplicaSet":
                "name": "default",
                "primary": "NODE03:3306",
                "ssl": "REQUIRED",
                "status": "OK_NO_TOLERANCE",
                "statusText": "Cluster is NOT tolerant to any failures. 1 member is not active",
                "topology":
                "NODE02:3306":
                "address": "NODE02:3306",
                "mode": "R/O",
                "readReplicas": ,
                "role": "HA",
                "status": "ONLINE"
                ,
                "NODE03:3306":
                "address": "NODE03:3306",
                "mode": "R/W",
                "readReplicas": ,
                "role": "HA",
                "status": "ONLINE"
                ,
                "NODE01:3306":
                "address": "NODE01:3306",
                "mode": "R/O",
                "readReplicas": ,
                "role": "HA",
                "status": "(MISSING)"


                ,
                "groupInformationSourceMember": "mysql://clusterAdmin@NODE03:3306"



                2) Take mysqldump from the master node (healthy node) using the following command.



                [root@NODE03 db_backup]# mysqldump --all-databases --add-drop-database --single-transaction --triggers --routines --port=mysql_port --user=root -p > /db_backup/mysql_dump_03062019.sql
                Enter password:
                Warning: A partial dump from a server that has GTIDs will by default include the GTIDs of all transactions, even those that changed suppressed parts of the database. If you don't want to restore GTIDs, pass --set-gtid-purged=OFF. To make a complete dump, pass --all-databases --triggers --routines --events.


                3) Execute below step to remove the failed node from the cluster.



                 MySQL NODE03:3306 ssl JS > var c=dba.getCluster()
                MySQL NODE03:3306 ssl JS > c.rescan()
                Rescanning the cluster...

                Result of the rescanning operation:

                "defaultReplicaSet":
                "name": "default",
                "newlyDiscoveredInstances": [],
                "unavailableInstances": [

                "host": "NODE01:3306",
                "label": "NODE01:3306",
                "member_id": "e2aa897d-1828-11e9-85b3-00505692188c"

                ]



                The instance 'NODE01:3306' is no longer part of the HA setup. It is either offline or left the HA group.
                You can try to add it to the cluster again with the cluster.rejoinInstance('NODE01:3306') command or you can remove it from the cluster configuration.
                Would you like to remove it from the cluster metadata? [Y/n]: Y
                Removing instance from the cluster metadata...

                The instance 'NODE01:3306' was successfully removed from the cluster metadata.

                MySQL NODE03:3306 ssl JS > c.status()

                "clusterName": "QACluster",
                "defaultReplicaSet":
                "name": "default",
                "primary": "NODE03:3306",
                "ssl": "REQUIRED",
                "status": "OK_NO_TOLERANCE",
                "statusText": "Cluster is NOT tolerant to any failures.",
                "topology":
                "NODE02:3306":
                "address": "NODE02:3306",
                "mode": "R/O",
                "readReplicas": ,
                "role": "HA",
                "status": "ONLINE"
                ,
                "NODE03:3306":
                "address": "NODE03:3306",
                "mode": "R/W",
                "readReplicas": ,
                "role": "HA",
                "status": "ONLINE"


                ,
                "groupInformationSourceMember": "mysql://clusterAdmin@NODE03:3306"



                4) Stop group replication if it is still running on failed node.



                mysql> STOP GROUP_REPLICATION;
                Query OK, 0 rows affected (1.01 sec)


                5) Reset "gtid_executed" on the failed node.



                mysql> show global variables like 'GTID_EXECUTED';
                +---------------+--------------------------------------------------------------------------------------------+
                | Variable_name | Value |
                +---------------+--------------------------------------------------------------------------------------------+
                | gtid_executed | 01f27b9c-182a-11e9-a199-00505692188c:1-14134172,
                e2aa897d-1828-11e9-85b3-00505692188c:1-12 |
                +---------------+--------------------------------------------------------------------------------------------+
                1 row in set (0.01 sec)

                mysql> reset master;
                Query OK, 0 rows affected (0.02 sec)

                mysql> reset slave;
                Query OK, 0 rows affected (0.02 sec)

                mysql> show global variables like 'GTID_EXECUTED';
                +---------------+-------+
                | Variable_name | Value |
                +---------------+-------+
                | gtid_executed | |
                +---------------+-------+
                1 row in set (0.00 sec)


                6) Disable "super_readonly_flag" on the failed node.



                mysql> SELECT @@global.read_only, @@global.super_read_only;
                +--------------------+--------------------------+
                | @@global.read_only | @@global.super_read_only |
                +--------------------+--------------------------+
                | 1 | 1 |
                +--------------------+--------------------------+
                1 row in set (0.00 sec)

                mysql> SET GLOBAL super_read_only = 0;
                Query OK, 0 rows affected (0.00 sec)

                mysql> SELECT @@global.read_only, @@global.super_read_only;
                +--------------------+--------------------------+
                | @@global.read_only | @@global.super_read_only |
                +--------------------+--------------------------+
                | 1 | 0 |
                +--------------------+--------------------------+
                1 row in set (0.00 sec)


                7) Restore the mysqldump from master on to the failed node.



                [root@E2LXQA1ALFDB01 db_backup]# mysql -uroot -p < mysql_dump_03062019.sql


                8) Once restore is completed enable "super_readonly_flag" on the failed node.



                mysql> SELECT @@global.read_only, @@global.super_read_only;
                +--------------------+--------------------------+
                | @@global.read_only | @@global.super_read_only |
                +--------------------+--------------------------+
                | 1 | 0 |
                +--------------------+--------------------------+
                1 row in set (0.00 sec)

                mysql> SET GLOBAL super_read_only = 1;
                Query OK, 0 rows affected (0.00 sec)


                mysql> SELECT @@global.read_only, @@global.super_read_only;
                +--------------------+--------------------------+
                | @@global.read_only | @@global.super_read_only |
                +--------------------+--------------------------+
                | 1 | 1 |
                +--------------------+--------------------------+
                1 row in set (0.00 sec)


                9) Finally add the failed node back to the innodb cluster.



                MySQL NODE03:3306 ssl JS > c.addInstance('clusterAdmin@NODE01:3306');
                A new instance will be added to the InnoDB cluster. Depending on the amount of
                data on the cluster this might take from a few seconds to several hours.

                Adding instance to the cluster ...

                Please provide the password for 'clusterAdmin@NODE01:3306': *******************
                Save password for 'clusterAdmin@NODE01:3306'? [Y]es/[N]o/Ne[v]er (default No):
                Validating instance at NODE01:3306...

                This instance reports its own address as NODE01
                WARNING: The following tables do not have a Primary Key or equivalent column:
                ephesoft.dlf, report.correction_type, report.field_details_ag, report_archive.correction_type, report_archive.field_details_ag, report_archive.global_data_ag

                Group Replication requires tables to use InnoDB and have a PRIMARY KEY or PRIMARY KEY Equivalent (non-null unique key). Tables that do not follow these requirements will be readable but not updateable when used with Group Replication. If your applications make updates (INSERT, UPDATE or DELETE) to these tables, ensure they use the InnoDB storage engine and have a PRIMARY KEY or PRIMARY KEY Equivalent.

                Instance configuration is suitable.
                WARNING: On instance 'NODE01:3306' membership change cannot be persisted since MySQL version 5.7.24 does not support the SET PERSIST command (MySQL version >= 8.0.11 required). Please use the .configureLocalInstance command locally to persist the changes.
                WARNING: On instance 'NODE02:3306' membership change cannot be persisted since MySQL version 5.7.24 does not support the SET PERSIST command (MySQL version >= 8.0.11 required). Please use the .configureLocalInstance command locally to persist the changes.
                WARNING: On instance 'NODE03:3306' membership change cannot be persisted since MySQL version 5.7.24 does not support the SET PERSIST command (MySQL version >= 8.0.11 required). Please use the .configureLocalInstance command locally to persist the changes.
                The instance 'clusterAdmin@NODE01:3306' was successfully added to the cluster.


                MySQL NODE03:3306 ssl JS > c.status()

                "clusterName": "QACluster",
                "defaultReplicaSet":
                "name": "default",
                "primary": "NODE03:3306",
                "ssl": "REQUIRED",
                "status": "OK",
                "statusText": "Cluster is ONLINE and can tolerate up to ONE failure.",
                "topology":
                "NODE01:3306":
                "address": "NODE01:3306",
                "mode": "R/O",
                "readReplicas": ,
                "role": "HA",
                "status": "ONLINE"
                ,
                "NODE02:3306":
                "address": "NODE02:3306",
                "mode": "R/O",
                "readReplicas": ,
                "role": "HA",
                "status": "ONLINE"
                ,
                "NODE03:3306":
                "address": "NODE03:3306",
                "mode": "R/W",
                "readReplicas": ,
                "role": "HA",
                "status": "ONLINE"


                ,
                "groupInformationSourceMember": "mysql://clusterAdmin@NODE03:3306"







                share|improve this answer














                share|improve this answer



                share|improve this answer








                edited Mar 15 at 3:37

























                answered Mar 9 at 4:31









                sqlcheckpointsqlcheckpoint

                351519




                351519



























                    draft saved

                    draft discarded
















































                    Thanks for contributing an answer to Stack Overflow!


                    • Please be sure to answer the question. Provide details and share your research!

                    But avoid


                    • Asking for help, clarification, or responding to other answers.

                    • Making statements based on opinion; back them up with references or personal experience.

                    To learn more, see our tips on writing great answers.




                    draft saved


                    draft discarded














                    StackExchange.ready(
                    function ()
                    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f55036255%2fa-node-in-mysql-5-7-innodb-cluster-is-crashed-and-unable-to-re-join-the-crashed%23new-answer', 'question_page');

                    );

                    Post as a guest















                    Required, but never shown





















































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown

































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown







                    Popular posts from this blog

                    1928 у кіно

                    Захаров Федір Захарович

                    Ель Греко