1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147
|
###############################################################################
# This test verifies that if the recovery endpoints contains multiple valid
# entries it shall succeed on first endpoint
#
# Test:
# 0. The test requires two servers
# 1. Install clone plugin on server1.
# 2. Configure recovery endpoints with two valid endpoints
# 3. Bootstrap server1 and add some data
# 4. Restart server 2 with a monitoring process (mysqld_safe) if needed
# 5. Setup the server so group replication starts on boot
# Install the Clone plugin
# 6. Ensure clone is used on recovery
# 7. On a empty server2 start group replication
# Wait for it to restart and come back
# 8. Cleanup
#
###############################################################################
--source include/have_mysqld_monitoring_process.inc
--source include/have_clone_plugin.inc
--source include/have_group_replication_plugin.inc
--let $rpl_skip_group_replication_start= 1
--source include/group_replication.inc
set session sql_log_bin=0;
call mtr.add_suppression("Failed to initialize TLS for channel: mysql_admin");
call mtr.add_suppression("Failed to set up SSL because of the following SSL library error");
set session sql_log_bin=1;
--echo
--echo # 1. Install clone plugin on server1.
--let $rpl_connection_name= server1
--source include/rpl_connection.inc
--let $server1_port= `SELECT @@GLOBAL.PORT`
--let $server1_admin_port= `SELECT @@GLOBAL.ADMIN_PORT`
--replace_result $CLONE_PLUGIN CLONE_PLUGIN
--eval INSTALL PLUGIN clone SONAME '$CLONE_PLUGIN'
--echo
--echo # 2. Configure recovery endpoints with two valid endpoints
--replace_result $server1_admin_port SERVER1_ADMIN_PORT $server1_port SERVER1_PORT
--eval SET @@GLOBAL.group_replication_advertise_recovery_endpoints = "127.0.0.1:$server1_admin_port,127.0.0.1:$server1_port"
--echo
--echo # 3. Bootstrap server1 and add some data
--source include/start_and_bootstrap_group_replication.inc
CREATE TABLE t1 (c1 INT NOT NULL PRIMARY KEY) ENGINE=InnoDB;
INSERT INTO t1 VALUES (1);
INSERT INTO t1 VALUES (2);
--echo
--echo # 4. Restart server 2 with a monitoring process (mysqld_safe) if needed
--let $rpl_connection_name= server2
--source include/rpl_connection.inc
--let $_group_replication_local_address= `SELECT @@GLOBAL.group_replication_local_address`
--let $_group_replication_group_seeds= `SELECT @@GLOBAL.group_replication_group_seeds`
--let $_group_replication_start_on_boot= `SELECT @@GLOBAL.group_replication_start_on_boot`
--let $_group_replication_comm_stack= `SELECT @@GLOBAL.group_replication_communication_stack`
--let $plugin_list= $GROUP_REPLICATION
--source include/spawn_monitoring_process.inc
--echo
--echo # 5. Setup the server so group replication starts on boot
--echo # Install the Clone plugin
--disable_query_log
--eval SET PERSIST group_replication_group_name= "$group_replication_group_name"
--eval SET PERSIST group_replication_local_address= "$_group_replication_local_address"
--eval SET PERSIST group_replication_group_seeds= "$_group_replication_group_seeds"
--eval SET PERSIST group_replication_communication_stack= "$_group_replication_comm_stack"
SET PERSIST group_replication_start_on_boot= ON;
--enable_query_log
--replace_result $CLONE_PLUGIN CLONE_PLUGIN
--eval INSTALL PLUGIN clone SONAME '$CLONE_PLUGIN'
--echo
--echo # 6. Ensure clone is used on recovery
--let $_group_replication_threshold_save= `SELECT @@GLOBAL.group_replication_clone_threshold`
SET GLOBAL group_replication_clone_threshold= 1;
--echo
--echo # 7. On a empty server2 start group replication
--echo # Wait for it to restart and come back
START GROUP_REPLICATION;
--source include/wait_until_disconnected.inc
--let $rpl_server_number= 2
--source include/rpl_reconnect.inc
--let $group_replication_member_state=ONLINE
--source include/gr_wait_for_member_state.inc
# See if the data has been properly cloned in server2
--let $diff_tables=server1:test.t1 ,server2:test.t1
--source include/diff_tables.inc
--echo
--echo # 8. Cleanup
--eval SET GLOBAL group_replication_clone_threshold= $_group_replication_threshold_save
RESET PERSIST IF EXISTS group_replication_group_name;
RESET PERSIST IF EXISTS group_replication_local_address;
RESET PERSIST IF EXISTS group_replication_group_seeds;
RESET PERSIST IF EXISTS group_replication_start_on_boot;
RESET PERSIST IF EXISTS group_replication_communication_stack;
--replace_result $_group_replication_start_on_boot START_ON_BOOT_VALUE
--eval SET GLOBAL group_replication_start_on_boot= $_group_replication_start_on_boot
DROP TABLE t1;
--source include/rpl_sync.inc
set session sql_log_bin=0;
call mtr.add_suppression("Due to the number of missing transactions being higher than the configured threshold of*");
call mtr.add_suppression("Clone removing all user data for provisioning: Started");
call mtr.add_suppression("Clone removing all user data for provisioning: Finished");
set session sql_log_bin=1;
--source include/clean_monitoring_process.inc
UNINSTALL PLUGIN clone;
--let $rpl_connection_name= server1
--source include/rpl_connection.inc
--eval SET @@GLOBAL.group_replication_advertise_recovery_endpoints= "DEFAULT"
UNINSTALL PLUGIN clone;
--source include/group_replication_end.inc
|