File: gr_recovery_endpoints_invalid_donor.test

package info (click to toggle)
mysql-8.0 8.0.43-3
  • links: PTS, VCS
  • area: main
  • in suites: sid
  • size: 1,273,924 kB
  • sloc: cpp: 4,684,605; ansic: 412,450; pascal: 108,398; java: 83,641; perl: 30,221; cs: 27,067; sql: 26,594; sh: 24,181; python: 21,816; yacc: 17,169; php: 11,522; xml: 7,388; javascript: 7,076; makefile: 2,194; lex: 1,075; awk: 670; asm: 520; objc: 183; ruby: 97; lisp: 86
file content (184 lines) | stat: -rw-r--r-- 6,890 bytes parent folder | download
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
###############################################################################
# Received a donor endpoints corrupted and fail recovery.
#
# Test:
#   0. The test requires two servers
#   1. Install clone plugin on server1.
#   2. Start server 1 and bootstrap group
#   3. Create some transactions to send to joiner
#   4. Change settings on recovery to speed up failure
#   5. Setup the server so group replication starts on boot
#      Install the Clone plugin
#   6. Ensure clone is used on recovery
#   7. Activate debug point that will simulate invalid recovery endpoints
#      from donor
#   8. Recovery will fail and server state change to ERROR
#   9. Clone and incremental distributed recovery will fail connecting to donor
#  10. Disable debug point ans server will be able to join
#  11. Stop server 2 and start it again, the information received
#      from donor will be valid
#  12. Clean debug point to allow to recovery be successful
#  13. Cleanup
#
###############################################################################

--source include/have_debug.inc
--source include/have_mysqld_monitoring_process.inc
--source include/have_clone_plugin.inc
--source include/have_group_replication_plugin.inc
--let $rpl_skip_group_replication_start= 1
--source include/group_replication.inc

--echo
--echo # 1. Install clone plugin on server1.

--let $rpl_connection_name= server1
--source include/rpl_connection.inc

--replace_result $CLONE_PLUGIN CLONE_PLUGIN
--eval INSTALL PLUGIN clone SONAME '$CLONE_PLUGIN'

--echo
--echo # 2. Start server 1 and bootstrap group

--source include/start_and_bootstrap_group_replication.inc

--echo
--echo # 3. Create some transactions to send to joiner

CREATE TABLE t1 (a INT NOT NULL AUTO_INCREMENT PRIMARY KEY , b INT);

INSERT INTO test.t1 (b) VALUES (1);
INSERT INTO test.t1 (b) VALUES (2);

--echo
--echo # 4. Change settings on recovery to speed up failure

--let $rpl_connection_name = server2
--source include/rpl_connection.inc

--let $_group_replication_local_address= `SELECT @@GLOBAL.group_replication_local_address`
--let $_group_replication_group_seeds= `SELECT @@GLOBAL.group_replication_group_seeds`
--let $_group_replication_start_on_boot= `SELECT @@GLOBAL.group_replication_start_on_boot`
--let $_group_replication_comm_stack= `SELECT @@GLOBAL.group_replication_communication_stack`

--let $plugin_list= $GROUP_REPLICATION
--source include/spawn_monitoring_process.inc

SET GLOBAL group_replication_recovery_reconnect_interval= 1;
SET GLOBAL group_replication_recovery_retry_count= 1;

--echo
--echo # 5. Setup the server so group replication starts on boot
--echo #    Install the Clone plugin

--disable_query_log
--eval SET PERSIST group_replication_group_name= "$group_replication_group_name"
--eval SET PERSIST group_replication_local_address= "$_group_replication_local_address"
--eval SET PERSIST group_replication_group_seeds= "$_group_replication_group_seeds"
--eval SET PERSIST group_replication_communication_stack= "$_group_replication_comm_stack"

SET PERSIST group_replication_start_on_boot= ON;
--enable_query_log

--replace_result $CLONE_PLUGIN CLONE_PLUGIN
--eval INSTALL PLUGIN clone SONAME '$CLONE_PLUGIN'

--echo
--echo # 6. Ensure clone is used on recovery

--let $_group_replication_threshold_save= `SELECT @@GLOBAL.group_replication_clone_threshold`
SET GLOBAL group_replication_clone_threshold= 1;

--echo
--echo # 7. Activate debug point that will simulate invalid recovery endpoints
--echo #    from donor

SET @debug_save= @@GLOBAL.DEBUG;
SET @@GLOBAL.DEBUG='+d,gr_recovery_endpoints_invalid_donor';

--echo
--echo # 8. Recovery will fail and server state change to ERROR

--let $group_replication_start_member_state= ERROR
--source include/start_group_replication.inc

--echo
--echo # 9. Clone and incremental distributed recovery will fail connecting to donor

--let $assert_only_after = CURRENT_TEST: group_replication.gr_recovery_endpoints_invalid_donor
--let $assert_file= $MYSQLTEST_VARDIR/log/mysqld.2.err
--let $assert_count = 2
--let $assert_select = Received invalid recovery endpoints configuration from donor. This member is not a valid donor for recovery, so it will be skipped.
--let $assert_text = recovery channel received an invalid recovery endpoints configuration from donor
--source include/assert_grep.inc

--echo
--echo # 10. Disable debug point ans server will be able to join

SET @@GLOBAL.DEBUG= @debug_save;

--echo
--echo # 11. Stop server 2 and start it again, the information received
--echo #     from donor will be valid

--let $rpl_connection_name = server2
--source include/rpl_connection.inc

--source include/stop_group_replication.inc

--echo
--echo # 12. Clean debug point to allow to recovery be successful


START GROUP_REPLICATION;

--source include/wait_until_disconnected.inc

--let $rpl_server_number= 2
--source include/rpl_reconnect.inc

--let $group_replication_member_state=ONLINE
--source include/gr_wait_for_member_state.inc

--echo
--echo # 13. Cleanup

DROP TABLE test.t1;

--source include/rpl_sync.inc

set session sql_log_bin=0;
call mtr.add_suppression("Maximum number of retries when*");
call mtr.add_suppression("Fatal error during the incremental recovery process of Group Replication.*");
call mtr.add_suppression("Skipping leave operation: concurrent attempt to leave the group is on-going.");
call mtr.add_suppression("The server was automatically set into read only mode after an error was detected.");
call mtr.add_suppression("Due to the number of missing transactions being higher than the configured threshold of*");
call mtr.add_suppression("Clone removing all user data for provisioning: Started");
call mtr.add_suppression("Clone removing all user data for provisioning: Finished");
call mtr.add_suppression("Due to some issue on the previous step distributed recovery is now executing: Incremental Recovery.");
call mtr.add_suppression("Failed to shutdown components infrastructure.");
call mtr.add_suppression("Received invalid recovery endpoints configuration from donor. This member is not a valid donor for recovery, so it will be skipped.");
set session sql_log_bin=1;

--eval SET GLOBAL group_replication_clone_threshold= $_group_replication_threshold_save

RESET PERSIST IF EXISTS group_replication_group_name;
RESET PERSIST IF EXISTS group_replication_local_address;
RESET PERSIST IF EXISTS group_replication_group_seeds;
RESET PERSIST IF EXISTS group_replication_start_on_boot;
RESET PERSIST IF EXISTS group_replication_communication_stack;

--replace_result $_group_replication_start_on_boot START_ON_BOOT_VALUE
--eval SET GLOBAL group_replication_start_on_boot= $_group_replication_start_on_boot

--source include/clean_monitoring_process.inc

UNINSTALL PLUGIN clone;

--let $rpl_connection_name= server1
--source include/rpl_connection.inc

UNINSTALL PLUGIN clone;

--source include/group_replication_end.inc