File: galera_sr_cc_no_primary.test

package info (click to toggle)
mariadb 1%3A11.8.2-1
  • links: PTS, VCS
  • area: main
  • in suites: forky, trixie
  • size: 765,428 kB
  • sloc: ansic: 2,382,827; cpp: 1,803,532; asm: 378,315; perl: 63,176; sh: 46,496; pascal: 40,776; java: 39,363; yacc: 20,428; python: 19,506; sql: 17,864; xml: 12,463; ruby: 8,544; makefile: 6,059; cs: 5,855; ada: 1,700; lex: 1,193; javascript: 1,039; objc: 80; tcl: 73; awk: 46; php: 22
file content (85 lines) | stat: -rw-r--r-- 2,675 bytes parent folder | download | duplicates (5)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
#
# This test verifies that an orphaned SR gets cleanup upon cluster
# reconnection.  Specifically, the case where the cluster goes through
# a state of no primary components, and the nodes rejoin with the
# same IDs.
#

--source include/galera_cluster.inc

--connect node_1a, 127.0.0.1, root, , test, $NODE_MYPORT_1
--connect node_2a, 127.0.0.1, root, , test, $NODE_MYPORT_2

# Save original auto_increment_offset values.
--let $node_1=node_1
--let $node_2=node_2
--source ../galera/include/auto_increment_offset_save.inc

--connection node_1
CREATE TABLE t1 (f1 INTEGER PRIMARY KEY);

SET SESSION wsrep_trx_fragment_size=1;
BEGIN;
INSERT INTO t1 VALUES (10);
INSERT INTO t1 VALUES (11);
INSERT INTO t1 VALUES (12);

--connection node_2
SET SESSION wsrep_trx_fragment_size=1;
BEGIN;
INSERT INTO t1 VALUES (20);
INSERT INTO t1 VALUES (21);
INSERT INTO t1 VALUES (22);
SELECT COUNT(*) `expect 6` FROM mysql.wsrep_streaming_log;

# isolate node 2
--connection node_2a
SET GLOBAL wsrep_provider_options = 'gmcast.isolate = 1';
SET SESSION wsrep_sync_wait = 0;
--let $wait_condition = SELECT VARIABLE_VALUE = 'OFF' FROM INFORMATION_SCHEMA.GLOBAL_STATUS WHERE VARIABLE_NAME = 'wsrep_ready'
--source include/wait_condition.inc
SHOW STATUS LIKE 'wsrep_cluster_size';
SHOW STATUS LIKE 'wsrep_cluster_status';
SET SESSION wsrep_sync_wait = DEFAULT;

--connection node_1a
SET SESSION wsrep_sync_wait = 0;
--let $wait_condition = SELECT VARIABLE_VALUE = 'OFF' FROM INFORMATION_SCHEMA.GLOBAL_STATUS WHERE VARIABLE_NAME = 'wsrep_ready'
--source include/wait_condition.inc
SHOW STATUS LIKE 'wsrep_cluster_size';
SHOW STATUS LIKE 'wsrep_cluster_status';
SET SESSION wsrep_sync_wait = DEFAULT;

--connection node_2a
# reconnect node 2
SET GLOBAL wsrep_provider_options = 'gmcast.isolate = 0';
--source include/galera_wait_ready.inc
SHOW STATUS LIKE 'wsrep_cluster_size';

--connection node_1a
# wait for reconnection and check that the streaming log has been cleared
--source include/galera_wait_ready.inc
SHOW STATUS LIKE 'wsrep_cluster_size';
SELECT COUNT(*) `expect 0` FROM mysql.wsrep_streaming_log;

--connection node_2a
# check that the streaming log has been cleared and there are no locks
# from the SRs by issuing conflicting inserts
SELECT COUNT(*) `expect 0` FROM mysql.wsrep_streaming_log;
INSERT INTO t1 VALUES(10);
INSERT INTO t1 VALUES(20);
SELECT * FROM t1;

# check that both SRs have been rolled back
--connection node_1
--error ER_LOCK_DEADLOCK
INSERT INTO t1 VALUES(13);

--connection node_2
--error ER_LOCK_DEADLOCK
INSERT INTO t1 VALUES(23);

DROP TABLE t1;

# Restore original auto_increment_offset values.
--source ../galera/include/auto_increment_offset_restore.inc