1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329
|
# See the file LICENSE for redistribution information.
#
# Copyright (c) 2003, 2013 Oracle and/or its affiliates. All rights reserved.
#
# $Id$
#
# TEST rep015
# TEST Locking across multiple pages with replication.
# TEST
# TEST Open master and client with small pagesize and
# TEST generate more than one page and generate off-page
# TEST dups on the first page (second key) and last page
# TEST (next-to-last key).
# TEST Within a single transaction, for each database, open
# TEST 2 cursors and delete the first and last entries (this
# TEST exercises locks on regular pages). Intermittently
# TEST update client during the process.
# TEST Within a single transaction, for each database, open
# TEST 2 cursors. Walk to the off-page dups and delete one
# TEST from each end (this exercises locks on off-page dups).
# TEST Intermittently update client.
#
proc rep015 { method { nentries 100 } { tnum "015" } { ndb 3 } args } {
global repfiles_in_memory
global env_private
global rand_init
berkdb srand $rand_init
source ./include.tcl
# Run for btree only.
if { $checking_valid_methods } {
set test_methods { btree }
return $test_methods
}
if { [is_btree $method] == 0 } {
puts "Skipping rep$tnum for method $method."
return
}
set args [convert_args $method $args]
set logsets [create_logsets 2]
set msg2 "and on-disk replication files"
if { $repfiles_in_memory } {
set msg2 "and in-memory replication files"
}
set msg3 ""
if { $env_private } {
set msg3 "and private env"
}
# Run the body of the test with and without recovery.
foreach r $test_recopts {
foreach l $logsets {
set logindex [lsearch -exact $l "in-memory"]
if { $r == "-recover" && $logindex != -1 } {
puts "Rep$tnum: \
Skipping for in-memory logs with -recover."
continue
}
puts "Rep$tnum ($method $r):\
Replication and locking $msg2 $msg3."
puts "Rep$tnum: Master logs are [lindex $l 0]"
puts "Rep$tnum: Client logs are [lindex $l 1]"
rep015_sub $method $nentries $tnum $ndb $l $r $args
}
}
}
proc rep015_sub { method nentries tnum ndb logset recargs largs } {
global testdir
global repfiles_in_memory
global env_private
global rep_verbose
global verbose_type
set verbargs ""
if { $rep_verbose == 1 } {
set verbargs " -verbose {$verbose_type on} "
}
set repmemargs ""
if { $repfiles_in_memory } {
set repmemargs "-rep_inmem_files "
}
set privargs ""
if { $env_private == 1 } {
set privargs " -private "
}
env_cleanup $testdir
set omethod [convert_method $method]
replsetup $testdir/MSGQUEUEDIR
set masterdir $testdir/MASTERDIR
set clientdir $testdir/CLIENTDIR
file mkdir $masterdir
file mkdir $clientdir
set m_logtype [lindex $logset 0]
set c_logtype [lindex $logset 1]
# In-memory logs require a large log buffer, and cannot
# be used with -txn nosync.
set m_logargs [adjust_logargs $m_logtype]
set c_logargs [adjust_logargs $c_logtype]
set m_txnargs [adjust_txnargs $m_logtype]
set c_txnargs [adjust_txnargs $c_logtype]
# Open a master.
repladd 1
set ma_envcmd "berkdb_env_noerr -create $m_txnargs $m_logargs \
$verbargs -errpfx MASTER $repmemargs $privargs \
-home $masterdir -rep_transport \[list 1 replsend\]"
set masterenv [eval $ma_envcmd $recargs -rep_master]
# Open a client
repladd 2
set cl_envcmd "berkdb_env_noerr -create $c_txnargs $c_logargs \
$verbargs -errpfx CLIENT $repmemargs $privargs \
-home $clientdir -rep_transport \[list 2 replsend\]"
set clientenv [eval $cl_envcmd $recargs -rep_client]
# Bring the clients online by processing the startup messages.
set envlist "{$masterenv 1} {$clientenv 2}"
process_msgs $envlist
# Set up the master databases. The small pagesize quickly
# generates multiple pages and off-page dups.
set pagesize 512
puts "\tRep$tnum.a: Create and populate databases in master."
for { set i 0 } { $i < $ndb } { incr i } {
set db [eval berkdb_open_noerr -create $omethod -auto_commit \
-pagesize $pagesize -env $masterenv $largs -dup testdb$i.db]
set dblist($i) $db
#
# Populate, being sure to create multiple pages.
# The non-duplicate entries are pairs of the form
# {1, data1} {2, data2}. The duplicates are pairs of
# the form {2, dup1} {2, dup2}, {2, dup3}, etc.
#
for { set j 1 } { $j <= $nentries } { incr j } {
set t [$masterenv txn]
error_check_good put_$db [eval $db put -txn $t \
$j [chop_data $method data$j]] 0
error_check_good txn_commit [$t commit] 0
}
# Create off-page dups on key 2 and next-to-last key.
set t [$masterenv txn]
for { set j 1 } { $j <= $nentries } { incr j } {
error_check_good put_second [eval $db put -txn $t \
2 [chop_data $method dup$j]] 0
error_check_good put_next_to_last [eval $db put \
-txn $t \
[expr $nentries - 1] [chop_data $method dup$j]] 0
}
error_check_good txn_commit [$t commit] 0
# Make sure there are off-page dups.
set stat [$db stat]
error_check_bad stat:offpage \
[is_substr $stat "{{Internal pages} 0}"] 1
}
puts "\tRep$tnum.b: Propagate setup to clients."
process_msgs $envlist
# Open client databases so we can exercise locking there too.
for { set i 0 } { $i < $ndb } { incr i } {
set cdb [eval {berkdb_open_noerr} -auto_commit \
-env $clientenv $largs testdb$i.db]
set cdblist($i) $cdb
}
# Set up two cursors into each db. Randomly select a cursor
# and do the next thing: position, delete, or close.
foreach option { regular off-page } {
puts "\tRep$tnum.c: Transactional cursor deletes ($option)."
set t [$masterenv txn]
# Set up two cursors into each db, and initialize the next
# action to be done to POSITION.
for { set i 0 } { $i < [expr $ndb * 2] } { incr i } {
set db $dblist([expr $i / 2])
set mcurs($i) [eval {$db cursor} -txn $t]
error_check_good mcurs$i \
[is_valid_cursor $mcurs($i) $db] TRUE
set cnext($i) POSITION
}
set ct [$clientenv txn]
# Set up two cursors into each client db.
for { set i 0 } { $i < [expr $ndb * 2] } { incr i } {
set cdb $cdblist([expr $i / 2])
set ccurs($i) [eval {$cdb cursor} -txn $ct]
error_check_good ccurs$i \
[is_valid_cursor $ccurs($i) $cdb] TRUE
}
# Randomly pick a cursor to operate on and do the next thing.
# At POSITION, we position that cursor. At DELETE, we delete
# the current item. At CLOSE, we close the cursor. At DONE,
# we do nothing except check to see if all cursors have reached
# DONE, and quit when they have.
# On the off-page dup test, walk to reach an off-page entry,
# and delete that one.
set k 0
while { 1 } {
# Every nth time through, update the client.
# set n 5
# if {[expr $k % $n] == 0 } {
# puts "Updating clients"
# process_msgs $envlist
# }
# incr k
set i [berkdb random_int 0 [expr [expr $ndb * 2] - 1]]
set next $cnext($i)
switch -exact -- $next {
POSITION {
do_position $mcurs($i) \
$i $nentries $option
set cnext($i) DELETE
# Position the client cursors too.
do_position $ccurs($i) \
$i $nentries $option
}
DELETE {
error_check_good c_del \
[$mcurs($i) del] 0
set cnext($i) CLOSE
# Update clients after a delete.
process_msgs $envlist
}
CLOSE {
error_check_good c_close.$i \
[$mcurs($i) close] 0
set cnext($i) DONE
# Close the client cursor too.
error_check_good cc_close.$i \
[$ccurs($i) close] 0
}
DONE {
set breakflag 1
for { set j 0 } \
{ $j < [expr $ndb * 2] } \
{ incr j } {
if { $cnext($j) != "DONE" } {
set breakflag 0
}
}
if { $breakflag == 1 } {
break
}
}
default {
puts "FAIL: Unrecognized \
next action $next"
}
}
}
error_check_good txn_commit [$t commit] 0
error_check_good clienttxn_commit [$ct commit] 0
process_msgs $envlist
}
# Clean up.
for { set i 0 } { $i < $ndb } { incr i } {
set db $dblist($i)
error_check_good close_$db [$db close] 0
set cdb $cdblist($i)
error_check_good close_$cdb [$cdb close] 0
}
error_check_good masterenv_close [$masterenv close] 0
error_check_good clientenv_close [$clientenv close] 0
replclose $testdir/MSGQUEUEDIR
return
}
proc do_position { cursor i nentries option } {
if { [expr $i % 2] == 0 } {
if { $option == "regular" } {
set ret [$cursor get -first]
set key [lindex [lindex $ret 0] 0]
set data [lindex [lindex $ret 0] 1]
error_check_good get_first \
[string range $data 4 end] $key
} elseif { $option == "off-page" } {
set ret [$cursor get -set 2]
error_check_good get_key_2 \
[lindex [lindex $ret 0] 0] 2
error_check_good get_data_2 \
[lindex [lindex $ret 0] 1] data2
for { set j 1 } { $j <= 95 } { incr j } {
set ret [$cursor get -nextdup]
error_check_good key_nextdup$j \
[lindex [lindex $ret 0] 0] 2
error_check_good data_nextdup$j \
[lindex [lindex $ret 0] 1] dup$j
}
}
} else {
if { $option == "regular" } {
set ret [$cursor get -set $nentries]
set key [lindex [lindex $ret 0] 0]
set data [lindex [lindex $ret 0] 1]
error_check_good get_set_$nentries \
[string range $data 4 end] $key
} elseif { $option == "off-page" } {
set ret [$cursor get -last]
set key [lindex [lindex $ret 0] 0]
set data [lindex [lindex $ret 0] 1]
error_check_good get_last \
[string range $data 3 end] [expr $key + 1]
for { set j 1 } { $j <= 5 } { incr j } {
set ret [$cursor get -prev]
set key [lindex [lindex $ret 0] 0]
set data [lindex [lindex $ret 0] 1]
error_check_good get_prev \
[string range $data 3 end] \
[expr [expr $key + 1] - $j]
}
}
}
}
|