libdb/test/tcl/rep098.tcl

274 lines
8.1 KiB
Tcl
Raw Normal View History

2011-09-13 17:44:24 +00:00
# See the file LICENSE for redistribution information.
#
# Copyright (c) 2011 Oracle and/or its affiliates. All rights reserved.
#
# $Id$
#
# TEST rep098
# TEST Test of internal initialization and page deallocation.
# TEST
# TEST Use one master, one client.
# TEST Generate several log files.
# TEST Remove old master log files.
# TEST Start a client.
# TEST After client gets UPDATE file information, delete entries to
# TEST remove pages in the database.
#
proc rep098 { method { niter 200 } { tnum "098" } args } {
source ./include.tcl
global databases_in_memory
global repfiles_in_memory
# Run for btree and queue methods only.
if { $checking_valid_methods } {
set test_methods {}
foreach method $valid_methods {
if { [is_btree $method] == 1 || \
[is_queue $method] == 1 } {
lappend test_methods $method
}
}
return $test_methods
}
if { [is_btree $method] == 0 && [is_queue $method] == 0 } {
puts "Rep$tnum: skipping for non-btree, non-queue method."
return
}
set args [convert_args $method $args]
# This test needs to set its own pagesize.
set pgindex [lsearch -exact $args "-pagesize"]
if { $pgindex != -1 } {
puts "Rep$tnum: skipping for specific pagesizes"
return
}
set logsets [create_logsets 2]
# Set up for on-disk or in-memory databases.
set msg "using on-disk databases"
if { $databases_in_memory } {
set msg "using named in-memory databases"
if { [is_queueext $method] } {
puts -nonewline "Skipping rep$tnum for method "
puts "$method with named in-memory databases."
return
}
}
set msg2 "and on-disk replication files"
if { $repfiles_in_memory } {
set msg2 "and in-memory replication files"
}
# Run the body of the test with and without recovery,
# Skip recovery with in-memory logging - it doesn't make sense.
foreach r $test_recopts {
foreach l $logsets {
set logindex [lsearch -exact $l "in-memory"]
if { $r == "-recover" && $logindex != -1 } {
puts "Skipping rep$tnum for -recover\
with in-memory logs."
continue
}
puts "Rep$tnum ($method $r $args): Test of\
internal init with page deallocation $msg $msg2."
puts "Rep$tnum: Master logs are [lindex $l 0]"
puts "Rep$tnum: Client logs are [lindex $l 1]"
rep098_sub $method $niter $tnum $l $r $args
}
}
}
proc rep098_sub { method niter tnum logset recargs largs } {
global databases_in_memory
global repfiles_in_memory
global rep_verbose
global testdir
global util_path
global verbose_type
set verbargs ""
if { $rep_verbose == 1 } {
set verbargs " -verbose {$verbose_type on} "
}
set repmemargs ""
if { $repfiles_in_memory } {
set repmemargs "-rep_inmem_files "
}
env_cleanup $testdir
replsetup $testdir/MSGQUEUEDIR
set masterdir $testdir/MASTERDIR
set clientdir $testdir/CLIENTDIR
file mkdir $masterdir
file mkdir $clientdir
# Log size is small so we quickly create more than one.
# The documentation says that the log file must be at least
# four times the size of the in-memory log buffer.
set pagesize 4096
append largs " -pagesize $pagesize "
set log_max [expr $pagesize * 8]
set m_logtype [lindex $logset 0]
set c_logtype [lindex $logset 1]
# In-memory logs cannot be used with -txn nosync.
set m_logargs [adjust_logargs $m_logtype]
set c_logargs [adjust_logargs $c_logtype]
set m_txnargs [adjust_txnargs $m_logtype]
set c_txnargs [adjust_txnargs $c_logtype]
# Open a master.
repladd 1
set ma_envcmd "berkdb_env_noerr -create $m_txnargs $repmemargs \
$m_logargs -log_max $log_max -errpfx MASTER $verbargs \
-lock_max_locks 20000 \
-home $masterdir -rep_transport \[list 1 replsend\]"
set masterenv [eval $ma_envcmd $recargs -rep_master]
# Run rep_test in the master only.
puts "\tRep$tnum.a: Running rep_test in replicated env."
set start 0
if { $databases_in_memory } {
set testfile { "" "test.db" }
set bigfile { "" "big.db" }
} else {
set testfile "test.db"
set bigfile "big.db"
}
set omethod [convert_method $method]
set dbargs [convert_args $method $largs]
set mdb [eval {berkdb_open} -env $masterenv -auto_commit\
-create -mode 0644 $omethod $dbargs $testfile ]
error_check_good reptest_db [is_valid_db $mdb] TRUE
#
# Create a database with lots of items forcing lots of pages.
# We want two databases so that we
#
set bigdb [eval {berkdb_open} -env $masterenv -auto_commit\
-create -mode 0644 $omethod $dbargs $bigfile ]
error_check_good reptest_db [is_valid_db $bigdb] TRUE
set stop 0
set bigniter [expr $niter * 10]
while { $stop == 0 } {
eval rep_test \
$method $masterenv $mdb $niter $start $start 0 $largs
incr start $niter
eval rep_test \
$method $masterenv $bigdb $bigniter $start $start 0 $largs
incr start $bigniter
puts "\tRep$tnum.a.1: Run db_archive on master."
if { $m_logtype == "on-disk" } {
$masterenv log_flush
eval exec $util_path/db_archive -d -h $masterdir
}
#
# Make sure we have moved beyond the first log file.
#
set first_master_log [get_logfile $masterenv first]
if { $first_master_log > 1 } {
set stop 1
}
}
puts "\tRep$tnum.b: Open client."
repladd 2
set cl_envcmd "berkdb_env_noerr -create $c_txnargs $repmemargs \
$c_logargs -log_max $log_max -errpfx CLIENT $verbargs \
-home $clientdir -rep_transport \[list 2 replsend\]"
set clientenv [eval $cl_envcmd $recargs -rep_client]
set envlist "{$masterenv 1} {$clientenv 2}"
#
# Process messages until the client gets into the SYNC_PAGE state.
# We want to stop once we have the file information telling us
# how many pages there are, but before # we ask for the pages.
# Then we can remove some of the pages the master said were there.
#
set stop 0
while { $stop == 0 } {
set nproced 0
incr nproced [proc_msgs_once $envlist NONE err]
error_check_bad nproced $nproced 0
set clstat [exec $util_path/db_stat \
-N -r -R A -h $clientdir]
if { [is_substr $clstat "SYNC_PAGE"] } {
set stop 1
}
}
#
# At this point the client has all the file info about both
# databases. Now let's remove pages from big.db.
#
if { [is_queue $method] == 0 } {
set consume ""
set statfld "Page count"
} else {
set consume "-consume"
set statfld "Number of pages"
}
set pg1 [stat_field $bigdb stat $statfld]
set txn [$masterenv txn]
set dbc [$bigdb cursor -txn $txn]
#
# Note, we get the last item and then move to the prev so that
# we remove all but the last item. We use the last item instead
# of the first so that the queue head moves.
#
set kd [$dbc get -last]
for { set kd [$dbc get -prev] } { [llength $kd] > 0 } \
{ set kd [$dbc get -prev] } {
error_check_good del_item [eval {$dbc del} $consume] 0
}
error_check_good dbc_close [$dbc close] 0
if { [is_queue $method] == 0 } {
set stat [catch {eval $bigdb compact -txn $txn -freespace} ret]
error_check_good compact $stat 0
}
error_check_good txn_commit [$txn commit] 0
$bigdb sync
set pg2 [stat_field $bigdb stat $statfld]
error_check_good pgcnt [expr $pg2 < $pg1] 1
puts "\tRep$tnum.c: Process msgs after page count reduction $pg1 to $pg2"
#
# Now that we've removed pages, let the init complete.
#
process_msgs $envlist
puts "\tRep$tnum.d: Verify logs and databases"
set cdb [eval {berkdb_open_noerr} -env $clientenv -auto_commit\
-create -mode 0644 $omethod $dbargs $testfile]
error_check_good reptest_db [is_valid_db $cdb] TRUE
set cbdb [eval {berkdb_open_noerr} -env $clientenv -auto_commit\
-create -mode 0644 $omethod $dbargs $bigfile]
error_check_good reptest_db [is_valid_db $cdb] TRUE
if { $databases_in_memory } {
rep_verify_inmem $masterenv $clientenv $mdb $cdb
rep_verify_inmem $masterenv $clientenv $bigdb $cbdb
} else {
rep_verify $masterdir $masterenv $clientdir $clientenv 1
}
# Make sure log files are on-disk or not as expected.
check_log_location $masterenv
check_log_location $clientenv
error_check_good mdb_close [$mdb close] 0
error_check_good mdb_close [$bigdb close] 0
error_check_good cdb_close [$cdb close] 0
error_check_good cdb_close [$cbdb close] 0
error_check_good masterenv_close [$masterenv close] 0
error_check_good clientenv_close [$clientenv close] 0
replclose $testdir/MSGQUEUEDIR
}