libdb/test/tcl/test097.tcl

200 lines
5.3 KiB
Tcl
Raw Normal View History

2011-09-13 17:44:24 +00:00
# See the file LICENSE for redistribution information.
#
# Copyright (c) 1996, 2011 Oracle and/or its affiliates. All rights reserved.
#
# $Id$
#
# TEST test097
# TEST Open up a large set of database files simultaneously.
# TEST Adjust for local file descriptor resource limits.
# TEST Then use the first 1000 entries from the dictionary.
# TEST Insert each with self as key and a fixed, medium length data string;
# TEST retrieve each. After all are entered, retrieve all; compare output
# TEST to original.
proc test097 { method {ndbs 500} {nentries 400} args } {
global pad_datastr
source ./include.tcl
set largs [convert_args $method $args]
set encargs ""
set largs [split_encargs $largs encargs]
# Open an environment, with a 1MB cache.
set eindex [lsearch -exact $largs "-env"]
if { $eindex != -1 } {
incr eindex
set env [lindex $largs $eindex]
puts "Test097: $method: skipping for env $env"
return
}
env_cleanup $testdir
set env [eval {berkdb_env -create -log_regionmax 256000 \
-pagesize 512 -cachesize { 0 1048576 1 } -txn} \
-home $testdir $encargs]
error_check_good dbenv [is_valid_env $env] TRUE
if { [is_partitioned $args] == 1 } {
set ndbs [expr $ndbs / 10]
}
# heap opens 2 other files under the covers, so limit dbs for heap
if { [is_heap $method] == 1 } {
if { $ndbs > 325 } {
set ndbs 325
}
}
# Create the database and open the dictionary
set basename test097
set t1 $testdir/t1
set t2 $testdir/t2
set t3 $testdir/t3
#
# When running with HAVE_MUTEX_SYSTEM_RESOURCES,
# we can run out of mutex lock slots due to the nature of this test.
# So, for this test, increase the number of pages per extent
# to consume fewer resources.
#
if { [is_queueext $method] } {
set numdb [expr $ndbs / 4]
set eindex [lsearch -exact $largs "-extent"]
error_check_bad extent $eindex -1
incr eindex
set extval [lindex $largs $eindex]
set extval [expr $extval * 4]
set largs [lreplace $largs $eindex $eindex $extval]
}
puts -nonewline "Test097: $method ($largs) "
puts "$nentries entries in at most $ndbs simultaneous databases"
puts "\tTest097.a: Simultaneous open"
set numdb [test097_open tdb $ndbs $method $env $basename $largs]
if { $numdb == 0 } {
puts "\tTest097: Insufficient resources available -- skipping."
error_check_good envclose [$env close] 0
return
}
set did [open $dict]
set pflags ""
set gflags ""
set txn ""
set count 0
# Here is the loop where we put and get each key/data pair
if { [is_record_based $method] == 1 } {
append gflags "-recno"
}
puts "\tTest097.b: put/get on $numdb databases"
set datastr "abcdefghij"
set pad_datastr [pad_data $method $datastr]
while { [gets $did str] != -1 && $count < $nentries } {
if { [is_record_based $method] == 1 } {
set key [expr $count + 1]
} else {
set key $str
}
for { set i 1 } { $i <= $numdb } { incr i } {
set ret [eval {$tdb($i) put} $txn $pflags \
{$key [chop_data $method $datastr]}]
error_check_good put $ret 0
set ret [eval {$tdb($i) get} $gflags {$key}]
error_check_good get $ret [list [list $key \
[pad_data $method $datastr]]]
}
incr count
}
close $did
# Now we will get each key from the DB and compare the results
# to the original.
puts "\tTest097.c: dump and check files"
for { set j 1 } { $j <= $numdb } { incr j } {
dump_file $tdb($j) $txn $t1 test097.check
error_check_good db_close [$tdb($j) close] 0
# Now compare the keys to see if they match the dictionary
if { [is_record_based $method] == 1 } {
set oid [open $t2 w]
for {set i 1} {$i <= $nentries} {set i [incr i]} {
puts $oid $i
}
close $oid
filesort $t2 $t3
file rename -force $t3 $t2
} else {
set q q
filehead $nentries $dict $t3
filesort $t3 $t2
}
filesort $t1 $t3
error_check_good Test097:diff($t3,$t2) [filecmp $t3 $t2] 0
}
error_check_good envclose [$env close] 0
}
# Check function for test097; data should be fixed are identical
proc test097.check { key data } {
global pad_datastr
error_check_good "data mismatch for key $key" $data $pad_datastr
}
proc test097_open { tdb ndbs method env basename largs } {
global errorCode
upvar $tdb db
set j 0
set numdb $ndbs
if { [is_queueext $method] } {
set numdb [expr $ndbs / 4]
}
set omethod [convert_method $method]
for { set i 1 } {$i <= $numdb } { incr i } {
set stat [catch {eval {berkdb_open -env $env \
-pagesize 512 -create -mode 0644} \
$largs {$omethod $basename.$i.db}} db($i)]
#
# Check if we've reached our limit
#
if { $stat == 1 } {
set min 20
set em [is_substr $errorCode EMFILE]
set en [is_substr $errorCode ENFILE]
error_check_good open_ret [expr $em || $en] 1
puts \
"\tTest097.a.1 Encountered resource limits opening $i files, adjusting"
if { [is_queueext $method] } {
set end [expr $j / 4]
set min 10
} else {
set end [expr $j - 10]
}
#
# If we cannot open even $min files, then this test is
# not very useful. Close up shop and go back.
#
if { $end < $min } {
test097_close db 1 $j
return 0
}
test097_close db [expr $end + 1] $j
return $end
} else {
error_check_good dbopen [is_valid_db $db($i)] TRUE
set j $i
}
}
return $j
}
proc test097_close { tdb start end } {
upvar $tdb db
for { set i $start } { $i <= $end } { incr i } {
error_check_good db($i)close [$db($i) close] 0
}
}