summaryrefslogtreecommitdiff
path: root/ctdb/tests/INTEGRATION/database/recovery.002.large.sh
blob: 4f544fa8475bca39412e89659bf214fbff8dbaec (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
#!/bin/bash

test_info()
{
    cat <<EOF
Older style of recovery using PULL_DB and PUSH_DB controls tries to
construct a single large marshall buffer for all the records in the
database.  However, this approach is problematic as talloc restricts the
maximum size of buffer to 256M.  Also, trying to construct and send large
buffers is inefficient and can cause CTDB daemon to be tied up for long
periods of time.

Instead new style recovery is introduced using DB_PULL and
DB_PUSH_START/DB_PUSH_CONFIRM controls.  This sends the records in
batches of ~RecBufferSizeLimit in size at a time.

Expected results:

* The recovery should complete successfully

EOF
}

. "${TEST_SCRIPTS_DIR}/integration.bash"

ctdb_test_init

set -e

cluster_is_healthy

#
# Main test
#
TEST1DB="large_persistent_db.tdb"
TEST2DB="large_volatile_db.tdb"
RECDATA=$(onnode 0 mktemp)

# Create a persistent database to test
echo "create persistent test database $TEST1DB"
try_command_on_node 0 $CTDB attach $TEST1DB persistent

# Wipe Test database
echo "wipe test database $TEST1DB"
try_command_on_node 0 $CTDB wipedb $TEST1DB

# Create dummy record data
echo "creating dummy record data"
onnode 0 dd if=/dev/urandom of=$RECDATA bs=10K count=1

# Add 345 records
echo "Adding 345 records"
for i in $(seq 1 345) ; do
    try_command_on_node 0 $CTDB pstore $TEST1DB record$i $RECDATA || exit 1
done

num_records=$(db_ctdb_cattdb_count_records 0 $TEST1DB)
if [ $num_records = "345" ] ; then
	echo "OK: records added correctly"
else
	echo "BAD: persistent database has $num_records of 345 records"
	try_command_on_node -v 0 "$CTDB cattdb $TEST1DB | tail -n 1"
	exit 1
fi

# Create a volatile database to test
echo "create volatile test database $TEST2DB"
try_command_on_node 0 $CTDB attach $TEST2DB

# Wipe Test database
echo "wipe test database $TEST2DB"
try_command_on_node 0 $CTDB wipedb $TEST2DB

# Create dummy record data
v1="1234567890"
v2="$v1$v1$v1$v1$v1$v1$v1$v1$v1$v1"
v3="$v2$v2$v2$v2$v2$v2$v2$v2$v2$v2"

# Add 1234 records
echo "Adding 1234 records"
for i in $(seq 1 1234) ; do
    try_command_on_node 0 $CTDB writekey $TEST2DB record$i $v3 || exit 1
done

num_records=$(db_ctdb_cattdb_count_records 0 $TEST2DB)
if [ $num_records = "1234" ] ; then
	echo "OK: records added correctly"
else
	echo "BAD: volatile database has $num_records of 1234 records"
	try_command_on_node -v 0 "$CTDB cattdb $TEST2DB | tail -n 1"
	exit 1
fi

echo "Find out which node is recmaster"
try_command_on_node 0 $CTDB recmaster
recmaster="$out"

# Set RecBufferSizeLimit to 10000
try_command_on_node $recmaster $CTDB setvar RecBufferSizeLimit 10000

# Do a recovery
echo "force recovery"
try_command_on_node 0 $CTDB recover

wait_until_node_has_status 0 recovered 30

# check that there are correct number of records
num_records=$(db_ctdb_cattdb_count_records 0 $TEST1DB)
if [ $num_records = "345" ] ; then
	echo "OK: persistent database recovered correctly"
else
	echo "BAD: persistent database has $num_records of 345 records"
	try_command_on_node -v 0 "$CTDB cattdb $TEST1DB | tail -n 1"
	exit 1
fi

num_records=$(db_ctdb_cattdb_count_records 0 $TEST2DB)
if [ $num_records = "1234" ] ; then
	echo "OK: volatile database recovered correctly"
else
	echo "BAD: volatile database has $num_records of 1234 records"
	try_command_on_node -v 0 "$CTDB cattdb $TEST2DB | tail -n 1"
	exit 1
fi