mirror of
https://github.com/percona/percona-toolkit.git
synced 2025-09-09 07:30:02 +00:00
371 lines
9.7 KiB
Perl
371 lines
9.7 KiB
Perl
#!/usr/bin/env perl
|
|
|
|
BEGIN {
|
|
die "The PERCONA_TOOLKIT_BRANCH environment variable is not set.\n"
|
|
unless $ENV{PERCONA_TOOLKIT_BRANCH} && -d $ENV{PERCONA_TOOLKIT_BRANCH};
|
|
unshift @INC, "$ENV{PERCONA_TOOLKIT_BRANCH}/lib";
|
|
};
|
|
|
|
use strict;
|
|
use warnings FATAL => 'all';
|
|
use English qw(-no_match_vars);
|
|
use Test::More;
|
|
use Data::Dumper;
|
|
|
|
use File::Temp qw(tempfile);
|
|
|
|
use PerconaTest;
|
|
use Sandbox;
|
|
|
|
require "$trunk/bin/pt-heartbeat";
|
|
# Do this after requiring pt-hb, since it uses Mo
|
|
require VersionParser;
|
|
|
|
my $dp = new DSNParser(opts=>$dsn_opts);
|
|
my $sb = new Sandbox(basedir => '/tmp', DSNParser => $dp);
|
|
my $node1 = $sb->get_dbh_for('node1');
|
|
my $node2 = $sb->get_dbh_for('node2');
|
|
my $node3 = $sb->get_dbh_for('node3');
|
|
|
|
if ( !$node1 ) {
|
|
plan skip_all => 'Cannot connect to cluster node1';
|
|
}
|
|
elsif ( !$node2 ) {
|
|
plan skip_all => 'Cannot connect to cluster node2';
|
|
}
|
|
elsif ( !$node3 ) {
|
|
plan skip_all => 'Cannot connect to cluster node3';
|
|
}
|
|
|
|
my $db_flavor = VersionParser->new($node1)->flavor();
|
|
if ( $db_flavor !~ /XtraDB Cluster/ ) {
|
|
plan skip_all => "PXC tests";
|
|
}
|
|
|
|
my $node1_dsn = $sb->dsn_for('node1');
|
|
my $node2_dsn = $sb->dsn_for('node2');
|
|
my $node3_dsn = $sb->dsn_for('node3');
|
|
my $node1_port = $sb->port_for('node1');
|
|
my $node2_port = $sb->port_for('node2');
|
|
my $node3_port = $sb->port_for('node3');
|
|
|
|
my $output;
|
|
my $exit;
|
|
my $base_pidfile = (tempfile("/tmp/pt-heartbeat-test.XXXXXXXX", OPEN => 0, UNLINK => 0))[1];
|
|
my $sample = "t/pt-heartbeat/samples/";
|
|
|
|
diag(`rm -rf /tmp/pt-heartbeat-sentinel >/dev/null 2>&1`);
|
|
$sb->create_dbs($node1, ['test']);
|
|
|
|
my @pidfiles;
|
|
|
|
sub start_update_instance {
|
|
my ($port) = @_;
|
|
my $pidfile = "$base_pidfile.$port.pid";
|
|
push @pidfiles, $pidfile;
|
|
|
|
system("$trunk/bin/pt-heartbeat -h 127.0.0.1 -u msandbox -p msandbox -P $port --database test --table heartbeat --create-table --update --interval 0.5 --daemonize --pid $pidfile >/dev/null");
|
|
|
|
PerconaTest::wait_for_files($pidfile);
|
|
ok(
|
|
-f $pidfile,
|
|
"--update on $port started"
|
|
);
|
|
}
|
|
|
|
foreach my $port ( map { $sb->port_for($_) } qw(node1 node2 node3) ) {
|
|
start_update_instance($port);
|
|
}
|
|
|
|
# #############################################################################
|
|
# Basic cluster tests
|
|
# #############################################################################
|
|
|
|
my $rows = $node1->selectall_hashref("select * from test.heartbeat", 'server_id');
|
|
|
|
is(
|
|
scalar keys %$rows,
|
|
3,
|
|
"Sanity check: All nodes are in the heartbeat table"
|
|
);
|
|
|
|
my $only_slave_data = {
|
|
map {
|
|
$_ => {
|
|
relay_master_log_file => $rows->{$_}->{relay_master_log_file},
|
|
exec_master_log_pos => $rows->{$_}->{exec_master_log_pos},
|
|
} } keys %$rows
|
|
};
|
|
|
|
my $same_data = { relay_master_log_file => undef, exec_master_log_pos => undef };
|
|
is_deeply(
|
|
$only_slave_data,
|
|
{
|
|
12345 => $same_data,
|
|
12346 => $same_data,
|
|
12347 => $same_data,
|
|
},
|
|
"Sanity check: No slave data (relay log or master pos) is stored"
|
|
);
|
|
|
|
$output = output(sub{
|
|
pt_heartbeat::main($node1_dsn, qw(-D test --check)),
|
|
},
|
|
stderr => 1,
|
|
);
|
|
|
|
like(
|
|
$output,
|
|
qr/\QThe --master-server-id option must be specified because the heartbeat table `test`.`heartbeat`/,
|
|
"pt-heartbeat --check + PXC doesn't autodetect a master if there isn't any"
|
|
);
|
|
|
|
$output = output(sub{
|
|
pt_heartbeat::main($node1_dsn, qw(-D test --check),
|
|
'--master-server-id', $node3_port),
|
|
},
|
|
stderr => 1,
|
|
);
|
|
|
|
$output =~ s/\d\.\d{2}/0.00/g;
|
|
is(
|
|
$output,
|
|
"0.00\n",
|
|
"pt-heartbeat --check + PXC works with --master-server-id"
|
|
);
|
|
|
|
# Test --monitor
|
|
|
|
$output = output(sub {
|
|
pt_heartbeat::main($node1_dsn,
|
|
qw(-D test --monitor --run-time 1s),
|
|
'--master-server-id', $node3_port)
|
|
},
|
|
stderr => 1,
|
|
);
|
|
|
|
$output =~ s/\d\.\d{2}/0.00/g;
|
|
|
|
is(
|
|
$output,
|
|
"0.00s [ 0.00s, 0.00s, 0.00s ]\n",
|
|
"--monitor works"
|
|
);
|
|
|
|
# Try to generate some lag between cluster nodes. Rather brittle at the moment.
|
|
|
|
# Lifted from alter active table
|
|
my $pt_osc_sample = "t/pt-online-schema-change/samples";
|
|
|
|
my $query_table_stop = "/tmp/query_table.$PID.stop";
|
|
my $query_table_pid = "/tmp/query_table.$PID.pid";
|
|
my $query_table_output = "/tmp/query_table.$PID.output";
|
|
|
|
$sb->create_dbs($node1, ['pt_osc']);
|
|
$sb->load_file('master', "$pt_osc_sample/basic_no_fks_innodb.sql");
|
|
|
|
$node1->do("USE pt_osc");
|
|
$node1->do("TRUNCATE TABLE t");
|
|
$node1->do("LOAD DATA INFILE '$trunk/$pt_osc_sample/basic_no_fks.data' INTO TABLE t");
|
|
$node1->do("ANALYZE TABLE t");
|
|
$sb->wait_for_slaves();
|
|
|
|
diag(`rm -rf $query_table_stop`);
|
|
diag(`echo > $query_table_output`);
|
|
|
|
my $cmd = "$trunk/$pt_osc_sample/query_table.pl";
|
|
system("$cmd 127.0.0.1 $node1_port pt_osc t id $query_table_stop $query_table_pid >$query_table_output 2>&1 &");
|
|
wait_until(sub{-e $query_table_pid});
|
|
|
|
# Reload sakila
|
|
system "$trunk/sandbox/load-sakila-db $node1_port &";
|
|
|
|
$output = output(sub {
|
|
pt_heartbeat::main($node3_dsn,
|
|
qw(-D test --monitor --run-time 5s),
|
|
'--master-server-id', $node1_port)
|
|
},
|
|
stderr => 1,
|
|
);
|
|
|
|
like(
|
|
$output,
|
|
qr/^(?:0\.(?:\d[1-9]|[1-9]\d)|\d*[1-9]\d*\.\d{2})s\s+\[/,
|
|
"pt-heartbeat can detect replication lag between nodes"
|
|
);
|
|
|
|
diag(`touch $query_table_stop`);
|
|
chomp(my $p = slurp_file($query_table_pid));
|
|
wait_until(sub{!kill 0, $p});
|
|
|
|
$node1->do(q{DROP DATABASE pt_osc});
|
|
|
|
$sb->wait_for_slaves();
|
|
|
|
# #############################################################################
|
|
# cluster, node1 -> slave, run on node1
|
|
# #############################################################################
|
|
|
|
my ($slave_dbh, $slave_dsn) = $sb->start_sandbox(
|
|
server => 'cslave1',
|
|
type => 'slave',
|
|
master => 'node1',
|
|
env => q/BINLOG_FORMAT="ROW"/,
|
|
);
|
|
|
|
$sb->create_dbs($slave_dbh, ['test']);
|
|
|
|
start_update_instance($sb->port_for('cslave1'));
|
|
PerconaTest::wait_for_table($slave_dbh, "test.heartbeat", "1=1");
|
|
|
|
$output = output(sub{
|
|
pt_heartbeat::main($slave_dsn, qw(-D test --check)),
|
|
},
|
|
stderr => 1,
|
|
);
|
|
|
|
like(
|
|
$output,
|
|
qr/\d\.\d{2}\n/,
|
|
"pt-heartbeat --check works on a slave of a cluster node"
|
|
);
|
|
|
|
$output = output(sub {
|
|
pt_heartbeat::main($slave_dsn,
|
|
qw(-D test --monitor --run-time 2s))
|
|
},
|
|
stderr => 1,
|
|
);
|
|
|
|
like(
|
|
$output,
|
|
qr/^\d.\d{2}s\s+\[/,
|
|
"pt-heartbeat --monitor + slave of a node1, without --master-server-id"
|
|
);
|
|
|
|
$output = output(sub {
|
|
pt_heartbeat::main($slave_dsn,
|
|
qw(-D test --monitor --run-time 2s),
|
|
'--master-server-id', $node3_port)
|
|
},
|
|
stderr => 1,
|
|
);
|
|
|
|
like(
|
|
$output,
|
|
qr/^\d.\d{2}s\s+\[/,
|
|
"pt-heartbeat --monitor + slave of node1, --master-server-id pointing to node3"
|
|
);
|
|
|
|
# #############################################################################
|
|
# master -> node1 in cluster
|
|
# #############################################################################
|
|
|
|
# CAREFUL! See the comments in t/pt-table-checksum/pxc.t about cmaster.
|
|
# Nearly everything applies here.
|
|
|
|
my ($master_dbh, $master_dsn) = $sb->start_sandbox(
|
|
server => 'cmaster',
|
|
type => 'master',
|
|
env => q/BINLOG_FORMAT="ROW"/,
|
|
);
|
|
|
|
my $cmaster_port = $sb->port_for('cmaster');
|
|
|
|
$sb->create_dbs($master_dbh, ['test']);
|
|
|
|
$master_dbh->do("FLUSH LOGS");
|
|
$master_dbh->do("RESET MASTER");
|
|
|
|
$sb->set_as_slave('node1', 'cmaster');
|
|
|
|
start_update_instance($sb->port_for('cmaster'));
|
|
PerconaTest::wait_for_table($node1, "test.heartbeat", "server_id=$cmaster_port");
|
|
|
|
$output = output(sub{
|
|
pt_heartbeat::main($node1_dsn, qw(-D test --check --print-master-server-id)),
|
|
},
|
|
stderr => 1,
|
|
);
|
|
|
|
like(
|
|
$output,
|
|
qr/^\d.\d{2} $cmaster_port$/,
|
|
"--print-master-id works for master -> $node1_port, when run from $node1_port"
|
|
);
|
|
|
|
# Wait until node2 & node3 get cmaster in their heartbeat tables
|
|
$sb->wait_for_slaves(master => 'node1', slave => 'node2');
|
|
$sb->wait_for_slaves(master => 'node1', slave => 'node3');
|
|
|
|
foreach my $test (
|
|
[ $node2_port, $node2_dsn, $node2 ],
|
|
[ $node3_port, $node3_dsn, $node3 ],
|
|
) {
|
|
my ($port, $dsn, $dbh) = @$test;
|
|
|
|
$output = output(sub{
|
|
pt_heartbeat::main($dsn, qw(-D test --check --print-master-server-id)),
|
|
},
|
|
stderr => 1,
|
|
);
|
|
|
|
# This could be made to work, see the node autodiscovery branch
|
|
TODO: {
|
|
local $::TODO = "cmaster -> node1, other nodes can't autodetect the master";
|
|
like(
|
|
$output,
|
|
qr/$cmaster_port/,
|
|
"--print-master-id works for master -> $node1_port, when run from $port"
|
|
);
|
|
}
|
|
|
|
$output = output(sub{
|
|
pt_heartbeat::main($dsn, qw(-D test --check --master-server-id), $cmaster_port),
|
|
},
|
|
stderr => 1,
|
|
);
|
|
|
|
$output =~ s/\d\.\d{2}/0.00/g;
|
|
is(
|
|
$output,
|
|
"0.00\n",
|
|
"--check + explicit --master-server-id work for master -> node1, run from $port"
|
|
);
|
|
}
|
|
|
|
# ############################################################################
|
|
# Stop the --update instances.
|
|
# ############################################################################
|
|
diag(`$trunk/bin/pt-heartbeat --stop >/dev/null`);
|
|
sleep 1;
|
|
|
|
foreach my $pidfile (@pidfiles) {
|
|
ok(
|
|
!-f $pidfile,
|
|
"--update on $pidfile stopped"
|
|
);
|
|
}
|
|
|
|
# ############################################################################
|
|
# Disconnect & stop the two servers we started
|
|
# ############################################################################
|
|
|
|
# We have to do this after the --stop, otherwise the --update processes will
|
|
# spew a bunch of warnings and clog
|
|
|
|
$slave_dbh->disconnect;
|
|
$master_dbh->disconnect;
|
|
$sb->stop_sandbox('cslave1', 'cmaster');
|
|
$node1->do("STOP SLAVE");
|
|
$node1->do("RESET SLAVE");
|
|
|
|
# #############################################################################
|
|
# Done.
|
|
# #############################################################################
|
|
$sb->wipe_clean($node1);
|
|
diag(`/tmp/12345/stop`);
|
|
diag(`/tmp/12345/start`);
|
|
ok($sb->ok(), "Sandbox servers") or BAIL_OUT(__FILE__ . " broke the sandbox");
|
|
done_testing;
|