use strict;
use warnings;
use Test::More;
use lib qw(t/lib);
use DBICTest;
BEGIN {
require DBIx::Class;
plan skip_all => 'Test needs ' . DBIx::Class::Optional::Dependencies->req_missing_for ('test_replicated')
unless DBIx::Class::Optional::Dependencies->req_ok_for ('test_replicated');
}
if (DBICTest::RunMode->is_smoker) {
my $mver = Moose->VERSION;
plan skip_all => "A trial version $mver of Moose detected known to break replication - skipping test known to fail"
if ($mver >= 1.99 and $mver <= 1.9902);
}
use Test::Moose;
use Test::Exception;
use List::Util 'first';
use Scalar::Util 'reftype';
use File::Spec;
use IO::Handle;
use Moose();
use MooseX::Types();
note "Using Moose version $Moose::VERSION and MooseX::Types version $MooseX::Types::VERSION";
my $var_dir = quotemeta ( File::Spec->catdir(qw/t var/) );
use_ok 'DBIx::Class::Storage::DBI::Replicated::Pool';
use_ok 'DBIx::Class::Storage::DBI::Replicated::Balancer';
use_ok 'DBIx::Class::Storage::DBI::Replicated::Replicant';
use_ok 'DBIx::Class::Storage::DBI::Replicated';
=head1 HOW TO USE
This is a test of the replicated storage system. This will work in one of
two ways, either it was try to fake replication with a couple of SQLite DBs
and creative use of copy, or if you define a couple of %ENV vars correctly
will try to test those. If you do that, it will assume the setup is properly
replicating. Your results may vary, but I have demonstrated this to work with
mysql native replication.
=cut
## ----------------------------------------------------------------------------
## Build a class to hold all our required testing data and methods.
## ----------------------------------------------------------------------------
TESTSCHEMACLASSES: {
## --------------------------------------------------------------------- ##
## Create an object to contain your replicated stuff.
## --------------------------------------------------------------------- ##
package DBIx::Class::DBI::Replicated::TestReplication;
use DBICTest;
use base qw/Class::Accessor::Fast/;
__PACKAGE__->mk_accessors( qw/schema/ );
## Initialize the object
sub new {
my ($class, $schema_method) = (shift, shift);
my $self = $class->SUPER::new(@_);
$self->schema( $self->init_schema($schema_method) );
return $self;
}
## Get the Schema and set the replication storage type
sub init_schema {
#my ($class, $schema_getter) = @_;
shift->${\ ( 'get_schema_' . shift ) };
}
sub get_schema_by_storage_type {
DBICTest->init_schema(
sqlite_use_file => 1,
storage_type=>{
'::DBI::Replicated' => {
balancer_type=>'::Random',
balancer_args=>{
auto_validate_every=>100,
master_read_weight => 1
},
}
},
deploy_args=>{
add_drop_table => 1,
},
);
}
sub get_schema_by_connect_info {
DBICTest->init_schema(
sqlite_use_file => 1,
storage_type=> '::DBI::Replicated',
balancer_type=>'::Random',
balancer_args=> {
auto_validate_every=>100,
master_read_weight => 1
},
pool_args=>{
maximum_lag=>1,
},
deploy_args=>{
add_drop_table => 1,
},
);
}
sub generate_replicant_connect_info {}
sub replicate {}
sub cleanup {}
## --------------------------------------------------------------------- ##
## Add a connect_info option to test option merging.
## --------------------------------------------------------------------- ##
{
package DBIx::Class::Storage::DBI::Replicated;
use Moose;
__PACKAGE__->meta->make_mutable;
around connect_info => sub {
my ($next, $self, $info) = @_;
$info->[3]{master_option} = 1;
$self->$next($info);
};
__PACKAGE__->meta->make_immutable;
no Moose;
}
## --------------------------------------------------------------------- ##
## Subclass for when you are using SQLite for testing, this provides a fake
## replication support.
## --------------------------------------------------------------------- ##
package DBIx::Class::DBI::Replicated::TestReplication::SQLite;
use DBICTest;
use File::Copy;
use base 'DBIx::Class::DBI::Replicated::TestReplication';
__PACKAGE__->mk_accessors(qw/master_path slave_paths/);
## Set the master path from DBICTest
sub new {
my $class = shift @_;
my $self = $class->SUPER::new(@_);
$self->master_path( DBICTest->_sqlite_dbfilename );
$self->slave_paths([
File::Spec->catfile(qw/t var DBIxClass_slave1.db/),
File::Spec->catfile(qw/t var DBIxClass_slave2.db/),
]);
return $self;
}
## Return an Array of ArrayRefs where each ArrayRef is suitable to use for
## $storage->connect_info to be used for connecting replicants.
sub generate_replicant_connect_info {
my $self = shift @_;
my @dsn = map {
"dbi:SQLite:${_}";
} @{$self->slave_paths};
my @connect_infos = map { [$_,'','',{AutoCommit=>1}] } @dsn;
## Make sure nothing is left over from a failed test
$self->cleanup;
## try a hashref too
my $c = $connect_infos[0];
$connect_infos[0] = {
dsn => $c->[0],
user => $c->[1],
password => $c->[2],
%{ $c->[3] }
};
@connect_infos
}
## Do a 'good enough' replication by copying the master dbfile over each of
## the slave dbfiles. If the master is SQLite we do this, otherwise we
## just do a one second pause to let the slaves catch up.
sub replicate {
my $self = shift @_;
foreach my $slave (@{$self->slave_paths}) {
copy($self->master_path, $slave);
}
}
## Cleanup after ourselves. Unlink all gthe slave paths.
sub cleanup {
my $self = shift @_;
foreach my $slave (@{$self->slave_paths}) {
if(-e $slave) {
unlink $slave;
}
}
}
## --------------------------------------------------------------------- ##
## Subclass for when you are setting the databases via custom export vars
## This is for when you have a replicating database setup that you are
## going to test against. You'll need to define the correct $ENV and have
## two slave databases to test against, as well as a replication system
## that will replicate in less than 1 second.
## --------------------------------------------------------------------- ##
package DBIx::Class::DBI::Replicated::TestReplication::Custom;
use base 'DBIx::Class::DBI::Replicated::TestReplication';
## Return an Array of ArrayRefs where each ArrayRef is suitable to use for
## $storage->connect_info to be used for connecting replicants.
sub generate_replicant_connect_info {
return (
[$ENV{"DBICTEST_SLAVE0_DSN"}, $ENV{"DBICTEST_SLAVE0_DBUSER"}, $ENV{"DBICTEST_SLAVE0_DBPASS"}, {AutoCommit => 1}],
[$ENV{"DBICTEST_SLAVE1_DSN"}, $ENV{"DBICTEST_SLAVE1_DBUSER"}, $ENV{"DBICTEST_SLAVE1_DBPASS"}, {AutoCommit => 1}],
);
}
## pause a bit to let the replication catch up
sub replicate {
sleep 1;
}
}
## ----------------------------------------------------------------------------
## Create an object and run some tests
## ----------------------------------------------------------------------------
## Thi first bunch of tests are basic, just make sure all the bits are behaving
my $replicated_class = DBICTest->has_custom_dsn ?
'DBIx::Class::DBI::Replicated::TestReplication::Custom' :
'DBIx::Class::DBI::Replicated::TestReplication::SQLite';
my $replicated;
for my $method (qw/by_connect_info by_storage_type/) {
undef $replicated;
ok $replicated = $replicated_class->new($method)
=> "Created a replication object $method";
isa_ok $replicated->schema
=> 'DBIx::Class::Schema';
isa_ok $replicated->schema->storage
=> 'DBIx::Class::Storage::DBI::Replicated';
isa_ok $replicated->schema->storage->balancer
=> 'DBIx::Class::Storage::DBI::Replicated::Balancer::Random'
=> 'configured balancer_type';
}
### check that all Storage::DBI methods are handled by ::Replicated
{
my @storage_dbi_methods = Class::MOP::Class
->initialize('DBIx::Class::Storage::DBI')->get_all_method_names;
my @replicated_methods = DBIx::Class::Storage::DBI::Replicated->meta
->get_all_method_names;
# remove constants and OTHER_CRAP
@storage_dbi_methods = grep !/^[A-Z_]+\z/, @storage_dbi_methods;
# remove CAG accessors
@storage_dbi_methods = grep !/_accessor\z/, @storage_dbi_methods;
# remove DBIx::Class (the root parent, with CAG and stuff) methods
my @root_methods = Class::MOP::Class->initialize('DBIx::Class')
->get_all_method_names;
my %count;
$count{$_}++ for (@storage_dbi_methods, @root_methods);
@storage_dbi_methods = grep $count{$_} != 2, @storage_dbi_methods;
# make hashes
my %storage_dbi_methods;
@storage_dbi_methods{@storage_dbi_methods} = ();
my %replicated_methods;
@replicated_methods{@replicated_methods} = ();
# remove ::Replicated-specific methods
for my $method (@replicated_methods) {
delete $replicated_methods{$method}
unless exists $storage_dbi_methods{$method};
}
@replicated_methods = keys %replicated_methods;
# check that what's left is implemented
%count = ();
$count{$_}++ for (@storage_dbi_methods, @replicated_methods);
if ((grep $count{$_} == 2, @storage_dbi_methods) == @storage_dbi_methods) {
pass 'all DBIx::Class::Storage::DBI methods implemented';
}
else {
my @unimplemented = grep $count{$_} == 1, @storage_dbi_methods;
fail 'the following DBIx::Class::Storage::DBI methods are unimplemented: '
. "@unimplemented";
}
}
ok $replicated->schema->storage->meta
=> 'has a meta object';
isa_ok $replicated->schema->storage->master
=> 'DBIx::Class::Storage::DBI';
isa_ok $replicated->schema->storage->pool
=> 'DBIx::Class::Storage::DBI::Replicated::Pool';
does_ok $replicated->schema->storage->balancer
=> 'DBIx::Class::Storage::DBI::Replicated::Balancer';
ok my @replicant_connects = $replicated->generate_replicant_connect_info
=> 'got replication connect information';
ok my @replicated_storages = $replicated->schema->storage->connect_replicants(@replicant_connects)
=> 'Created some storages suitable for replicants';
our %debug;
$replicated->schema->storage->debug(1);
$replicated->schema->storage->debugcb(sub {
my ($op, $info) = @_;
##warn "\n$op, $info\n";
%debug = (
op => $op,
info => $info,
dsn => ($info=~m/\[(.+)\]/)[0],
storage_type => $info=~m/REPLICANT/ ? 'REPLICANT' : 'MASTER',
);
});
ok my @all_storages = $replicated->schema->storage->all_storages
=> '->all_storages';
is scalar @all_storages,
3
=> 'correct number of ->all_storages';
is ((grep $_->isa('DBIx::Class::Storage::DBI'), @all_storages),
3
=> '->all_storages are correct type');
my @all_storage_opts =
grep { (reftype($_)||'') eq 'HASH' }
map @{ $_->_connect_info }, @all_storages;
is ((grep $_->{master_option}, @all_storage_opts),
3
=> 'connect_info was merged from master to replicants');
my @replicant_names = keys %{ $replicated->schema->storage->replicants };
ok @replicant_names, "found replicant names @replicant_names";
## Silence warning about not supporting the is_replicating method if using the
## sqlite dbs.
$replicated->schema->storage->debugobj->silence(1)
if first { $_ =~ /$var_dir/ } @replicant_names;
isa_ok $replicated->schema->storage->balancer->current_replicant
=> 'DBIx::Class::Storage::DBI';
$replicated->schema->storage->debugobj->silence(0);
ok $replicated->schema->storage->pool->has_replicants
=> 'does have replicants';
is $replicated->schema->storage->pool->num_replicants => 2
=> 'has two replicants';
does_ok $replicated_storages[0]
=> 'DBIx::Class::Storage::DBI::Replicated::Replicant';
does_ok $replicated_storages[1]
=> 'DBIx::Class::Storage::DBI::Replicated::Replicant';
does_ok $replicated->schema->storage->replicants->{$replicant_names[0]}
=> 'DBIx::Class::Storage::DBI::Replicated::Replicant';
does_ok $replicated->schema->storage->replicants->{$replicant_names[1]}
=> 'DBIx::Class::Storage::DBI::Replicated::Replicant';
## Add some info to the database
$replicated
->schema
->populate('Artist', [
[ qw/artistid name/ ],
[ 4, "Ozric Tentacles"],
]);
is $debug{storage_type}, 'MASTER',
"got last query from a master: $debug{dsn}";
like $debug{info}, qr/INSERT/, 'Last was an insert';
## Make sure all the slaves have the table definitions
$replicated->replicate;
$replicated->schema->storage->replicants->{$replicant_names[0]}->active(1);
$replicated->schema->storage->replicants->{$replicant_names[1]}->active(1);
## Silence warning about not supporting the is_replicating method if using the
## sqlite dbs.
$replicated->schema->storage->debugobj->silence(1)
if first { $_ =~ /$var_dir/ } @replicant_names;
$replicated->schema->storage->pool->validate_replicants;
$replicated->schema->storage->debugobj->silence(0);
## Make sure we can read the data.
ok my $artist1 = $replicated->schema->resultset('Artist')->find(4)
=> 'Created Result';
## We removed testing here since master read weight is on, so we can't tell in
## advance what storage to expect. We turn master read weight off a bit lower
## is $debug{storage_type}, 'REPLICANT'
## => "got last query from a replicant: $debug{dsn}, $debug{info}";
isa_ok $artist1
=> 'DBICTest::Artist';
is $artist1->name, 'Ozric Tentacles'
=> 'Found expected name for first result';
## Check that master_read_weight is honored
{
no warnings qw/once redefine/;
local
*DBIx::Class::Storage::DBI::Replicated::Balancer::Random::_random_number =
sub { 999 };
$replicated->schema->storage->balancer->increment_storage;
is $replicated->schema->storage->balancer->current_replicant,
$replicated->schema->storage->master
=> 'master_read_weight is honored';
## turn it off for the duration of the test
$replicated->schema->storage->balancer->master_read_weight(0);
$replicated->schema->storage->balancer->increment_storage;
}
## Add some new rows that only the master will have This is because
## we overload any type of write operation so that is must hit the master
## database.
$replicated
->schema
->populate('Artist', [
[ qw/artistid name/ ],
[ 5, "Doom's Children"],
[ 6, "Dead On Arrival"],
[ 7, "Watergate"],
]);
is $debug{storage_type}, 'MASTER',
"got last query from a master: $debug{dsn}";
like $debug{info}, qr/INSERT/, 'Last was an insert';
## Make sure all the slaves have the table definitions
$replicated->replicate;
## Should find some data now
ok my $artist2 = $replicated->schema->resultset('Artist')->find(5)
=> 'Sync succeed';
is $debug{storage_type}, 'REPLICANT'
=> "got last query from a replicant: $debug{dsn}";
isa_ok $artist2
=> 'DBICTest::Artist';
is $artist2->name, "Doom's Children"
=> 'Found expected name for first result';
## What happens when we disconnect all the replicants?
is $replicated->schema->storage->pool->connected_replicants => 2
=> "both replicants are connected";
$replicated->schema->storage->replicants->{$replicant_names[0]}->disconnect;
$replicated->schema->storage->replicants->{$replicant_names[1]}->disconnect;
is $replicated->schema->storage->pool->connected_replicants => 0
=> "both replicants are now disconnected";
## All these should pass, since the database should automatically reconnect
ok my $artist3 = $replicated->schema->resultset('Artist')->find(6)
=> 'Still finding stuff.';
is $debug{storage_type}, 'REPLICANT'
=> "got last query from a replicant: $debug{dsn}";
isa_ok $artist3
=> 'DBICTest::Artist';
is $artist3->name, "Dead On Arrival"
=> 'Found expected name for first result';
is $replicated->schema->storage->pool->connected_replicants => 1
=> "At Least One replicant reconnected to handle the job";
## What happens when we try to select something that doesn't exist?
ok ! $replicated->schema->resultset('Artist')->find(666)
=> 'Correctly failed to find something.';
is $debug{storage_type}, 'REPLICANT'
=> "got last query from a replicant: $debug{dsn}";
## test the reliable option
TESTRELIABLE: {
$replicated->schema->storage->set_reliable_storage;
ok $replicated->schema->resultset('Artist')->find(2)
=> 'Read from master 1';
is $debug{storage_type}, 'MASTER',
"got last query from a master: $debug{dsn}";
ok $replicated->schema->resultset('Artist')->find(5)
=> 'Read from master 2';
is $debug{storage_type}, 'MASTER',
"got last query from a master: $debug{dsn}";
$replicated->schema->storage->set_balanced_storage;
ok $replicated->schema->resultset('Artist')->find(3)
=> 'Read from replicant';
is $debug{storage_type}, 'REPLICANT',
"got last query from a replicant: $debug{dsn}";
}
## Make sure when reliable goes out of scope, we are using replicants again
ok $replicated->schema->resultset('Artist')->find(1)
=> 'back to replicant 1.';
is $debug{storage_type}, 'REPLICANT',
"got last query from a replicant: $debug{dsn}";
ok $replicated->schema->resultset('Artist')->find(2)
=> 'back to replicant 2.';
is $debug{storage_type}, 'REPLICANT',
"got last query from a replicant: $debug{dsn}";
## set all the replicants to inactive, and make sure the balancer falls back to
## the master.
$replicated->schema->storage->replicants->{$replicant_names[0]}->active(0);
$replicated->schema->storage->replicants->{$replicant_names[1]}->active(0);
{
## catch the fallback to master warning
open my $debugfh, '>', \my $fallback_warning;
my $oldfh = $replicated->schema->storage->debugfh;
$replicated->schema->storage->debugfh($debugfh);
ok $replicated->schema->resultset('Artist')->find(2)
=> 'Fallback to master';
is $debug{storage_type}, 'MASTER',
"got last query from a master: $debug{dsn}";
like $fallback_warning, qr/falling back to master/
=> 'emits falling back to master debug';
$replicated->schema->storage->debugfh($oldfh);
}
$replicated->schema->storage->replicants->{$replicant_names[0]}->active(1);
$replicated->schema->storage->replicants->{$replicant_names[1]}->active(1);
## Silence warning about not supporting the is_replicating method if using the
## sqlite dbs.
$replicated->schema->storage->debugobj->silence(1)
if first { $_ =~ /$var_dir/ } @replicant_names;
$replicated->schema->storage->pool->validate_replicants;
$replicated->schema->storage->debugobj->silence(0);
{
## catch the fallback to master warning
open my $debugfh, '>', \my $return_warning;
my $oldfh = $replicated->schema->storage->debugfh;
$replicated->schema->storage->debugfh($debugfh);
ok $replicated->schema->resultset('Artist')->find(2)
=> 'Return to replicants';
is $debug{storage_type}, 'REPLICANT',
"got last query from a replicant: $debug{dsn}";
like $return_warning, qr/Moved back to slave/
=> 'emits returning to slave debug';
$replicated->schema->storage->debugfh($oldfh);
}
## Getting slave status tests
SKIP: {
## We skip this tests unless you have a custom replicants, since the default
## sqlite based replication tests don't support these functions.
skip 'Cannot Test Replicant Status on Non Replicating Database', 10
unless DBICTest->has_custom_dsn && $ENV{"DBICTEST_SLAVE0_DSN"};
$replicated->replicate; ## Give the slaves a chance to catchup.
ok $replicated->schema->storage->replicants->{$replicant_names[0]}->is_replicating
=> 'Replicants are replicating';
is $replicated->schema->storage->replicants->{$replicant_names[0]}->lag_behind_master, 0
=> 'Replicant is zero seconds behind master';
## Test the validate replicants
$replicated->schema->storage->pool->validate_replicants;
is $replicated->schema->storage->pool->active_replicants, 2
=> 'Still have 2 replicants after validation';
## Force the replicants to fail the validate test by required their lag to
## be negative (ie ahead of the master!)
$replicated->schema->storage->pool->maximum_lag(-10);
$replicated->schema->storage->pool->validate_replicants;
is $replicated->schema->storage->pool->active_replicants, 0
=> 'No way a replicant be be ahead of the master';
## Let's be fair to the replicants again. Let them lag up to 5
$replicated->schema->storage->pool->maximum_lag(5);
$replicated->schema->storage->pool->validate_replicants;
is $replicated->schema->storage->pool->active_replicants, 2
=> 'Both replicants in good standing again';
## Check auto validate
is $replicated->schema->storage->balancer->auto_validate_every, 100
=> "Got the expected value for auto validate";
## This will make sure we auto validatge everytime
$replicated->schema->storage->balancer->auto_validate_every(0);
## set all the replicants to inactive, and make sure the balancer falls back to
## the master.
$replicated->schema->storage->replicants->{$replicant_names[0]}->active(0);
$replicated->schema->storage->replicants->{$replicant_names[1]}->active(0);
## Ok, now when we go to run a query, autovalidate SHOULD reconnect
is $replicated->schema->storage->pool->active_replicants => 0
=> "both replicants turned off";
ok $replicated->schema->resultset('Artist')->find(5)
=> 'replicant reactivated';
is $debug{storage_type}, 'REPLICANT',
"got last query from a replicant: $debug{dsn}";
is $replicated->schema->storage->pool->active_replicants => 2
=> "both replicants reactivated";
}
## Test the reliably callback
ok my $reliably = sub {
ok $replicated->schema->resultset('Artist')->find(5)
=> 'replicant reactivated';
is $debug{storage_type}, 'MASTER',
"got last query from a master: $debug{dsn}";
} => 'created coderef properly';
$replicated->schema->storage->execute_reliably($reliably);
## Try something with an error
ok my $unreliably = sub {
ok $replicated->schema->resultset('ArtistXX')->find(5)
=> 'replicant reactivated';
} => 'created coderef properly';
throws_ok {$replicated->schema->storage->execute_reliably($unreliably)}
qr/Can't find source for ArtistXX/
=> 'Bad coderef throws proper error';
## Make sure replication came back
ok $replicated->schema->resultset('Artist')->find(3)
=> 'replicant reactivated';
is $debug{storage_type}, 'REPLICANT', "got last query from a replicant: $debug{dsn}";
## make sure transactions are set to execute_reliably
ok my $transaction = sub {
my $id = shift @_;
$replicated
->schema
->populate('Artist', [
[ qw/artistid name/ ],
[ $id, "Children of the Grave $id"],
]);
ok my $result = $replicated->schema->resultset('Artist')->find($id)
=> "Found expected artist for $id";
is $debug{storage_type}, 'MASTER',
"got last query from a master: $debug{dsn}";
ok my $more = $replicated->schema->resultset('Artist')->find(1)
=> 'Found expected artist again for 1';
is $debug{storage_type}, 'MASTER',
"got last query from a master: $debug{dsn}";
return ($result, $more);
} => 'Created a coderef properly';
## Test the transaction with multi return
{
ok my @return = $replicated->schema->txn_do($transaction, 666)
=> 'did transaction';
is $return[0]->id, 666
=> 'first returned value is correct';
is $debug{storage_type}, 'MASTER',
"got last query from a master: $debug{dsn}";
is $return[1]->id, 1
=> 'second returned value is correct';
is $debug{storage_type}, 'MASTER',
"got last query from a master: $debug{dsn}";
}
## Test that asking for single return works
{
ok my @return = $replicated->schema->txn_do($transaction, 777)
=> 'did transaction';
is $return[0]->id, 777
=> 'first returned value is correct';
is $return[1]->id, 1
=> 'second returned value is correct';
}
## Test transaction returning a single value
{
ok my $result = $replicated->schema->txn_do(sub {
ok my $more = $replicated->schema->resultset('Artist')->find(1)
=> 'found inside a transaction';
is $debug{storage_type}, 'MASTER', "got last query from a master: $debug{dsn}";
return $more;
}) => 'successfully processed transaction';
is $result->id, 1
=> 'Got expected single result from transaction';
}
## Make sure replication came back
ok $replicated->schema->resultset('Artist')->find(1)
=> 'replicant reactivated';
is $debug{storage_type}, 'REPLICANT', "got last query from a replicant: $debug{dsn}";
## Test Discard changes
{
ok my $artist = $replicated->schema->resultset('Artist')->find(2)
=> 'got an artist to test discard changes';
is $debug{storage_type}, 'REPLICANT', "got last query from a replicant: $debug{dsn}";
ok $artist->get_from_storage({force_pool=>'master'})
=> 'properly discard changes';
is $debug{storage_type}, 'MASTER', "got last query from a master: $debug{dsn}";
ok $artist->discard_changes({force_pool=>'master'})
=> 'properly called discard_changes against master (manual attrs)';
is $debug{storage_type}, 'MASTER', "got last query from a master: $debug{dsn}";
ok $artist->discard_changes()
=> 'properly called discard_changes against master (default attrs)';
is $debug{storage_type}, 'MASTER', "got last query from a master: $debug{dsn}";
ok $artist->discard_changes({force_pool=>$replicant_names[0]})
=> 'properly able to override the default attributes';
is $debug{storage_type}, 'REPLICANT', "got last query from a replicant: $debug{dsn}"
}
## Test some edge cases, like trying to do a transaction inside a transaction, etc
{
ok my $result = $replicated->schema->txn_do(sub {
return $replicated->schema->txn_do(sub {
ok my $more = $replicated->schema->resultset('Artist')->find(1)
=> 'found inside a transaction inside a transaction';
is $debug{storage_type}, 'MASTER', "got last query from a master: $debug{dsn}";
return $more;
});
}) => 'successfully processed transaction';
is $result->id, 1
=> 'Got expected single result from transaction';
}
{
ok my $result = $replicated->schema->txn_do(sub {
return $replicated->schema->storage->execute_reliably(sub {
return $replicated->schema->txn_do(sub {
return $replicated->schema->storage->execute_reliably(sub {
ok my $more = $replicated->schema->resultset('Artist')->find(1)
=> 'found inside crazy deep transactions and execute_reliably';
is $debug{storage_type}, 'MASTER', "got last query from a master: $debug{dsn}";
return $more;
});
});
});
}) => 'successfully processed transaction';
is $result->id, 1
=> 'Got expected single result from transaction';
}
## Test the force_pool resultset attribute.
{
ok my $artist_rs = $replicated->schema->resultset('Artist')
=> 'got artist resultset';
## Turn on Forced Pool Storage
ok my $reliable_artist_rs = $artist_rs->search(undef, {force_pool=>'master'})
=> 'Created a resultset using force_pool storage';
ok my $artist = $reliable_artist_rs->find(2)
=> 'got an artist result via force_pool storage';
is $debug{storage_type}, 'MASTER', "got last query from a master: $debug{dsn}";
}
## Test the force_pool resultset attribute part two.
{
ok my $artist_rs = $replicated->schema->resultset('Artist')
=> 'got artist resultset';
## Turn on Forced Pool Storage
ok my $reliable_artist_rs = $artist_rs->search(undef, {force_pool=>$replicant_names[0]})
=> 'Created a resultset using force_pool storage';
ok my $artist = $reliable_artist_rs->find(2)
=> 'got an artist result via force_pool storage';
is $debug{storage_type}, 'REPLICANT', "got last query from a replicant: $debug{dsn}";
}
## Delete the old database files
$replicated->cleanup;
done_testing;
# vim: sw=4 sts=4 :