# work around brain damage in PPerl (yes, it has to be a global)
$SIG{__WARN__} = sub {
warn @_ unless $_[0] =~ /\QUse of "goto" to jump into a construct is deprecated/
} if ($ENV{DBICTEST_IN_PERSISTENT_ENV});
# the persistent environments run with this flag first to see if
# we will run at all (e.g. it will fail if $^X doesn't match)
exit 0 if $ENV{DBICTEST_PERSISTENT_ENV_BAIL_EARLY};
# Do the override as early as possible so that CORE::bless doesn't get compiled away
# We will replace $bless_override only if we are in author mode
my $bless_override;
BEGIN {
$bless_override = sub {
CORE::bless( $_[0], (@_ > 1) ? $_[1] : caller() );
};
*CORE::GLOBAL::bless = sub { goto $bless_override };
}
use strict;
use warnings;
use Test::More;
use lib qw(t/lib);
use DBICTest::RunMode;
use DBICTest::Util::LeakTracer qw(populate_weakregistry assert_empty_weakregistry visit_refs);
use Scalar::Util qw(weaken blessed reftype);
use DBIx::Class::_Util qw(hrefaddr sigwarn_silencer modver_gt_or_eq modver_gt_or_eq_and_lt);
BEGIN {
plan skip_all => "Your perl version $] appears to leak like a sieve - skipping test"
if DBIx::Class::_ENV_::PEEPEENESS;
}
my $TB = Test::More->builder;
if ($ENV{DBICTEST_IN_PERSISTENT_ENV}) {
# without this explicit close TB warns in END after a ->reset
close ($TB->$_) for qw(output failure_output todo_output);
# newer TB does not auto-reopen handles
if ( modver_gt_or_eq( 'Test::More', '1.200' ) ) {
open ($TB->$_, '>&', *STDERR)
for qw( failure_output todo_output );
open ($TB->output, '>&', *STDOUT);
}
# so done_testing can work on every persistent pass
$TB->reset;
}
# this is what holds all weakened refs to be checked for leakage
my $weak_registry = {};
# whether or to invoke IC::DT
my $has_dt;
# Skip the heavy-duty leak tracing when just doing an install
unless (DBICTest::RunMode->is_plain) {
# redefine the bless override so that we can catch each and every object created
no warnings qw/redefine once/;
no strict qw/refs/;
$bless_override = sub {
my $obj = CORE::bless(
$_[0], (@_ > 1) ? $_[1] : do {
my ($class, $fn, $line) = caller();
fail ("bless() of $_[0] into $class without explicit class specification at $fn line $line")
if $class =~ /^ (?: DBIx\:\:Class | DBICTest ) /x;
$class;
}
);
# unicode is tricky, and now we happen to invoke it early via a
# regex in connection()
return $obj if (ref $obj) =~ /^utf8/;
# Test Builder is now making a new object for every pass/fail (que bloat?)
# and as such we can't really store any of its objects (since it will
# re-populate the registry while checking it, ewwww!)
return $obj if (ref $obj) =~ /^TB2::|^Test::Stream/;
# populate immediately to avoid weird side effects
return populate_weakregistry ($weak_registry, $obj );
};
require Try::Tiny;
for my $func (qw/try catch finally/) {
my $orig = \&{"Try::Tiny::$func"};
*{"Try::Tiny::$func"} = sub (&;@) {
populate_weakregistry( $weak_registry, $_[0] );
goto $orig;
}
}
# Some modules are known to install singletons on-load
# Load them and empty the registry
# this loads the DT armada
$has_dt = DBIx::Class::Optional::Dependencies->req_ok_for('test_dt_sqlite');
require Errno;
require DBI;
require DBD::SQLite;
require FileHandle;
require Moo;
%$weak_registry = ();
}
{
use_ok ('DBICTest');
my $schema = DBICTest->init_schema;
my $rs = $schema->resultset ('Artist');
my $storage = $schema->storage;
ok ($storage->connected, 'we are connected');
my $row_obj = $rs->search({}, { rows => 1})->next; # so that commits/rollbacks work
ok ($row_obj, 'row from db');
# txn_do to invoke more codepaths
my ($mc_row_obj, $pager, $pager_explicit_count) = $schema->txn_do (sub {
my $artist = $schema->resultset('Artist')->create ({
name => 'foo artist',
cds => [{
title => 'foo cd',
year => 1984,
tracks => [
{ title => 't1' },
{ title => 't2' },
],
genre => { name => 'mauve' },
}],
});
my $pg = $rs->search({}, { rows => 1})->page(2)->pager;
my $pg_wcount = $rs->page(4)->pager->total_entries (66);
return ($artist, $pg, $pg_wcount);
});
# more codepaths - error handling in txn_do
{
eval { $schema->txn_do ( sub {
$storage->_dbh->begin_work;
fail ('how did we get so far?!');
} ) };
eval { $schema->txn_do ( sub {
$schema->txn_do ( sub {
die "It's called EXCEPTION";
fail ('how did we get so far?!');
} );
fail ('how did we get so far?!');
} ) };
like( $@, qr/It\'s called EXCEPTION/, 'Exception correctly propagated in nested txn_do' );
}
# dbh_do codepath
my ($rs_bind_circref, $cond_rowobj) = $schema->storage->dbh_do ( sub {
my $row = $_[0]->schema->resultset('Artist')->new({});
my $rs = $_[0]->schema->resultset('Artist')->search({
name => $row, # this is deliberately bogus, see FIXME below!
});
return ($rs, $row);
});
is ($pager->next_page, 3, 'There is one more page available');
# based on 66 per 10 pages
is ($pager_explicit_count->last_page, 7, 'Correct last page');
# do some population (invokes some extra codepaths)
# also exercise the guard code and the manual txn control
{
my $guard = $schema->txn_scope_guard;
# populate with bindvars
$rs->populate([{ name => 'James Bound' }]);
$guard->commit;
$schema->txn_begin;
# populate mixed
$rs->populate([{ name => 'James Rebound', rank => \ '11' }]);
$schema->txn_commit;
$schema->txn_begin;
# and without bindvars
$rs->populate([{ name => \ '"James Unbound"' }]);
$schema->txn_rollback;
}
# prefetching
my $cds_rs = $schema->resultset('CD');
my $cds_with_artist = $cds_rs->search({}, { prefetch => 'artist' });
my $cds_with_tracks = $cds_rs->search({}, { prefetch => 'tracks' });
my $cds_with_stuff = $cds_rs->search({}, { prefetch => [ 'genre', { artist => { cds => { tracks => 'cd_single' } } } ] });
# implicit pref
my $cds_with_impl_artist = $cds_rs->search({}, { columns => [qw/me.title artist.name/], join => 'artist' });
# get_column
my $getcol_rs = $cds_rs->get_column('me.cdid');
my $pref_getcol_rs = $cds_with_stuff->get_column('me.cdid');
my $base_collection = {
resultset => $rs,
pref_precursor => $cds_rs,
pref_rs_single => $cds_with_artist,
pref_rs_multi => $cds_with_tracks,
pref_rs_nested => $cds_with_stuff,
pref_rs_implicit => $cds_with_impl_artist,
pref_row_single => $cds_with_artist->next,
pref_row_multi => $cds_with_tracks->next,
pref_row_nested => $cds_with_stuff->next,
# even though this does not leak Storable croaks on it :(((
#pref_row_implicit => $cds_with_impl_artist->next,
get_column_rs_plain => $getcol_rs,
get_column_rs_pref => $pref_getcol_rs,
# twice so that we make sure only one H::M object spawned
chained_resultset => $rs->search_rs ({}, { '+columns' => { foo => 'artistid' } } ),
chained_resultset2 => $rs->search_rs ({}, { '+columns' => { bar => 'artistid' } } ),
row_object => $row_obj,
mc_row_object => $mc_row_obj,
result_source => $rs->result_source,
result_source_handle => $rs->result_source->handle,
pager_explicit_count => $pager_explicit_count,
leaky_resultset => $rs_bind_circref,
leaky_resultset_cond => $cond_rowobj,
};
# fire all resultsets multiple times, once here, more below
# some of these can't find anything (notably leaky_resultset)
my @rsets = grep {
blessed $_
and
(
$_->isa('DBIx::Class::ResultSet')
or
$_->isa('DBIx::Class::ResultSetColumn')
)
} values %$base_collection;
my $fire_resultsets = sub {
local $ENV{DBIC_COLUMNS_INCLUDE_FILTER_RELS} = 1;
local $SIG{__WARN__} = sigwarn_silencer(
qr/Unable to deflate 'filter'-type relationship 'artist'.+related object primary key not retrieved/
);
map
{ $_, (blessed($_) ? { $_->get_columns } : ()) }
map
{ $_->all }
@rsets
;
};
push @{$base_collection->{random_results}}, $fire_resultsets->();
# FIXME - something throws a Storable for a spin if we keep
# the results in-collection. The same problem is seen above,
# swept under the rug back in 0a03206a, damned lazy ribantainer
{
local $base_collection->{random_results};
require Storable;
%$base_collection = (
%$base_collection,
refrozen => Storable::dclone( $base_collection ),
rerefrozen => Storable::dclone( Storable::dclone( $base_collection ) ),
pref_row_implicit => $cds_with_impl_artist->next,
schema => $schema,
storage => $storage,
sql_maker => $storage->sql_maker,
dbh => $storage->_dbh,
fresh_pager => $rs->page(5)->pager,
pager => $pager,
);
}
# FIXME - ideally this kind of collector ought to be global, but attempts
# with an invasive debugger-based tracer did not quite work out... yet
# Manually scan the innards of everything we have in the base collection
# we assembled so far (skip the DT madness below) *recursively*
#
# Only do this when we do have the bits to look inside CVs properly,
# without it we are liable to pick up object defaults that are locked
# in method closures
if (DBICTest::Util::LeakTracer::CV_TRACING) {
visit_refs(
refs => [ $base_collection ],
action => sub {
populate_weakregistry ($weak_registry, $_[0]);
1; # true means "keep descending"
},
);
# do a heavy-duty fire-and-compare loop on all resultsets
# this is expensive - not running on install
my $typecounts = {};
if (
! DBICTest::RunMode->is_plain
and
! $ENV{DBICTEST_IN_PERSISTENT_ENV}
) {
# FIXME - ideally we should be able to just populate an alternative
# registry, subtract everything from the main one, and arrive at
# an "empty" resulting hash
# However due to gross inefficiencies in the ::ResultSet code we
# end up recalculating a new set of aliasmaps which could have very
# well been cached if it wasn't for... anyhow
# What we do here for the time being is similar to the lazy approach
# of Devel::LeakTrace - we just make sure we do not end up with more
# reftypes than when we started. At least we are not blanket-counting
# SVs like D::LT does, but going by reftype... sigh...
for (values %$weak_registry) {
if ( my $r = reftype($_->{weakref}) ) {
$typecounts->{$r}--;
}
}
# For now we can only reuse the same registry, see FIXME above/below
#for my $interim_wr ({}, {}) {
for my $interim_wr ( ($weak_registry) x 4 ) {
visit_refs(
refs => [ $fire_resultsets->(), @rsets ],
action => sub {
populate_weakregistry ($interim_wr, $_[0]);
1; # true means "keep descending"
},
);
# FIXME - this is what *should* be here
#
## anything we have seen so far is cool
#delete @{$interim_wr}{keys %$weak_registry};
#
## moment of truth - the rest ought to be gone
#assert_empty_weakregistry($interim_wr);
}
for (values %$weak_registry) {
if ( my $r = reftype($_->{weakref}) ) {
$typecounts->{$r}++;
}
}
}
for (keys %$typecounts) {
fail ("Amount of $_ refs changed by $typecounts->{$_} during resultset mass-execution")
if ( abs ($typecounts->{$_}) > 1 ); # there is a pad caught somewhere, the +1/-1 can be ignored
}
}
if ($has_dt) {
my $rs = $base_collection->{icdt_rs} = $schema->resultset('Event');
my $now = DateTime->now;
for (1..5) {
$base_collection->{"icdt_row_$_"} = $rs->create({
created_on => DateTime->new(year => 2011, month => 1, day => $_, time_zone => "-0${_}00" ),
starts_at => $now->clone->add(days => $_),
});
}
# re-search
my @dummy = $rs->all;
}
# dbh's are created in XS space, so pull them separately
for ( grep { defined } map { @{$_->{ChildHandles}} } values %{ {DBI->installed_drivers()} } ) {
$base_collection->{"DBI handle $_"} = $_;
}
populate_weakregistry ($weak_registry, $base_collection->{$_}, "basic $_")
for keys %$base_collection;
}
# check that "phantom-chaining" works - we never lose track of the original $schema
# and have access to the entire tree without leaking anything
{
my $phantom;
for (
sub { DBICTest->init_schema( sqlite_use_file => 0 ) },
sub { shift->source('Artist') },
sub { shift->resultset },
sub { shift->result_source },
sub { shift->schema },
sub { shift->resultset('Artist') },
sub { shift->find_or_create({ name => 'detachable' }) },
sub { shift->result_source },
sub { shift->schema },
sub { shift->clone },
sub { shift->resultset('CD') },
sub { shift->next },
sub { shift->artist },
sub { shift->search_related('cds') },
sub { shift->next },
sub { shift->search_related('artist') },
sub { shift->result_source },
sub { shift->resultset },
sub { shift->create({ name => 'detached' }) },
sub { shift->update({ name => 'reattached' }) },
sub { shift->discard_changes },
sub { shift->delete },
sub { shift->insert },
) {
$phantom = populate_weakregistry ( $weak_registry, scalar $_->($phantom) );
}
ok( $phantom->in_storage, 'Properly deleted/reinserted' );
is( $phantom->name, 'reattached', 'Still correct name' );
}
# Naturally we have some exceptions
my $cleared;
for my $addr (keys %$weak_registry) {
my $names = join "\n", keys %{$weak_registry->{$addr}{slot_names}};
if ($names =~ /^Test::Builder/m) {
# T::B 2.0 has result objects and other fancyness
delete $weak_registry->{$addr};
}
elsif ($names =~ /^Hash::Merge/m) {
# only clear one object of a specific behavior - more would indicate trouble
delete $weak_registry->{$addr}
unless $cleared->{hash_merge_singleton}{$weak_registry->{$addr}{weakref}{behavior}}++;
}
elsif ($names =~ /^DateTime::TimeZone::UTC/m) {
# DT is going through a refactor it seems - let it leak zones for now
delete $weak_registry->{$addr};
}
elsif (
# # if we can look at closed over pieces - we will register it as a global
# !DBICTest::Util::LeakTracer::CV_TRACING
# and
$names =~ /^SQL::Translator::Generator::DDL::SQLite/m
) {
# SQLT::Producer::SQLite keeps global generators around for quoted
# and non-quoted DDL, allow one for each quoting style
delete $weak_registry->{$addr}
unless $cleared->{sqlt_ddl_sqlite}->{@{$weak_registry->{$addr}{weakref}->quote_chars}}++;
}
}
# FIXME !!!
# There is an actual strong circular reference taking place here, but because
# half of it is in XS, so it is a bit harder to track down (it stumps D::FR)
# (our tracker does not yet do it, but it'd be nice)
# The problem is:
#
# $cond_object --> result_source --> schema --> storage --> $dbh --> {CachedKids}
# ^ /
# \-------- bound value on prepared/cached STH <-----------/
#
{
my @circreffed;
for my $r (map
{ $_->{weakref} }
grep
{ $_->{slot_names}{'basic leaky_resultset_cond'} }
values %$weak_registry
) {
local $TODO = 'Needs Data::Entangled or somesuch - see RT#82942';
ok(! defined $r, 'Self-referential RS conditions no longer leak!')
or push @circreffed, $r;
}
if (@circreffed) {
is (scalar @circreffed, 1, 'One resultset expected to leak');
# this is useless on its own, it is to showcase the circref-diag
# and eventually test it when it is operational
local $TODO = 'Needs Data::Entangled or somesuch - see RT#82942';
while (@circreffed) {
weaken (my $r = shift @circreffed);
populate_weakregistry( (my $mini_registry = {}), $r );
assert_empty_weakregistry( $mini_registry );
$r->result_source(undef);
}
}
}
assert_empty_weakregistry ($weak_registry);
# we got so far without a failure - this is a good thing
# now let's try to rerun this script under a "persistent" environment
# this is ugly and dirty but we do not yet have a Test::Embedded or
# similar
# set up -I
require Config;
$ENV{PERL5LIB} = join ($Config::Config{path_sep}, @INC);
($ENV{PATH}) = $ENV{PATH} =~ /(.+)/;
my $persistence_tests;
SKIP: {
skip 'Test already in a persistent loop', 1
if $ENV{DBICTEST_IN_PERSISTENT_ENV};
skip 'Main test failed - skipping persistent env tests', 1
unless $TB->is_passing;
skip "Test::Builder\@@{[ Test::Builder->VERSION ]} known to break persistence tests", 1
if modver_gt_or_eq_and_lt( 'Test::More', '1.200', '1.301001_099' );
local $ENV{DBICTEST_IN_PERSISTENT_ENV} = 1;
$persistence_tests = {
PPerl => {
cmd => [qw/pperl --prefork=1/, __FILE__],
},
'CGI::SpeedyCGI' => {
cmd => [qw/speedy -- -t5/, __FILE__],
},
};
# scgi is smart and will auto-reap after -t amount of seconds
# pperl needs an actual killer :(
$persistence_tests->{PPerl}{termcmd} = [
$persistence_tests->{PPerl}{cmd}[0],
'--kill',
@{$persistence_tests->{PPerl}{cmd}}[ 1 .. $#{$persistence_tests->{PPerl}{cmd}} ],
];
require IPC::Open2;
for my $type (keys %$persistence_tests) { SKIP: {
unless (eval "require $type") {
# Don't terminate what we didn't start
delete $persistence_tests->{$type}{termcmd};
skip "$type module not found", 1;
}
my @cmd = @{$persistence_tests->{$type}{cmd}};
# since PPerl is racy and sucks - just prime the "server"
{
local $ENV{DBICTEST_PERSISTENT_ENV_BAIL_EARLY} = 1;
system(@cmd);
sleep 1;
# see if the thing actually runs, if not - might as well bail now
skip "Something is wrong with $type ($!)", 1
if system(@cmd);
}
for (1,2,3) {
note ("Starting run in persistent env ($type pass $_)");
IPC::Open2::open2(my $out, undef, @cmd);
my @out_lines;
while (my $ln = <$out>) {
next if $ln =~ /^\s*$/;
push @out_lines, " $ln";
last if $ln =~ /^\d+\.\.\d+$/; # this is persistence, we need to terminate reading on our end
}
print $_ for @out_lines;
close $out;
wait;
ok (!$?, "Run in persistent env ($type pass $_): exit $?");
ok (scalar @out_lines, "Run in persistent env ($type pass $_): got output");
}
ok (! system (@{$persistence_tests->{$type}{termcmd}}), "killed $type server instance")
if $persistence_tests->{$type}{termcmd};
}}
}
done_testing;
# just an extra precaution in case we blew away from the SKIP - since there are no
# PID files to go by (man does pperl really suck :(
END {
if ($persistence_tests->{PPerl}{termcmd}) {
local $?; # otherwise test will inherit $? of the system()
require IPC::Open3;
open my $null, ">", File::Spec->devnull;
waitpid(
IPC::Open3::open3(undef, $null, $null, @{$persistence_tests->{PPerl}{termcmd}}),
0,
);
}
}