X-Git-Url: http://git.shadowcat.co.uk/gitweb/gitweb.cgi?a=blobdiff_plain;f=t%2F52leaks.t;h=bd159a7ff337346695512e0f18a2e515ef3d281d;hb=d6c13bfdf6656317fedbf7e9deeb450cf42efb5b;hp=7ecec4946b8544940ee4f8b378f8fb019bbbdcd2;hpb=226d1c35c31ebc68b3214fe8b33912d06d2c5b2f;p=dbsrgits%2FDBIx-Class.git diff --git a/t/52leaks.t b/t/52leaks.t index 7ecec49..bd159a7 100644 --- a/t/52leaks.t +++ b/t/52leaks.t @@ -1,58 +1,640 @@ +BEGIN { do "./t/lib/ANFANG.pm" or die ( $@ || $! ) } + +# work around brain damage in PPerl (yes, it has to be a global) +$SIG{__WARN__} = sub { + warn @_ unless $_[0] =~ /\QUse of "goto" to jump into a construct is deprecated/ +} if ($ENV{DBICTEST_IN_PERSISTENT_ENV}); + +# the persistent environments run with this flag first to see if +# we will run at all (e.g. it will fail if $^X doesn't match) +exit 0 if $ENV{DBICTEST_PERSISTENT_ENV_BAIL_EARLY}; + +# Do the override as early as possible so that CORE::bless doesn't get compiled away +# We will replace $bless_override only if we are in author mode +my $bless_override; +BEGIN { + $bless_override = sub { + CORE::bless( $_[0], (@_ > 1) ? $_[1] : caller() ); + }; + *CORE::GLOBAL::bless = sub { goto $bless_override }; +} + use strict; use warnings; - use Test::More; -use lib qw(t/lib); - BEGIN { - require DBIx::Class::Optional::Dependencies; - plan skip_all => 'Test needs: ' . DBIx::Class::Optional::Dependencies->req_missing_for ('test_leaks') - unless ( DBIx::Class::Optional::Dependencies->req_ok_for ('test_leaks') ); + require DBICTest::Util; + plan skip_all => "Your perl version $] appears to leak like a sieve - skipping test" + if DBICTest::Util::PEEPEENESS(); } -use DBICTest; -use DBICTest::Schema; -use Scalar::Util 'weaken'; -use namespace::clean; +use DBICTest::RunMode; +use DBICTest::Util::LeakTracer qw(populate_weakregistry assert_empty_weakregistry visit_refs); +use Scalar::Util qw(weaken blessed reftype); +use DBIx::Class::_Util qw(hrefaddr sigwarn_silencer modver_gt_or_eq modver_gt_or_eq_and_lt); +use DBIx::Class::Optional::Dependencies; + +my $TB = Test::More->builder; +if ($ENV{DBICTEST_IN_PERSISTENT_ENV}) { + # without this explicit close TB warns in END after a ->reset + close ($TB->$_) for qw(output failure_output todo_output); + + # newer TB does not auto-reopen handles + if ( modver_gt_or_eq( 'Test::More', '1.200' ) ) { + open ($TB->$_, '>&', *STDERR) + for qw( failure_output todo_output ); + open ($TB->output, '>&', *STDOUT); + } + + # so done_testing can work on every persistent pass + $TB->reset; +} + +# this is what holds all weakened refs to be checked for leakage +my $weak_registry = {}; + +# whether or to invoke IC::DT +my $has_dt; + +# Skip the heavy-duty leak tracing when just doing an install +# or when having Moose crap all over everything +# FIXME - remove when Replicated gets off Moose +if ( !$ENV{DBICTEST_VIA_REPLICATED} and !DBICTest::RunMode->is_plain ) { + + # redefine the bless override so that we can catch each and every object created + no warnings qw/redefine once/; + no strict qw/refs/; + + $bless_override = sub { + + my $obj = CORE::bless( + $_[0], (@_ > 1) ? $_[1] : do { + my ($class, $fn, $line) = caller(); + fail ("bless() of $_[0] into $class without explicit class specification at $fn line $line") + if $class =~ /^ (?: DBIx\:\:Class | DBICTest ) /x; + $class; + } + ); + + # unicode is tricky, and now we happen to invoke it early via a + # regex in connection() + return $obj if (ref $obj) =~ /^utf8/; + + # Test Builder is now making a new object for every pass/fail (que bloat?) + # and as such we can't really store any of its objects (since it will + # re-populate the registry while checking it, ewwww!) + return $obj if (ref $obj) =~ /^TB2::|^Test::Stream/; + + # populate immediately to avoid weird side effects + return populate_weakregistry ($weak_registry, $obj ); + }; + + require Try::Tiny; + for my $func (qw/try catch finally/) { + my $orig = \&{"Try::Tiny::$func"}; + *{"Try::Tiny::$func"} = sub (&;@) { + populate_weakregistry( $weak_registry, $_[0] ); + goto $orig; + } + } + + # Some modules are known to install singletons on-load + # Load them and empty the registry + + # this loads the DT armada + $has_dt = DBIx::Class::Optional::Dependencies->req_ok_for([qw( test_rdbms_sqlite ic_dt )]); + + require DBI; + require DBD::SQLite; + require Moo; + require Math::BigInt; + + %$weak_registry = (); +} + +{ + use_ok ('DBICTest'); + + my $schema = DBICTest->init_schema; + my $rs = $schema->resultset ('Artist'); + my $storage = $schema->storage; + + my $row_obj = $rs->search({}, { rows => 1})->next; # so that commits/rollbacks work + ok ($row_obj, 'row from db'); + + # txn_do to invoke more codepaths + my ($mc_row_obj, $pager, $pager_explicit_count) = $schema->txn_do (sub { + + my $artist = $schema->resultset('Artist')->create ({ + name => 'foo artist', + cds => [{ + title => 'foo cd', + year => 1984, + tracks => [ + { title => 't1' }, + { title => 't2' }, + ], + genre => { name => 'mauve' }, + }], + }); + + my $pg = $rs->search({}, { rows => 1})->page(2)->pager; + + my $pg_wcount = $rs->page(4)->pager->total_entries (66); + + return ($artist, $pg, $pg_wcount); + }); + + # more codepaths - error handling in txn_do + { + eval { $schema->txn_do ( sub { + $storage->_dbh->begin_work; + fail ('how did we get so far?!'); + } ) }; + + eval { $schema->txn_do ( sub { + $schema->txn_do ( sub { + die "It's called EXCEPTION"; + fail ('how did we get so far?!'); + } ); + fail ('how did we get so far?!'); + } ) }; + like( $@, qr/It\'s called EXCEPTION/, 'Exception correctly propagated in nested txn_do' ); + } + + # dbh_do codepath + my ($rs_bind_circref, $cond_rowobj) = $schema->storage->dbh_do ( sub { + my $row = $_[0]->schema->resultset('Artist')->new({}); + my $rs = $_[0]->schema->resultset('Artist')->search({ + name => $row, # this is deliberately bogus, see FIXME below! + }); + return ($rs, $row); + }); + + is ($pager->next_page, 3, 'There is one more page available'); + + # based on 66 per 10 pages + is ($pager_explicit_count->last_page, 7, 'Correct last page'); + + # do some population (invokes some extra codepaths) + # also exercise the guard code and the manual txn control + { + my $guard = $schema->txn_scope_guard; + # populate with bindvars + $rs->populate([{ name => 'James Bound' }]); + $guard->commit; + + $schema->txn_begin; + # populate mixed + $rs->populate([{ name => 'James Rebound', rank => \ '11' }]); + $schema->txn_commit; + + $schema->txn_begin; + # and without bindvars + $rs->populate([{ name => \ '"James Unbound"' }]); + $schema->txn_rollback; + } + + # prefetching + my $cds_rs = $schema->resultset('CD'); + my $cds_with_artist = $cds_rs->search({}, { prefetch => 'artist' }); + my $cds_with_tracks = $cds_rs->search({}, { prefetch => 'tracks' }); + my $cds_with_stuff = $cds_rs->search({}, { prefetch => [ 'genre', { artist => { cds => { tracks => 'cd_single' } } } ] }); + + # implicit pref + my $cds_with_impl_artist = $cds_rs->search({}, { columns => [qw/me.title artist.name/], join => 'artist' }); + + # get_column + my $getcol_rs = $cds_rs->get_column('me.cdid'); + my $pref_getcol_rs = $cds_with_stuff->get_column('me.cdid'); + + my $base_collection = { + resultset => $rs, + + pref_precursor => $cds_rs, + + pref_rs_single => $cds_with_artist, + pref_rs_multi => $cds_with_tracks, + pref_rs_nested => $cds_with_stuff, + + pref_rs_implicit => $cds_with_impl_artist, + + pref_row_single => $cds_with_artist->next, + pref_row_multi => $cds_with_tracks->next, + pref_row_nested => $cds_with_stuff->next, + + # even though this does not leak Storable croaks on it :((( + #pref_row_implicit => $cds_with_impl_artist->next, + + get_column_rs_plain => $getcol_rs, + get_column_rs_pref => $pref_getcol_rs, + + # twice so that we make sure only one H::M object spawned + chained_resultset => $rs->search_rs ({}, { '+columns' => { foo => 'artistid' } } ), + chained_resultset2 => $rs->search_rs ({}, { '+columns' => { bar => 'artistid' } } ), + + row_object => $row_obj, + + mc_row_object => $mc_row_obj, + + result_source => $rs->result_source, -import Test::Memory::Cycle; + result_source_handle => $rs->result_source->handle, -my $weak; + pager_explicit_count => $pager_explicit_count, + leaky_resultset => $rs_bind_circref, + leaky_resultset_cond => $cond_rowobj, + }; + + # fire all resultsets multiple times, once here, more below + # some of these can't find anything (notably leaky_resultset) + my @rsets = grep { + blessed $_ + and + ( + $_->isa('DBIx::Class::ResultSet') + or + $_->isa('DBIx::Class::ResultSetColumn') + ) + } values %$base_collection; + + + my $fire_resultsets = sub { + local $ENV{DBIC_COLUMNS_INCLUDE_FILTER_RELS} = 1; + local $SIG{__WARN__} = sigwarn_silencer( + qr/Unable to deflate 'filter'-type relationship 'artist'.+related object primary key not retrieved/ + ); + + map + { $_, (blessed($_) ? { $_->get_columns } : ()) } + map + { $_->all } + @rsets + ; + }; + + push @{$base_collection->{random_results}}, $fire_resultsets->(); + + # FIXME - something throws a Storable for a spin if we keep + # the results in-collection. The same problem is seen above, + # swept under the rug back in 0a03206a, damned lazy ribantainer { - my $s = $weak->{schema} = DBICTest->init_schema; - ok ($s->storage->connected, 'we are connected'); - memory_cycle_ok($s, 'No cycles in schema'); + local $base_collection->{random_results}; + + require Storable; + %$base_collection = ( + %$base_collection, + refrozen => Storable::dclone( $base_collection ), + rerefrozen => Storable::dclone( Storable::dclone( $base_collection ) ), + pref_row_implicit => $cds_with_impl_artist->next, + schema => $schema, + storage => $storage, + sql_maker => $storage->sql_maker, + dbh => $storage->_dbh, + fresh_pager => $rs->page(5)->pager, + pager => $pager, + ); +} - my $storage = $weak->{storage} = $s->storage; - memory_cycle_ok($storage, 'No cycles in storage'); + # FIXME - ideally this kind of collector ought to be global, but attempts + # with an invasive debugger-based tracer did not quite work out... yet + # Manually scan the innards of everything we have in the base collection + # we assembled so far (skip the DT madness below) *recursively* + # + # Only do this when we do have the bits to look inside CVs properly, + # without it we are liable to pick up object defaults that are locked + # in method closures + if (DBICTest::Util::LeakTracer::CV_TRACING) { + visit_refs( + refs => [ $base_collection ], + action => sub { + populate_weakregistry ($weak_registry, $_[0]); + 1; # true means "keep descending" + }, + ); - my $rs = $weak->{resultset} = $s->resultset ('Artist'); - memory_cycle_ok($rs, 'No cycles in resultset'); + # do a heavy-duty fire-and-compare loop on all resultsets + # this is expensive - not running on install + my $typecounts = {}; + if ( + ! DBICTest::RunMode->is_plain + and + ! $ENV{DBICTEST_IN_PERSISTENT_ENV} + ) { + + # FIXME - ideally we should be able to just populate an alternative + # registry, subtract everything from the main one, and arrive at + # an "empty" resulting hash + # However due to gross inefficiencies in the ::ResultSet code we + # end up recalculating a new set of aliasmaps which could have very + # well been cached if it wasn't for... anyhow + # What we do here for the time being is similar to the lazy approach + # of Devel::LeakTrace - we just make sure we do not end up with more + # reftypes than when we started. At least we are not blanket-counting + # SVs like D::LT does, but going by reftype... sigh... + + for (values %$weak_registry) { + if ( my $r = reftype($_->{weakref}) ) { + $typecounts->{$r}--; + } + } + + # For now we can only reuse the same registry, see FIXME above/below + #for my $interim_wr ({}, {}) { + for my $interim_wr ( ($weak_registry) x 4 ) { + + visit_refs( + refs => [ $fire_resultsets->(), @rsets ], + action => sub { + populate_weakregistry ($interim_wr, $_[0]); + 1; # true means "keep descending" + }, + ); + + # FIXME - this is what *should* be here + # + ## anything we have seen so far is cool + #delete @{$interim_wr}{keys %$weak_registry}; + # + ## moment of truth - the rest ought to be gone + #assert_empty_weakregistry($interim_wr); + } + + for (values %$weak_registry) { + if ( my $r = reftype($_->{weakref}) ) { + $typecounts->{$r}++; + } + } + } + + for (keys %$typecounts) { + fail ("Amount of $_ refs changed by $typecounts->{$_} during resultset mass-execution") + if ( abs ($typecounts->{$_}) > 1 ); # there is a pad caught somewhere, the +1/-1 can be ignored + } + } - my $rsrc = $weak->{resultsource} = $rs->result_source; - memory_cycle_ok($rsrc, 'No cycles in resultsource'); + if ($has_dt) { + my $rs = $base_collection->{icdt_rs} = $schema->resultset('Event'); - my $row = $weak->{row} = $rs->first; - memory_cycle_ok($row, 'No cycles in row'); + my $now = DateTime->now; + for (1..5) { + $base_collection->{"icdt_row_$_"} = $rs->create({ + created_on => DateTime->new(year => 2011, month => 1, day => $_, time_zone => "-0${_}00" ), + starts_at => $now->clone->add(days => $_), + }); + } - my $sqla = $weak->{sqla} = $s->storage->sql_maker; - memory_cycle_ok($sqla, 'No cycles in SQL maker'); + # re-search + my @dummy = $rs->all; + } + + # dbh's are created in XS space, so pull them separately + for ( grep { defined } map { @{$_->{ChildHandles}} } values %{ {DBI->installed_drivers()} } ) { + $base_collection->{"DBI handle $_"} = $_; + } + + populate_weakregistry ($weak_registry, $base_collection->{$_}, "basic $_") + for keys %$base_collection; +} + +# check that "phantom-chaining" works - we never lose track of the original $schema +# and have access to the entire tree without leaking anything +{ + my $phantom; + for ( + sub { DBICTest->init_schema( sqlite_use_file => 0 ) }, + sub { shift->source('Artist') }, + sub { shift->resultset }, + sub { shift->result_source }, + sub { shift->schema }, + sub { shift->resultset('Artist') }, + sub { shift->find_or_create({ name => 'detachable' }) }, + sub { shift->result_source }, + sub { shift->schema }, + sub { shift->clone }, + sub { shift->resultset('CD') }, + sub { shift->next }, + sub { shift->artist }, + sub { shift->search_related('cds') }, + sub { shift->next }, + sub { shift->search_related('artist') }, + sub { shift->result_source }, + sub { shift->resultset }, + sub { shift->create({ name => 'detached' }) }, + sub { shift->update({ name => 'reattached' }) }, + sub { shift->discard_changes }, + sub { shift->delete }, + sub { shift->insert }, + ) { + $phantom = populate_weakregistry ( $weak_registry, scalar $_->($phantom) ); + } + + ok( $phantom->in_storage, 'Properly deleted/reinserted' ); + is( $phantom->name, 'reattached', 'Still correct name' ); +} + +# Naturally we have some exceptions +my $cleared; +for my $addr (keys %$weak_registry) { + my $names = join "\n", keys %{$weak_registry->{$addr}{slot_names}}; + + if ($names =~ /^Test::Builder/m) { + # T::B 2.0 has result objects and other fancyness + delete $weak_registry->{$addr}; + } + # remove this when IO::Dir is gone from SQLT + elsif ($INC{"IO/Dir.pm"} and $names =~ /^Class::Struct::Tie_ISA/m) { + delete $weak_registry->{$addr}; + } + elsif ($names =~ /^Hash::Merge/m) { + # only clear one object of a specific behavior - more would indicate trouble + delete $weak_registry->{$addr} + unless $cleared->{hash_merge_singleton}{$weak_registry->{$addr}{weakref}{behavior}}++; + } + elsif ($names =~ /^B::Hooks::EndOfScope::PP::_TieHintHashFieldHash/m) { + # there is one tied lexical which stays alive until GC time + # https://metacpan.org/source/ETHER/B-Hooks-EndOfScope-0.15/lib/B/Hooks/EndOfScope/PP/FieldHash.pm#L24 + # simply ignore it here, instead of teaching the leaktracer to examine ties + # the latter is possible yet terrible: https://github.com/dbsrgits/dbix-class/blob/v0.082820/t/lib/DBICTest/Util/LeakTracer.pm#L113-L117 + delete $weak_registry->{$addr} + unless $cleared->{bheos_pptiehinthashfieldhash}++; + } + elsif ( + $names =~ /^Data::Dumper/m + and + $weak_registry->{$addr}{stacktrace} =~ /\bDBIx::Class::SQLMaker::Util::lax_serialize\b/ + ) { + # only clear one object of a specific behavior - more would indicate trouble + delete $weak_registry->{$addr} + unless $cleared->{dd_lax_serializer}++; + } + elsif ($names =~ /^DateTime::TimeZone::UTC/m) { + # DT is going through a refactor it seems - let it leak zones for now + delete $weak_registry->{$addr}; + } + elsif ( +# # if we can look at closed over pieces - we will register it as a global +# !DBICTest::Util::LeakTracer::CV_TRACING +# and + $names =~ /^SQL::Translator::Generator::DDL::SQLite/m + ) { + # SQLT::Producer::SQLite keeps global generators around for quoted + # and non-quoted DDL, allow one for each quoting style + delete $weak_registry->{$addr} + unless $cleared->{sqlt_ddl_sqlite}->{@{$weak_registry->{$addr}{weakref}->quote_chars}}++; + } +} - my $dbh = $weak->{dbh} = $s->storage->_get_dbh; - memory_cycle_ok($dbh, 'No cycles in DBI handle'); +# FIXME !!! +# There is an actual strong circular reference taking place here, but because +# half of it is in XS, so it is a bit harder to track down (it stumps D::FR) +# (our tracker does not yet do it, but it'd be nice) +# The problem is: +# +# $cond_object --> result_source --> schema --> storage --> $dbh --> {CachedKids} +# ^ / +# \-------- bound value on prepared/cached STH <-----------/ +# +{ + my @circreffed; - for (@{$dbh->{ChildHandles}}) { - $weak->{"$_"} = $_ if $_; + for my $r (map + { $_->{weakref} } + grep + { $_->{slot_names}{'basic leaky_resultset_cond'} } + values %$weak_registry + ) { + local $TODO = 'Needs Data::Entangled or somesuch - see RT#82942'; + ok(! defined $r, 'Self-referential RS conditions no longer leak!') + or push @circreffed, $r; } - weaken $_ for values %$weak; - memory_cycle_ok($weak, 'No cycles in weak object collection'); + if (@circreffed) { + is (scalar @circreffed, 1, 'One resultset expected to leak'); + + # this is useless on its own, it is to showcase the circref-diag + # and eventually test it when it is operational + local $TODO = 'Needs Data::Entangled or somesuch - see RT#82942'; + while (@circreffed) { + weaken (my $r = shift @circreffed); + + populate_weakregistry( (my $mini_registry = {}), $r ); + assert_empty_weakregistry( $mini_registry ); + + $r->result_source(undef); + } + } } -for (keys %$weak) { - ok (! $weak->{$_}, "No $_ leaks"); +assert_empty_weakregistry ($weak_registry); + +# we got so far without a failure - this is a good thing +# now let's try to rerun this script under a "persistent" environment +# this is ugly and dirty but we do not yet have a Test::Embedded or +# similar + +my $persistence_tests; +SKIP: { + skip 'Test already in a persistent loop', 1 + if $ENV{DBICTEST_IN_PERSISTENT_ENV}; + + skip 'Main test failed - skipping persistent env tests', 1 + unless $TB->is_passing; + + skip "Test::Builder\@@{[ Test::Builder->VERSION ]} known to break persistence tests", 1 + if modver_gt_or_eq_and_lt( 'Test::More', '1.200', '1.301001_099' ); + + local $ENV{DBICTEST_IN_PERSISTENT_ENV} = 1; + local $ENV{DBICTEST_ANFANG_DEFANG} = 1; + + require File::Spec; + + $persistence_tests = { + PPerl => { + cmd => [qw/pperl --prefork=1/, __FILE__], + }, + 'CGI::SpeedyCGI' => { + cmd => [qw/speedy -- -t5/, __FILE__], + }, + }; + + # scgi is smart and will auto-reap after -t amount of seconds + # pperl needs an actual killer :( + $persistence_tests->{PPerl}{termcmd} = [ + $persistence_tests->{PPerl}{cmd}[0], + '--kill', + @{$persistence_tests->{PPerl}{cmd}}[ 1 .. $#{$persistence_tests->{PPerl}{cmd}} ], + ]; + + # set up -I + require Config; + $ENV{PERL5LIB} = join ($Config::Config{path_sep}, @INC); + + # adjust PATH for -T + if (length $ENV{PATH}) { + ( $ENV{PATH} ) = join ( $Config::Config{path_sep}, + map { length($_) ? File::Spec->rel2abs($_) : () } + split /\Q$Config::Config{path_sep}/, $ENV{PATH} + ) =~ /\A(.+)\z/; + } + + for my $type (keys %$persistence_tests) { SKIP: { + unless (eval "require $type") { + # Don't terminate what we didn't start + delete $persistence_tests->{$type}{termcmd}; + skip "$type module not found", 1; + } + + my @cmd = @{$persistence_tests->{$type}{cmd}}; + + # since PPerl is racy and sucks - just prime the "server" + { + local $ENV{DBICTEST_PERSISTENT_ENV_BAIL_EARLY} = 1; + system(@cmd); + sleep 1; + + # see if the thing actually runs, if not - might as well bail now + skip "Something is wrong with $type ($!)", 1 + if system(@cmd); + } + + require IPC::Open2; + + for (1,2,3) { + note ("Starting run in persistent env ($type pass $_)"); + IPC::Open2::open2(my $out, undef, @cmd); + my @out_lines; + while (my $ln = <$out>) { + next if $ln =~ /^\s*$/; + push @out_lines, " $ln"; + last if $ln =~ /^\d+\.\.\d+$/; # this is persistence, we need to terminate reading on our end + } + print $_ for @out_lines; + close $out; + wait; + ok (!$?, "Run in persistent env ($type pass $_): exit $?"); + ok (scalar @out_lines, "Run in persistent env ($type pass $_): got output"); + } + + ok (! system (@{$persistence_tests->{$type}{termcmd}}), "killed $type server instance") + if $persistence_tests->{$type}{termcmd}; + }} } done_testing; + +# just an extra precaution in case we blew away from the SKIP - since there are no +# PID files to go by (man does pperl really suck :( +END { + if ($persistence_tests->{PPerl}{termcmd}) { + local $?; # otherwise test will inherit $? of the system() + require IPC::Open3; + open my $null, ">", File::Spec->devnull; + waitpid( + IPC::Open3::open3(undef, $null, $null, @{$persistence_tests->{PPerl}{termcmd}}), + 0, + ); + } +}