-
Notifications
You must be signed in to change notification settings - Fork 151
/
52leaks.t
640 lines (533 loc) · 20.2 KB
/
52leaks.t
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
BEGIN { do "./t/lib/ANFANG.pm" or die ( $@ || $! ) }
# work around brain damage in PPerl (yes, it has to be a global)
$SIG{__WARN__} = sub {
warn @_ unless $_[0] =~ /\QUse of "goto" to jump into a construct is deprecated/
} if ($ENV{DBICTEST_IN_PERSISTENT_ENV});
# the persistent environments run with this flag first to see if
# we will run at all (e.g. it will fail if $^X doesn't match)
exit 0 if $ENV{DBICTEST_PERSISTENT_ENV_BAIL_EARLY};
# Do the override as early as possible so that CORE::bless doesn't get compiled away
# We will replace $bless_override only if we are in author mode
my $bless_override;
BEGIN {
$bless_override = sub {
CORE::bless( $_[0], (@_ > 1) ? $_[1] : caller() );
};
*CORE::GLOBAL::bless = sub { goto $bless_override };
}
use strict;
use warnings;
use Test::More;
BEGIN {
require DBICTest::Util;
plan skip_all => "Your perl version $] appears to leak like a sieve - skipping test"
if DBICTest::Util::PEEPEENESS();
}
use DBICTest::RunMode;
use DBICTest::Util::LeakTracer qw(populate_weakregistry assert_empty_weakregistry visit_refs);
use Scalar::Util qw(weaken blessed reftype);
use DBIx::Class::_Util qw(hrefaddr sigwarn_silencer modver_gt_or_eq modver_gt_or_eq_and_lt);
use DBIx::Class::Optional::Dependencies;
my $TB = Test::More->builder;
if ($ENV{DBICTEST_IN_PERSISTENT_ENV}) {
# without this explicit close TB warns in END after a ->reset
close ($TB->$_) for qw(output failure_output todo_output);
# newer TB does not auto-reopen handles
if ( modver_gt_or_eq( 'Test::More', '1.200' ) ) {
open ($TB->$_, '>&', *STDERR)
for qw( failure_output todo_output );
open ($TB->output, '>&', *STDOUT);
}
# so done_testing can work on every persistent pass
$TB->reset;
}
# this is what holds all weakened refs to be checked for leakage
my $weak_registry = {};
# whether or to invoke IC::DT
my $has_dt;
# Skip the heavy-duty leak tracing when just doing an install
# or when having Moose crap all over everything
# FIXME - remove when Replicated gets off Moose
if ( !$ENV{DBICTEST_VIA_REPLICATED} and !DBICTest::RunMode->is_plain ) {
# redefine the bless override so that we can catch each and every object created
no warnings qw/redefine once/;
no strict qw/refs/;
$bless_override = sub {
my $obj = CORE::bless(
$_[0], (@_ > 1) ? $_[1] : do {
my ($class, $fn, $line) = caller();
fail ("bless() of $_[0] into $class without explicit class specification at $fn line $line")
if $class =~ /^ (?: DBIx\:\:Class | DBICTest ) /x;
$class;
}
);
# unicode is tricky, and now we happen to invoke it early via a
# regex in connection()
return $obj if (ref $obj) =~ /^utf8/;
# Test Builder is now making a new object for every pass/fail (que bloat?)
# and as such we can't really store any of its objects (since it will
# re-populate the registry while checking it, ewwww!)
return $obj if (ref $obj) =~ /^TB2::|^Test::Stream/;
# populate immediately to avoid weird side effects
return populate_weakregistry ($weak_registry, $obj );
};
require Try::Tiny;
for my $func (qw/try catch finally/) {
my $orig = \&{"Try::Tiny::$func"};
*{"Try::Tiny::$func"} = sub (&;@) {
populate_weakregistry( $weak_registry, $_[0] );
goto $orig;
}
}
# Some modules are known to install singletons on-load
# Load them and empty the registry
# this loads the DT armada
$has_dt = DBIx::Class::Optional::Dependencies->req_ok_for([qw( test_rdbms_sqlite ic_dt )]);
require DBI;
require DBD::SQLite;
require Moo;
require Math::BigInt;
%$weak_registry = ();
}
{
use_ok ('DBICTest');
my $schema = DBICTest->init_schema;
my $rs = $schema->resultset ('Artist');
my $storage = $schema->storage;
my $row_obj = $rs->search({}, { rows => 1})->next; # so that commits/rollbacks work
ok ($row_obj, 'row from db');
# txn_do to invoke more codepaths
my ($mc_row_obj, $pager, $pager_explicit_count) = $schema->txn_do (sub {
my $artist = $schema->resultset('Artist')->create ({
name => 'foo artist',
cds => [{
title => 'foo cd',
year => 1984,
tracks => [
{ title => 't1' },
{ title => 't2' },
],
genre => { name => 'mauve' },
}],
});
my $pg = $rs->search({}, { rows => 1})->page(2)->pager;
my $pg_wcount = $rs->page(4)->pager->total_entries (66);
return ($artist, $pg, $pg_wcount);
});
# more codepaths - error handling in txn_do
{
eval { $schema->txn_do ( sub {
$storage->_dbh->begin_work;
fail ('how did we get so far?!');
} ) };
eval { $schema->txn_do ( sub {
$schema->txn_do ( sub {
die "It's called EXCEPTION";
fail ('how did we get so far?!');
} );
fail ('how did we get so far?!');
} ) };
like( $@, qr/It\'s called EXCEPTION/, 'Exception correctly propagated in nested txn_do' );
}
# dbh_do codepath
my ($rs_bind_circref, $cond_rowobj) = $schema->storage->dbh_do ( sub {
my $row = $_[0]->schema->resultset('Artist')->new({});
my $rs = $_[0]->schema->resultset('Artist')->search({
name => $row, # this is deliberately bogus, see FIXME below!
});
return ($rs, $row);
});
is ($pager->next_page, 3, 'There is one more page available');
# based on 66 per 10 pages
is ($pager_explicit_count->last_page, 7, 'Correct last page');
# do some population (invokes some extra codepaths)
# also exercise the guard code and the manual txn control
{
my $guard = $schema->txn_scope_guard;
# populate with bindvars
$rs->populate([{ name => 'James Bound' }]);
$guard->commit;
$schema->txn_begin;
# populate mixed
$rs->populate([{ name => 'James Rebound', rank => \ '11' }]);
$schema->txn_commit;
$schema->txn_begin;
# and without bindvars
$rs->populate([{ name => \ '"James Unbound"' }]);
$schema->txn_rollback;
}
# prefetching
my $cds_rs = $schema->resultset('CD');
my $cds_with_artist = $cds_rs->search({}, { prefetch => 'artist' });
my $cds_with_tracks = $cds_rs->search({}, { prefetch => 'tracks' });
my $cds_with_stuff = $cds_rs->search({}, { prefetch => [ 'genre', { artist => { cds => { tracks => 'cd_single' } } } ] });
# implicit pref
my $cds_with_impl_artist = $cds_rs->search({}, { columns => [qw/me.title artist.name/], join => 'artist' });
# get_column
my $getcol_rs = $cds_rs->get_column('me.cdid');
my $pref_getcol_rs = $cds_with_stuff->get_column('me.cdid');
my $base_collection = {
resultset => $rs,
pref_precursor => $cds_rs,
pref_rs_single => $cds_with_artist,
pref_rs_multi => $cds_with_tracks,
pref_rs_nested => $cds_with_stuff,
pref_rs_implicit => $cds_with_impl_artist,
pref_row_single => $cds_with_artist->next,
pref_row_multi => $cds_with_tracks->next,
pref_row_nested => $cds_with_stuff->next,
# even though this does not leak Storable croaks on it :(((
#pref_row_implicit => $cds_with_impl_artist->next,
get_column_rs_plain => $getcol_rs,
get_column_rs_pref => $pref_getcol_rs,
# twice so that we make sure only one H::M object spawned
chained_resultset => $rs->search_rs ({}, { '+columns' => { foo => 'artistid' } } ),
chained_resultset2 => $rs->search_rs ({}, { '+columns' => { bar => 'artistid' } } ),
row_object => $row_obj,
mc_row_object => $mc_row_obj,
result_source => $rs->result_source,
result_source_handle => $rs->result_source->handle,
pager_explicit_count => $pager_explicit_count,
leaky_resultset => $rs_bind_circref,
leaky_resultset_cond => $cond_rowobj,
};
# fire all resultsets multiple times, once here, more below
# some of these can't find anything (notably leaky_resultset)
my @rsets = grep {
blessed $_
and
(
$_->isa('DBIx::Class::ResultSet')
or
$_->isa('DBIx::Class::ResultSetColumn')
)
} values %$base_collection;
my $fire_resultsets = sub {
local $ENV{DBIC_COLUMNS_INCLUDE_FILTER_RELS} = 1;
local $SIG{__WARN__} = sigwarn_silencer(
qr/Unable to deflate 'filter'-type relationship 'artist'.+related object primary key not retrieved/
);
map
{ $_, (blessed($_) ? { $_->get_columns } : ()) }
map
{ $_->all }
@rsets
;
};
push @{$base_collection->{random_results}}, $fire_resultsets->();
# FIXME - something throws a Storable for a spin if we keep
# the results in-collection. The same problem is seen above,
# swept under the rug back in 0a03206a, damned lazy ribantainer
{
local $base_collection->{random_results};
require Storable;
%$base_collection = (
%$base_collection,
refrozen => Storable::dclone( $base_collection ),
rerefrozen => Storable::dclone( Storable::dclone( $base_collection ) ),
pref_row_implicit => $cds_with_impl_artist->next,
schema => $schema,
storage => $storage,
sql_maker => $storage->sql_maker,
dbh => $storage->_dbh,
fresh_pager => $rs->page(5)->pager,
pager => $pager,
);
}
# FIXME - ideally this kind of collector ought to be global, but attempts
# with an invasive debugger-based tracer did not quite work out... yet
# Manually scan the innards of everything we have in the base collection
# we assembled so far (skip the DT madness below) *recursively*
#
# Only do this when we do have the bits to look inside CVs properly,
# without it we are liable to pick up object defaults that are locked
# in method closures
if (DBICTest::Util::LeakTracer::CV_TRACING) {
visit_refs(
refs => [ $base_collection ],
action => sub {
populate_weakregistry ($weak_registry, $_[0]);
1; # true means "keep descending"
},
);
# do a heavy-duty fire-and-compare loop on all resultsets
# this is expensive - not running on install
my $typecounts = {};
if (
! DBICTest::RunMode->is_plain
and
! $ENV{DBICTEST_IN_PERSISTENT_ENV}
) {
# FIXME - ideally we should be able to just populate an alternative
# registry, subtract everything from the main one, and arrive at
# an "empty" resulting hash
# However due to gross inefficiencies in the ::ResultSet code we
# end up recalculating a new set of aliasmaps which could have very
# well been cached if it wasn't for... anyhow
# What we do here for the time being is similar to the lazy approach
# of Devel::LeakTrace - we just make sure we do not end up with more
# reftypes than when we started. At least we are not blanket-counting
# SVs like D::LT does, but going by reftype... sigh...
for (values %$weak_registry) {
if ( my $r = reftype($_->{weakref}) ) {
$typecounts->{$r}--;
}
}
# For now we can only reuse the same registry, see FIXME above/below
#for my $interim_wr ({}, {}) {
for my $interim_wr ( ($weak_registry) x 4 ) {
visit_refs(
refs => [ $fire_resultsets->(), @rsets ],
action => sub {
populate_weakregistry ($interim_wr, $_[0]);
1; # true means "keep descending"
},
);
# FIXME - this is what *should* be here
#
## anything we have seen so far is cool
#delete @{$interim_wr}{keys %$weak_registry};
#
## moment of truth - the rest ought to be gone
#assert_empty_weakregistry($interim_wr);
}
for (values %$weak_registry) {
if ( my $r = reftype($_->{weakref}) ) {
$typecounts->{$r}++;
}
}
}
for (keys %$typecounts) {
fail ("Amount of $_ refs changed by $typecounts->{$_} during resultset mass-execution")
if ( abs ($typecounts->{$_}) > 1 ); # there is a pad caught somewhere, the +1/-1 can be ignored
}
}
if ($has_dt) {
my $rs = $base_collection->{icdt_rs} = $schema->resultset('Event');
my $now = DateTime->now;
for (1..5) {
$base_collection->{"icdt_row_$_"} = $rs->create({
created_on => DateTime->new(year => 2011, month => 1, day => $_, time_zone => "-0${_}00" ),
starts_at => $now->clone->add(days => $_),
});
}
# re-search
my @dummy = $rs->all;
}
# dbh's are created in XS space, so pull them separately
for ( grep { defined } map { @{$_->{ChildHandles}} } values %{ {DBI->installed_drivers()} } ) {
$base_collection->{"DBI handle $_"} = $_;
}
populate_weakregistry ($weak_registry, $base_collection->{$_}, "basic $_")
for keys %$base_collection;
}
# check that "phantom-chaining" works - we never lose track of the original $schema
# and have access to the entire tree without leaking anything
{
my $phantom;
for (
sub { DBICTest->init_schema( sqlite_use_file => 0 ) },
sub { shift->source('Artist') },
sub { shift->resultset },
sub { shift->result_source },
sub { shift->schema },
sub { shift->resultset('Artist') },
sub { shift->find_or_create({ name => 'detachable' }) },
sub { shift->result_source },
sub { shift->schema },
sub { shift->clone },
sub { shift->resultset('CD') },
sub { shift->next },
sub { shift->artist },
sub { shift->search_related('cds') },
sub { shift->next },
sub { shift->search_related('artist') },
sub { shift->result_source },
sub { shift->resultset },
sub { shift->create({ name => 'detached' }) },
sub { shift->update({ name => 'reattached' }) },
sub { shift->discard_changes },
sub { shift->delete },
sub { shift->insert },
) {
$phantom = populate_weakregistry ( $weak_registry, scalar $_->($phantom) );
}
ok( $phantom->in_storage, 'Properly deleted/reinserted' );
is( $phantom->name, 'reattached', 'Still correct name' );
}
# Naturally we have some exceptions
my $cleared;
for my $addr (keys %$weak_registry) {
my $names = join "\n", keys %{$weak_registry->{$addr}{slot_names}};
if ($names =~ /^Test::Builder/m) {
# T::B 2.0 has result objects and other fancyness
delete $weak_registry->{$addr};
}
# remove this when IO::Dir is gone from SQLT
elsif ($INC{"IO/Dir.pm"} and $names =~ /^Class::Struct::Tie_ISA/m) {
delete $weak_registry->{$addr};
}
elsif ($names =~ /^Hash::Merge/m) {
# only clear one object of a specific behavior - more would indicate trouble
delete $weak_registry->{$addr}
unless $cleared->{hash_merge_singleton}{$weak_registry->{$addr}{weakref}{behavior}}++;
}
elsif ($names =~ /^B::Hooks::EndOfScope::PP::_TieHintHashFieldHash/m) {
# there is one tied lexical which stays alive until GC time
# https://metacpan.org/source/ETHER/B-Hooks-EndOfScope-0.15/lib/B/Hooks/EndOfScope/PP/FieldHash.pm#L24
# simply ignore it here, instead of teaching the leaktracer to examine ties
# the latter is possible yet terrible: https://github.com/dbsrgits/dbix-class/blob/v0.082820/t/lib/DBICTest/Util/LeakTracer.pm#L113-L117
delete $weak_registry->{$addr}
unless $cleared->{bheos_pptiehinthashfieldhash}++;
}
elsif (
$names =~ /^Data::Dumper/m
and
$weak_registry->{$addr}{stacktrace} =~ /\bDBIx::Class::SQLMaker::Util::lax_serialize\b/
) {
# only clear one object of a specific behavior - more would indicate trouble
delete $weak_registry->{$addr}
unless $cleared->{dd_lax_serializer}++;
}
elsif ($names =~ /^DateTime::TimeZone::UTC/m) {
# DT is going through a refactor it seems - let it leak zones for now
delete $weak_registry->{$addr};
}
elsif (
# # if we can look at closed over pieces - we will register it as a global
# !DBICTest::Util::LeakTracer::CV_TRACING
# and
$names =~ /^SQL::Translator::Generator::DDL::SQLite/m
) {
# SQLT::Producer::SQLite keeps global generators around for quoted
# and non-quoted DDL, allow one for each quoting style
delete $weak_registry->{$addr}
unless $cleared->{sqlt_ddl_sqlite}->{@{$weak_registry->{$addr}{weakref}->quote_chars}}++;
}
}
# FIXME !!!
# There is an actual strong circular reference taking place here, but because
# half of it is in XS, so it is a bit harder to track down (it stumps D::FR)
# (our tracker does not yet do it, but it'd be nice)
# The problem is:
#
# $cond_object --> result_source --> schema --> storage --> $dbh --> {CachedKids}
# ^ /
# \-------- bound value on prepared/cached STH <-----------/
#
{
my @circreffed;
for my $r (map
{ $_->{weakref} }
grep
{ $_->{slot_names}{'basic leaky_resultset_cond'} }
values %$weak_registry
) {
local $TODO = 'Needs Data::Entangled or somesuch - see RT#82942';
ok(! defined $r, 'Self-referential RS conditions no longer leak!')
or push @circreffed, $r;
}
if (@circreffed) {
is (scalar @circreffed, 1, 'One resultset expected to leak');
# this is useless on its own, it is to showcase the circref-diag
# and eventually test it when it is operational
local $TODO = 'Needs Data::Entangled or somesuch - see RT#82942';
while (@circreffed) {
weaken (my $r = shift @circreffed);
populate_weakregistry( (my $mini_registry = {}), $r );
assert_empty_weakregistry( $mini_registry );
$r->result_source(undef);
}
}
}
assert_empty_weakregistry ($weak_registry);
# we got so far without a failure - this is a good thing
# now let's try to rerun this script under a "persistent" environment
# this is ugly and dirty but we do not yet have a Test::Embedded or
# similar
my $persistence_tests;
SKIP: {
skip 'Test already in a persistent loop', 1
if $ENV{DBICTEST_IN_PERSISTENT_ENV};
skip 'Main test failed - skipping persistent env tests', 1
unless $TB->is_passing;
skip "Test::Builder\@@{[ Test::Builder->VERSION ]} known to break persistence tests", 1
if modver_gt_or_eq_and_lt( 'Test::More', '1.200', '1.301001_099' );
local $ENV{DBICTEST_IN_PERSISTENT_ENV} = 1;
local $ENV{DBICTEST_ANFANG_DEFANG} = 1;
require File::Spec;
$persistence_tests = {
PPerl => {
cmd => [qw/pperl --prefork=1/, __FILE__],
},
'CGI::SpeedyCGI' => {
cmd => [qw/speedy -- -t5/, __FILE__],
},
};
# scgi is smart and will auto-reap after -t amount of seconds
# pperl needs an actual killer :(
$persistence_tests->{PPerl}{termcmd} = [
$persistence_tests->{PPerl}{cmd}[0],
'--kill',
@{$persistence_tests->{PPerl}{cmd}}[ 1 .. $#{$persistence_tests->{PPerl}{cmd}} ],
];
# set up -I
require Config;
$ENV{PERL5LIB} = join ($Config::Config{path_sep}, @INC);
# adjust PATH for -T
if (length $ENV{PATH}) {
( $ENV{PATH} ) = join ( $Config::Config{path_sep},
map { length($_) ? File::Spec->rel2abs($_) : () }
split /\Q$Config::Config{path_sep}/, $ENV{PATH}
) =~ /\A(.+)\z/;
}
for my $type (keys %$persistence_tests) { SKIP: {
unless (eval "require $type") {
# Don't terminate what we didn't start
delete $persistence_tests->{$type}{termcmd};
skip "$type module not found", 1;
}
my @cmd = @{$persistence_tests->{$type}{cmd}};
# since PPerl is racy and sucks - just prime the "server"
{
local $ENV{DBICTEST_PERSISTENT_ENV_BAIL_EARLY} = 1;
system(@cmd);
sleep 1;
# see if the thing actually runs, if not - might as well bail now
skip "Something is wrong with $type ($!)", 1
if system(@cmd);
}
require IPC::Open2;
for (1,2,3) {
note ("Starting run in persistent env ($type pass $_)");
IPC::Open2::open2(my $out, undef, @cmd);
my @out_lines;
while (my $ln = <$out>) {
next if $ln =~ /^\s*$/;
push @out_lines, " $ln";
last if $ln =~ /^\d+\.\.\d+$/; # this is persistence, we need to terminate reading on our end
}
print $_ for @out_lines;
close $out;
wait;
ok (!$?, "Run in persistent env ($type pass $_): exit $?");
ok (scalar @out_lines, "Run in persistent env ($type pass $_): got output");
}
ok (! system (@{$persistence_tests->{$type}{termcmd}}), "killed $type server instance")
if $persistence_tests->{$type}{termcmd};
}}
}
done_testing;
# just an extra precaution in case we blew away from the SKIP - since there are no
# PID files to go by (man does pperl really suck :(
END {
if ($persistence_tests->{PPerl}{termcmd}) {
local $?; # otherwise test will inherit $? of the system()
require IPC::Open3;
open my $null, ">", File::Spec->devnull;
waitpid(
IPC::Open3::open3(undef, $null, $null, @{$persistence_tests->{PPerl}{termcmd}}),
0,
);
}
}