X-Git-Url: http://git.shadowcat.co.uk/gitweb/gitweb.cgi?a=blobdiff_plain;f=lib%2FTest%2FHarness.pm;h=2024d24305271854933112ea973d5b9b8e6089df;hb=cdfe229e642682dd52e04cdd1232a90648b35fe3;hp=788042ada3016684c3431c07c8d081bc35596a30;hpb=356733dafb64c27d060c217f81fd00dc55b1e995;p=p5sagit%2Fp5-mst-13.2.git diff --git a/lib/Test/Harness.pm b/lib/Test/Harness.pm index 788042a..2024d24 100644 --- a/lib/Test/Harness.pm +++ b/lib/Test/Harness.pm @@ -1,9 +1,8 @@ # -*- Mode: cperl; cperl-indent-level: 4 -*- -# $Id: Harness.pm,v 1.14.2.18 2002/04/25 05:04:35 schwern Exp $ package Test::Harness; -require 5.004; +require 5.00405; use Test::Harness::Straps; use Test::Harness::Assert; use Exporter; @@ -11,46 +10,66 @@ use Benchmark; use Config; use strict; -use vars qw($VERSION $Verbose $Switches $Have_Devel_Corestack $Curtest - $Columns $verbose $switches - @ISA @EXPORT @EXPORT_OK - ); -# Backwards compatibility for exportable variable names. -*verbose = \$Verbose; -*switches = \$Switches; +use vars qw( + $VERSION + @ISA @EXPORT @EXPORT_OK + $Verbose $Switches $Debug + $verbose $switches $debug + $Columns + $Timer + $ML $Last_ML_Print + $Strap + $has_time_hires +); + +BEGIN { + eval "use Time::HiRes 'time'"; + $has_time_hires = !$@; +} + +=head1 NAME + +Test::Harness - Run Perl standard test scripts with statistics -$Have_Devel_Corestack = 0; +=head1 VERSION -$VERSION = '2.03'; +Version 2.58 + +=cut + +$VERSION = '2.58_01'; + +# Backwards compatibility for exportable variable names. +*verbose = *Verbose; +*switches = *Switches; +*debug = *Debug; $ENV{HARNESS_ACTIVE} = 1; +$ENV{HARNESS_VERSION} = $VERSION; END { # For VMS. delete $ENV{HARNESS_ACTIVE}; + delete $ENV{HARNESS_VERSION}; } -# Some experimental versions of OS/2 build have broken $? -my $Ignore_Exitcode = $ENV{HARNESS_IGNORE_EXITCODE}; - my $Files_In_Dir = $ENV{HARNESS_FILELEAK_IN_DIR}; -my $Strap = Test::Harness::Straps->new; +$Strap = Test::Harness::Straps->new; + +sub strap { return $Strap }; @ISA = ('Exporter'); @EXPORT = qw(&runtests); -@EXPORT_OK = qw($verbose $switches); +@EXPORT_OK = qw(&execute_tests $verbose $switches); $Verbose = $ENV{HARNESS_VERBOSE} || 0; +$Debug = $ENV{HARNESS_DEBUG} || 0; $Switches = "-w"; $Columns = $ENV{HARNESS_COLUMNS} || $ENV{COLUMNS} || 80; $Columns--; # Some shells have trouble with a full line of text. - - -=head1 NAME - -Test::Harness - run perl standard test scripts with statistics +$Timer = $ENV{HARNESS_TIMER} || 0; =head1 SYNOPSIS @@ -60,175 +79,31 @@ Test::Harness - run perl standard test scripts with statistics =head1 DESCRIPTION -B If all you want to do is write a test script, consider using -Test::Simple. Otherwise, read on. - -(By using the Test module, you can write test scripts without -knowing the exact output this module expects. However, if you need to -know the specifics, read on!) - -Perl test scripts print to standard output C<"ok N"> for each single -test, where C is an increasing sequence of integers. The first line -output by a standard test script is C<"1..M"> with C being the -number of tests that should be run within the test -script. Test::Harness::runtests(@tests) runs all the testscripts -named as arguments and checks standard output for the expected -C<"ok N"> strings. - -After all tests have been performed, runtests() prints some -performance statistics that are computed by the Benchmark module. - -=head2 The test script output - -The following explains how Test::Harness interprets the output of your -test program. - -=over 4 - -=item B<'1..M'> - -This header tells how many tests there will be. For example, C<1..10> -means you plan on running 10 tests. This is a safeguard in case your -test dies quietly in the middle of its run. - -It should be the first non-comment line output by your test program. - -In certain instances, you may not know how many tests you will -ultimately be running. In this case, it is permitted for the 1..M -header to appear as the B line output by your test (again, it -can be followed by further comments). - -Under B circumstances should 1..M appear in the middle of your -output or more than once. - - -=item B<'ok', 'not ok'. Ok?> - -Any output from the testscript to standard error is ignored and -bypassed, thus will be seen by the user. Lines written to standard -output containing C are interpreted as feedback for -runtests(). All other lines are discarded. - -C indicates a failed test. C is a successful test. - - -=item B +B If all you want to do is write a test script, consider +using Test::Simple. Test::Harness is the module that reads the +output from Test::Simple, Test::More and other modules based on +Test::Builder. You don't need to know about Test::Harness to use +those modules. -Perl normally expects the 'ok' or 'not ok' to be followed by a test -number. It is tolerated if the test numbers after 'ok' are -omitted. In this case Test::Harness maintains temporarily its own -counter until the script supplies test numbers again. So the following -test script +Test::Harness runs tests and expects output from the test in a +certain format. That format is called TAP, the Test Anything +Protocol. It is defined in L. - print < - -Anything after the test number but before the # is considered to be -the name of the test. - - ok 42 this is the name of the test - -Currently, Test::Harness does nothing with this information. - -=item B - -If the standard output line contains the substring C< # Skip> (with -variations in spacing and case) after C or C, it is -counted as a skipped test. If the whole testscript succeeds, the -count of skipped tests is included in the generated output. -C reports the text after C< # Skip\S*\s+> as a reason -for skipping. - - ok 23 # skip Insufficient flogiston pressure. - -Similarly, one can include a similar explanation in a C<1..0> line -emitted if the test script is skipped completely: - - 1..0 # Skipped: no leverage found - -=item B - -If the standard output line contains the substring C< # TODO> after -C or C, it is counted as a todo test. The text -afterwards is the thing that has to be done before this test will -succeed. - - not ok 13 # TODO harness the power of the atom - -=begin _deprecated - -Alternatively, you can specify a list of what tests are todo as part -of the test header. - - 1..23 todo 5 12 23 - -This only works if the header appears at the beginning of the test. - -This style is B. - -=end _deprecated - -These tests represent a feature to be implemented or a bug to be fixed -and act as something of an executable "thing to do" list. They are -B expected to succeed. Should a todo test begin succeeding, -Test::Harness will report it as a bonus. This indicates that whatever -you were supposed to do has been done and you should promote this to a -normal test. - -=item B - -As an emergency measure, a test script can decide that further tests -are useless (e.g. missing dependencies) and testing should stop -immediately. In that case the test script prints the magic words - - Bail out! - -to standard output. Any message after these words will be displayed by -C as the reason why testing is stopped. - -=item B - -Additional comments may be put into the testing output on their own -lines. Comment lines should begin with a '#', Test::Harness will -ignore them. - - ok 1 - # Life is good, the sun is shining, RAM is cheap. - not ok 2 - # got 'Bush' expected 'Gore' - -=item B - -Any other output Test::Harness sees it will silently ignore B If you wish to place additional output in your -test script, please use a comment. - -=back +C runs all the testscripts named +as arguments and checks standard output for the expected strings +in TAP format. +The F utility is a thin wrapper around Test::Harness. =head2 Taint mode -Test::Harness will honor the C<-T> in the #! line on your test files. So -if you begin a test with: +Test::Harness will honor the C<-T> or C<-t> in the #! line on your +test files. So if you begin a test with: #!perl -T the test will be run with taint mode on. - =head2 Configuration variables. These variables can be used to configure the behavior of @@ -236,25 +111,30 @@ Test::Harness. They are exported on request. =over 4 -=item B<$Test::Harness::verbose> +=item C<$Test::Harness::Verbose> -The global variable $Test::Harness::verbose is exportable and can be -used to let runtests() display the standard output of the script -without altering the behavior otherwise. +The package variable C<$Test::Harness::Verbose> is exportable and can be +used to let C display the standard output of the script +without altering the behavior otherwise. The F utility's C<-v> +flag will set this. -=item B<$Test::Harness::switches> +=item C<$Test::Harness::switches> -The global variable $Test::Harness::switches is exportable and can be +The package variable C<$Test::Harness::switches> is exportable and can be used to set perl command line options used for running the test -script(s). The default value is C<-w>. +script(s). The default value is C<-w>. It overrides C. + +=item C<$Test::Harness::Timer> + +If set to true, and C is available, print elapsed seconds +after each test file. =back =head2 Failure -It will happen, your tests will fail. After you mop up your ego, you -can begin examining the summary report: +When tests fail, analyze the summary report: t/base..............ok t/nonumbers.........ok @@ -264,12 +144,12 @@ can begin examining the summary report: Test returned status 3 (wstat 768, 0x300) DIED. FAILED tests 1, 3, 5, 7, 9, 11, 13, 15, 17, 19 Failed 10/20 tests, 50.00% okay - Failed Test Stat Wstat Total Fail Failed List of Failed - ----------------------------------------------------------------------- - t/waterloo.t 3 768 20 10 50.00% 1 3 5 7 9 11 13 15 17 19 + Failed Test Stat Wstat Total Fail List of Failed + --------------------------------------------------------------- + t/waterloo.t 3 768 20 10 1 3 5 7 9 11 13 15 17 19 Failed 1/5 test scripts, 80.00% okay. 10/44 subtests failed, 77.27% okay. -Everything passed but t/waterloo.t. It failed 10 of 20 tests and +Everything passed but F. It failed 10 of 20 tests and exited with non-zero status indicating something dubious happened. The columns in the summary report mean: @@ -286,7 +166,7 @@ If the test exited with non-zero, this is its exit status. =item B -The wait status of the test I. +The wait status of the test. =item B @@ -296,10 +176,6 @@ Total number of tests expected to run. Number which failed, either from "not ok" or because they never ran. -=item B - -Percentage of the total tests which failed. - =item B A list of the tests which failed. Successive failures may be @@ -309,25 +185,19 @@ abbreviated (ie. 15-20 to indicate that tests 15, 16, 17, 18, 19 and =back -=head2 Functions +=head1 FUNCTIONS -Test::Harness currently only has one function, here it is. - -=over 4 +The following functions are available. -=item B +=head2 runtests( @test_files ) - my $allok = runtests(@test_files); - -This runs all the given @test_files and divines whether they passed +This runs all the given I<@test_files> and divines whether they passed or failed based on their output to STDOUT (details above). It prints out each individual test which failed along with a summary report and a how long it all took. -It returns true if everything was ok, false otherwise. - -=for _private -This is just _run_all_tests() plus _show_results() +It returns true if everything was ok. Otherwise it will C with +one of the messages in the DIAGNOSTICS section. =cut @@ -336,8 +206,8 @@ sub runtests { local ($\, $,); - my($tot, $failedtests) = _run_all_tests(@tests); - _show_results($tot, $failedtests); + my ($tot, $failedtests,$todo_passed) = execute_tests(tests => \@tests); + print get_results($tot, $failedtests,$todo_passed); my $ok = _all_ok($tot); @@ -347,15 +217,8 @@ sub runtests { return $ok; } -=begin _private - -=item B<_all_ok> - - my $ok = _all_ok(\%tot); - -Tells you if this test run is overall successful or not. - -=cut +# my $ok = _all_ok(\%tot); +# Tells you if this test run is overall successful or not. sub _all_ok { my($tot) = shift; @@ -363,30 +226,30 @@ sub _all_ok { return $tot->{bad} == 0 && ($tot->{max} || $tot->{skipped}) ? 1 : 0; } -=item B<_globdir> - - my @files = _globdir $dir; +# Returns all the files in a directory. This is shorthand for backwards +# compatibility on systems where C doesn't work right. -Returns all the files in a directory. This is shorthand for backwards -compatibility on systems where glob() doesn't work right. +sub _globdir { + local *DIRH; -=cut - -sub _globdir { - opendir DIRH, shift; - my @f = readdir DIRH; - closedir DIRH; + opendir DIRH, shift; + my @f = readdir DIRH; + closedir DIRH; return @f; } -=item B<_run_all_tests> +=head2 execute_tests( tests => \@test_files, out => \*FH ) - my($total, $failed) = _run_all_tests(@test_files); +Runs all the given C<@test_files> (just like C) but +doesn't generate the final report. During testing, progress +information will be written to the currently selected output +filehandle (usually C), or to the filehandle given by the +C parameter. The I is optional. -Runs all the given @test_files (as runtests()) but does it quietly (no -report). $total is a hash ref summary of all the tests run. Its keys -and values are this: +Returns a list of two values, C<$total> and C<$failed>, describing the +results. C<$total> is a hash ref summary of all the tests run. Its +keys and values are this: bonus Number of individual todo tests unexpectedly passed max Number of individual tests ran @@ -400,10 +263,10 @@ and values are this: tests Number of test files originally given skipped Number of test files skipped -If $total->{bad} == 0 and $total->{max} > 0, you've got a successful -test. +If C<< $total->{bad} == 0 >> and C<< $total->{max} > 0 >>, you've +got a successful test. -$failed is a hash ref of all the test scripts which failed. Each key +C<$failed> is a hash ref of all the test scripts that failed. Each key is the name of a test script, each value is another hash representing how that script failed. Its keys are these: @@ -412,19 +275,24 @@ how that script failed. Its keys are these: wstat Script's wait status max Number of individual tests failed Number which failed - percent Percentage of tests which failed canon List of tests which failed (as string). -Needless to say, $failed should be empty if everything passed. - -B Currently this function is still noisy. I'm working on it. +C<$failed> should be empty if everything passed. =cut -sub _run_all_tests { - my(@tests) = @_; - local($|) = 1; - my(%failedtests); +sub execute_tests { + my %args = @_; + my @tests = @{$args{tests}}; + my $out = $args{out} || select(); + + # We allow filehandles that are symbolic refs + no strict 'refs'; + _autoflush($out); + _autoflush(\*STDERR); + + my %failedtests; + my %todo_passed; # Test-wide totals. my(%tot) = ( @@ -441,128 +309,163 @@ sub _run_all_tests { bench => 0, ); - local($ENV{'PERL5LIB'}) = $Strap->_INC2PERL5LIB; - - my @dir_files = _globdir $Files_In_Dir if defined $Files_In_Dir; - my $t_start = new Benchmark; + my @dir_files; + @dir_files = _globdir $Files_In_Dir if defined $Files_In_Dir; + my $run_start_time = new Benchmark; my $width = _leader_width(@tests); foreach my $tfile (@tests) { - $Strap->_reset_file_state; - + $Last_ML_Print = 0; # so each test prints at least once my($leader, $ml) = _mk_leader($tfile, $width); - print $leader; + local $ML = $ml; - my $fh = _open_test($tfile); + print $out $leader; $tot{files}++; + $Strap->{_seen_header} = 0; + if ( $Test::Harness::Debug ) { + print $out "# Running: ", $Strap->_command_line($tfile), "\n"; + } + my $test_start_time = $Timer ? time : 0; + my %results = $Strap->analyze_file($tfile) or + do { warn $Strap->{error}, "\n"; next }; + my $elapsed; + if ( $Timer ) { + $elapsed = time - $test_start_time; + if ( $has_time_hires ) { + $elapsed = sprintf( " %8d ms", $elapsed*1000 ); + } + else { + $elapsed = sprintf( " %8s s", $elapsed ? $elapsed : "<1" ); + } + } + else { + $elapsed = ""; + } + # state of the current test. + my @failed = grep { !$results{details}[$_-1]{ok} } + 1..@{$results{details}}; + my @todo_pass = grep { $results{details}[$_-1]{actual_ok} && + $results{details}[$_-1]{type} eq 'todo' } + 1..@{$results{details}}; + my %test = ( - ok => 0, - 'next' => 0, - max => 0, - failed => [], - todo => {}, - bonus => 0, - skipped => 0, - skip_reason => undef, + ok => $results{ok}, + 'next' => $Strap->{'next'}, + max => $results{max}, + failed => \@failed, + todo_pass => \@todo_pass, + todo => $results{todo}, + bonus => $results{bonus}, + skipped => $results{skip}, + skip_reason => $results{skip_reason}, + skip_all => $Strap->{skip_all}, ml => $ml, ); - my($seen_header, $tests_seen) = (0,0); - while (<$fh>) { - print if $Verbose; + $tot{bonus} += $results{bonus}; + $tot{max} += $results{max}; + $tot{ok} += $results{ok}; + $tot{todo} += $results{todo}; + $tot{sub_skipped} += $results{skip}; - $Strap->{line}++; - if( _parse_header($_, \%test, \%tot) ) { - warn "Test header seen twice!\n" if $seen_header; + my($estatus, $wstatus) = @results{qw(exit wait)}; - $seen_header = 1; - - warn "1..M can only appear at the beginning or end of tests\n" - if $tests_seen && $test{max} < $tests_seen; - } - elsif( _parse_test_line($_, \%test, \%tot) ) { - $tests_seen++; - } - # else, ignore it. - } - - my($estatus, $wstatus) = _close_fh($fh); - - my $allok = $test{ok} == $test{max} && $test{'next'} == $test{max}+1; - - if ($wstatus) { - $failedtests{$tfile} = _dubious_return(\%test, \%tot, - $estatus, $wstatus); - $failedtests{$tfile}{name} = $tfile; - } - elsif ($allok) { + if ($results{passing}) { + # XXX Combine these first two if ($test{max} and $test{skipped} + $test{bonus}) { my @msg; push(@msg, "$test{skipped}/$test{max} skipped: $test{skip_reason}") if $test{skipped}; - push(@msg, "$test{bonus}/$test{max} unexpectedly succeeded") - if $test{bonus}; - print "$test{ml}ok, ".join(', ', @msg)."\n"; - } elsif ($test{max}) { - print "$test{ml}ok\n"; - } elsif (defined $test{skip_reason}) { - print "skipped: $test{skip_reason}\n"; + if ($test{bonus}) { + my ($txt, $canon) = _canondetail($test{todo},0,'TODO passed', + @{$test{todo_pass}}); + $todo_passed{$tfile} = { + canon => $canon, + max => $test{todo}, + failed => $test{bonus}, + name => $tfile, + estat => '', + wstat => '', + }; + + push(@msg, "$test{bonus}/$test{max} unexpectedly succeeded\n$txt"); + } + print $out "$test{ml}ok$elapsed\n ".join(', ', @msg)."\n"; + } + elsif ( $test{max} ) { + print $out "$test{ml}ok$elapsed\n"; + } + elsif ( defined $test{skip_all} and length $test{skip_all} ) { + print $out "skipped\n all skipped: $test{skip_all}\n"; $tot{skipped}++; - } else { - print "skipped test on this platform\n"; + } + else { + print $out "skipped\n all skipped: no reason given\n"; $tot{skipped}++; } $tot{good}++; } else { - if ($test{max}) { - if ($test{'next'} <= $test{max}) { - push @{$test{failed}}, $test{'next'}..$test{max}; + # List unrun tests as failures. + if ($test{'next'} <= $test{max}) { + push @{$test{failed}}, $test{'next'}..$test{max}; + } + # List overruns as failures. + else { + my $details = $results{details}; + foreach my $overrun ($test{max}+1..@$details) { + next unless ref $details->[$overrun-1]; + push @{$test{failed}}, $overrun } - if (@{$test{failed}}) { - my ($txt, $canon) = canonfailed($test{max},$test{skipped}, + } + + if ($wstatus) { + $failedtests{$tfile} = _dubious_return(\%test, \%tot, + $estatus, $wstatus); + $failedtests{$tfile}{name} = $tfile; + } + elsif($results{seen}) { + if (@{$test{failed}} and $test{max}) { + my ($txt, $canon) = _canondetail($test{max},$test{skipped},'Failed', @{$test{failed}}); - print "$test{ml}$txt"; + print $out "$test{ml}$txt"; $failedtests{$tfile} = { canon => $canon, max => $test{max}, failed => scalar @{$test{failed}}, name => $tfile, - percent => 100*(scalar @{$test{failed}})/$test{max}, estat => '', wstat => '', }; - } else { - print "Don't know which tests failed: got $test{ok} ok, ". + } + else { + print $out "Don't know which tests failed: got $test{ok} ok, ". "expected $test{max}\n"; $failedtests{$tfile} = { canon => '??', max => $test{max}, failed => '??', name => $tfile, - percent => undef, estat => '', wstat => '', }; } $tot{bad}++; - } elsif ($test{'next'} == 0) { - print "FAILED before any test output arrived\n"; + } + else { + print $out "FAILED before any test output arrived\n"; $tot{bad}++; $failedtests{$tfile} = { canon => '??', max => '??', failed => '??', name => $tfile, - percent => undef, estat => '', wstat => '', }; } } - $tot{sub_skipped} += $test{skipped}; - if (defined $Files_In_Dir) { my @new_dir_files = _globdir $Files_In_Dir; if (@new_dir_files != @dir_files) { @@ -570,30 +473,36 @@ sub _run_all_tests { @f{@new_dir_files} = (1) x @new_dir_files; delete @f{@dir_files}; my @f = sort keys %f; - print "LEAKED FILES: @f\n"; + print $out "LEAKED FILES: @f\n"; @dir_files = @new_dir_files; } } - - close $fh; - } - $tot{bench} = timediff(new Benchmark, $t_start); + } # foreach test + $tot{bench} = timediff(new Benchmark, $run_start_time); $Strap->_restore_PERL5LIB; - return(\%tot, \%failedtests); + return(\%tot, \%failedtests, \%todo_passed); +} + +# Turns on autoflush for the handle passed +sub _autoflush { + my $flushy_fh = shift; + my $old_fh = select $flushy_fh; + $| = 1; + select $old_fh; } -=item B<_mk_leader> +=for private _mk_leader - my($leader, $ml) = _mk_leader($test_file, $width); + my($leader, $ml) = _mk_leader($test_file, $width); -Generates the 't/foo........' $leader for the given $test_file as well +Generates the 't/foo........' leader for the given C<$test_file> as well as a similar version which will overwrite the current line (by use of -\r and such). $ml may be empty if Test::Harness doesn't think you're +\r and such). C<$ml> may be empty if Test::Harness doesn't think you're on TTY. -The $width is the width of the "yada/blah.." string. +The C<$width> is the width of the "yada/blah.." string. =cut @@ -602,18 +511,20 @@ sub _mk_leader { chomp($te); $te =~ s/\.\w+$/./; - if ($^O eq 'VMS') { $te =~ s/^.*\.t\./\[.t./s; } - my $blank = (' ' x 77); + if ($^O eq 'VMS') { + $te =~ s/^.*\.t\./\[.t./s; + } my $leader = "$te" . '.' x ($width - length($te)); my $ml = ""; - $ml = "\r$blank\r$leader" - if -t STDOUT and not $ENV{HARNESS_NOTTY} and not $Verbose; + if ( -t STDOUT and not $ENV{HARNESS_NOTTY} and not $Verbose ) { + $ml = "\r" . (' ' x 77) . "\r$leader" + } return($leader, $ml); } -=item B<_leader_width> +=for private _leader_width my($width) = _leader_width(@test_files); @@ -636,176 +547,152 @@ sub _leader_width { return $maxlen + 3 - $maxsuflen; } +sub get_results { + my $tot = shift; + my $failedtests = shift; + my $todo_passed = shift; -sub _show_results { - my($tot, $failedtests) = @_; + my $out = ''; - my $pct; my $bonusmsg = _bonusmsg($tot); if (_all_ok($tot)) { - print "All tests successful$bonusmsg.\n"; - } elsif (!$tot->{tests}){ + $out .= "All tests successful$bonusmsg.\n"; + if ($tot->{bonus}) { + my($fmt_top, $fmt) = _create_fmts("Passed TODO",$todo_passed); + # Now write to formats + $out .= swrite( $fmt_top ); + for my $script (sort keys %{$todo_passed||{}}) { + my $Curtest = $todo_passed->{$script}; + $out .= swrite( $fmt, @{ $Curtest }{qw(name estat wstat max failed canon)} ); + } + } + } + elsif (!$tot->{tests}){ die "FAILED--no tests were run for some reason.\n"; - } elsif (!$tot->{max}) { + } + elsif (!$tot->{max}) { my $blurb = $tot->{tests}==1 ? "script" : "scripts"; die "FAILED--$tot->{tests} test $blurb could be run, ". "alas--no output ever seen\n"; - } else { - $pct = sprintf("%.2f", $tot->{good} / $tot->{tests} * 100); - my $percent_ok = 100*$tot->{ok}/$tot->{max}; - my $subpct = sprintf " %d/%d subtests failed, %.2f%% okay.", - $tot->{max} - $tot->{ok}, $tot->{max}, - $percent_ok; + } + else { + my $subresults = sprintf( " %d/%d subtests failed.", + $tot->{max} - $tot->{ok}, $tot->{max} ); - my($fmt_top, $fmt) = _create_fmts($failedtests); + my($fmt_top, $fmt1, $fmt2) = _create_fmts("Failed Test",$failedtests); # Now write to formats + $out .= swrite( $fmt_top ); for my $script (sort keys %$failedtests) { - $Curtest = $failedtests->{$script}; - write; + my $Curtest = $failedtests->{$script}; + $out .= swrite( $fmt1, @{ $Curtest }{qw(name estat wstat max failed canon)} ); + $out .= swrite( $fmt2, $Curtest->{canon} ); } if ($tot->{bad}) { $bonusmsg =~ s/^,\s*//; - print "$bonusmsg.\n" if $bonusmsg; - die "Failed $tot->{bad}/$tot->{tests} test scripts, $pct% okay.". - "$subpct\n"; + $out .= "$bonusmsg.\n" if $bonusmsg; + $out .= "Failed $tot->{bad}/$tot->{tests} test scripts.$subresults\n"; } } - printf("Files=%d, Tests=%d, %s\n", + $out .= sprintf("Files=%d, Tests=%d, %s\n", $tot->{files}, $tot->{max}, timestr($tot->{bench}, 'nop')); + return $out; } +sub swrite { + my $format = shift; + $^A = ''; + formline($format,@_); + my $out = $^A; + $^A = ''; + return $out; +} -sub _parse_header { - my($line, $test, $tot) = @_; - my $is_header = 0; +my %Handlers = ( + header => \&header_handler, + test => \&test_handler, + bailout => \&bailout_handler, +); - if( $Strap->_is_header($line) ) { - $is_header = 1; +$Strap->{callback} = \&strap_callback; +sub strap_callback { + my($self, $line, $type, $totals) = @_; + print $line if $Verbose; - $test->{max} = $Strap->{max}; - for ( keys %{$Strap->{todo}} ) { $test->{todo}{$_} = 1; } + my $meth = $Handlers{$type}; + $meth->($self, $line, $type, $totals) if $meth; +}; - $test->{skip_reason} = $Strap->{skip_all} - if not $test->{max} and defined $Strap->{skip_all}; - $test->{'next'} = 1 unless $test->{'next'}; +sub header_handler { + my($self, $line, $type, $totals) = @_; + warn "Test header seen more than once!\n" if $self->{_seen_header}; - $tot->{max} += $test->{max}; - } - else { - $is_header = 0; - } - - return $is_header; -} + $self->{_seen_header}++; + warn "1..M can only appear at the beginning or end of tests\n" + if $totals->{seen} && + $totals->{max} < $totals->{seen}; +}; -sub _open_test { - my($test) = shift; +sub test_handler { + my($self, $line, $type, $totals) = @_; - my $s = _set_switches($test); + my $curr = $totals->{seen}; + my $next = $self->{'next'}; + my $max = $totals->{max}; + my $detail = $totals->{details}[-1]; - my $perl = -x $^X ? $^X : $Config{perlpath}; + if( $detail->{ok} ) { + _print_ml_less("ok $curr/$max"); - # XXX This is WAY too core specific! - my $cmd = ($ENV{'HARNESS_COMPILE_TEST'}) - ? "./perl -I../lib ../utils/perlcc $test " - . "-r 2>> ./compilelog |" - : "$perl $s $test|"; - $cmd = "MCR $cmd" if $^O eq 'VMS'; - - if( open(PERL, $cmd) ) { - return \*PERL; + if( $detail->{type} eq 'skip' ) { + $totals->{skip_reason} = $detail->{reason} + unless defined $totals->{skip_reason}; + $totals->{skip_reason} = 'various reasons' + if $totals->{skip_reason} ne $detail->{reason}; + } } else { - print "can't run $test. $!\n"; - return; + _print_ml("NOK $curr"); } -} - - -sub _parse_test_line { - my($line, $test, $tot) = @_; - - my %result; - if ( $Strap->_is_test($line, \%result) ) { - $test->{'next'} ||= 1; - my $this = $test->{'next'}; - my($not, $tnum) = (!$result{ok}, $result{number}); - - $this = $tnum if $tnum; + if( $curr > $next ) { + print "Test output counter mismatch [test $curr]\n"; + } + elsif( $curr < $next ) { + print "Confused test output: test $curr answered after ". + "test ", $next - 1, "\n"; + } - my($type, $reason) = ($result{type}, $result{reason}); +}; - my($istodo, $isskip); - if( defined $type ) { - $istodo = 1 if $type eq 'todo'; - $isskip = 1 if $type eq 'skip'; - } +sub bailout_handler { + my($self, $line, $type, $totals) = @_; - $test->{todo}{$this} = 1 if $istodo; - if( $test->{todo}{$this} ) { - $tot->{todo}++; - $test->{bonus}++, $tot->{bonus}++ unless $not; - } + die "FAILED--Further testing stopped" . + ($self->{bailout_reason} ? ": $self->{bailout_reason}\n" : ".\n"); +}; - if( $not && !$test->{todo}{$this} ) { - print "$test->{ml}NOK $this" if $test->{ml}; - push @{$test->{failed}}, $this; - } - else { - print "$test->{ml}ok $this/$test->{max}" if $test->{ml}; - $test->{ok}++; - $tot->{ok}++; - $test->{skipped}++ if $isskip; - - $reason = '[no reason given]' - if $isskip and not defined $reason; - if (defined $reason and defined $test->{skip_reason}) { - # print "was: '$skip_reason' new '$reason'\n"; - $test->{skip_reason} = 'various reasons' - if $test->{skip_reason} ne $reason; - } elsif (defined $reason) { - $test->{skip_reason} = $reason; - } - } - if ($this > $test->{'next'}) { - print "Test output counter mismatch [test $this]\n"; +sub _print_ml { + print join '', $ML, @_ if $ML; +} - # Guard against resource starvation. - if( $this > 100000 ) { - print "Enourmous test number seen [test $this]\n"; - } - else { - push @{$test->{failed}}, $test->{'next'}..$this-1; - } - } - elsif ($this < $test->{'next'}) { - #we have seen more "ok" lines than the number suggests - print "Confused test output: test $this answered after ". - "test ", $test->{'next'}-1, "\n"; - $test->{'next'} = $this; - } - $test->{'next'} = $this + 1; - } - else { - my $bail_reason; - if( $Strap->_is_bail_out($line, \$bail_reason) ) { # bail out! - die "FAILED--Further testing stopped" . - ($bail_reason ? ": $bail_reason\n" : ".\n"); - } +# Print updates only once per second. +sub _print_ml_less { + my $now = CORE::time; + if ( $Last_ML_Print != $now ) { + _print_ml(@_); + $Last_ML_Print = $now; } } - sub _bonusmsg { my($tot) = @_; @@ -828,92 +715,52 @@ sub _bonusmsg { . ($tot->{sub_skipped} != 1 ? 's' : '') . " skipped"; } - return $bonusmsg; } -# VMS has some subtle nastiness with closing the test files. -sub _close_fh { - my($fh) = shift; - - close($fh); # must close to reap child resource values - - my $wstatus = $Ignore_Exitcode ? 0 : $?; # Can trust $? ? - my $estatus; - $estatus = ($^O eq 'VMS' - ? eval 'use vmsish "status"; $estatus = $?' - : $wstatus >> 8); - - return($estatus, $wstatus); -} - - -# Set up the command-line switches to run perl as. -sub _set_switches { - my($test) = shift; - - my $s = $Switches; - $s .= $Strap->_switches($test); - - return $s; -} - - # Test program go boom. sub _dubious_return { my($test, $tot, $estatus, $wstatus) = @_; - my ($failed, $canon, $percent) = ('??', '??'); + + my $failed = '??'; + my $canon = '??'; printf "$test->{ml}dubious\n\tTest returned status $estatus ". "(wstat %d, 0x%x)\n", $wstatus,$wstatus; print "\t\t(VMS status is $estatus)\n" if $^O eq 'VMS'; - if (corestatus($wstatus)) { # until we have a wait module - if ($Have_Devel_Corestack) { - Devel::CoreStack::stack($^X); - } else { - print "\ttest program seems to have generated a core\n"; - } - } - $tot->{bad}++; if ($test->{max}) { if ($test->{'next'} == $test->{max} + 1 and not @{$test->{failed}}) { print "\tafter all the subtests completed successfully\n"; - $percent = 0; $failed = 0; # But we do not set $canon! } else { push @{$test->{failed}}, $test->{'next'}..$test->{max}; $failed = @{$test->{failed}}; - (my $txt, $canon) = canonfailed($test->{max},$test->{skipped},@{$test->{failed}}); - $percent = 100*(scalar @{$test->{failed}})/$test->{max}; + (my $txt, $canon) = _canondetail($test->{max},$test->{skipped},'Failed',@{$test->{failed}}); print "DIED. ",$txt; } } return { canon => $canon, max => $test->{max} || '??', failed => $failed, - percent => $percent, estat => $estatus, wstat => $wstatus, }; } -sub _garbled_output { - my($gibberish) = shift; - warn "Confusing test output: '$gibberish'\n"; -} - - sub _create_fmts { - my($failedtests) = @_; + my $failed_str = shift; + my $failedtests = shift; - my $failed_str = "Failed Test"; - my $middle_str = " Stat Wstat Total Fail Failed "; - my $list_str = "List of Failed"; + my ($type) = split /\s/,$failed_str; + my $short = substr($type,0,4); + my $total = $short eq 'Pass' ? 'TODOs' : 'Total'; + my $middle_str = " Stat Wstat $total $short "; + my $list_str = "List of $type"; # Figure out our longest name string for formatting purposes. my $max_namelen = length($failed_str); @@ -932,115 +779,89 @@ sub _create_fmts { } } - my $fmt_top = "format STDOUT_TOP =\n" - . sprintf("%-${max_namelen}s", $failed_str) + my $fmt_top = sprintf("%-${max_namelen}s", $failed_str) . $middle_str . $list_str . "\n" . "-" x $Columns - . "\n.\n"; - - my $fmt = "format STDOUT =\n" - . "@" . "<" x ($max_namelen - 1) - . " @>> @>>>> @>>>> @>>> ^##.##% " - . "^" . "<" x ($list_len - 1) . "\n" - . '{ $Curtest->{name}, $Curtest->{estat},' - . ' $Curtest->{wstat}, $Curtest->{max},' - . ' $Curtest->{failed}, $Curtest->{percent},' - . ' $Curtest->{canon}' - . "\n}\n" - . "~~" . " " x ($Columns - $list_len - 2) . "^" - . "<" x ($list_len - 1) . "\n" - . '$Curtest->{canon}' - . "\n.\n"; - - eval $fmt_top; - die $@ if $@; - eval $fmt; - die $@ if $@; - - return($fmt_top, $fmt); -} - -{ - my $tried_devel_corestack; - - sub corestatus { - my($st) = @_; - - eval { - local $^W = 0; # *.ph files are often *very* noisy - require 'wait.ph' - }; - return if $@; - my $did_core = defined &WCOREDUMP ? WCOREDUMP($st) : $st & 0200; + . "\n"; - eval { require Devel::CoreStack; $Have_Devel_Corestack++ } - unless $tried_devel_corestack++; + my $fmt1 = "@" . "<" x ($max_namelen - 1) + . " @>> @>>>> @>>>> @>>> " + . "^" . "<" x ($list_len - 1) . "\n"; + my $fmt2 = "~~" . " " x ($Columns - $list_len - 2) . "^" + . "<" x ($list_len - 1) . "\n"; - return $did_core; - } + return($fmt_top, $fmt1, $fmt2); } -sub canonfailed ($@) { - my($max,$skipped,@failed) = @_; +sub _canondetail { + my $max = shift; + my $skipped = shift; + my $type = shift; + my @detail = @_; my %seen; - @failed = sort {$a <=> $b} grep !$seen{$_}++, @failed; - my $failed = @failed; + @detail = sort {$a <=> $b} grep !$seen{$_}++, @detail; + my $detail = @detail; my @result = (); my @canon = (); my $min; - my $last = $min = shift @failed; + my $last = $min = shift @detail; my $canon; - if (@failed) { - for (@failed, $failed[-1]) { # don't forget the last one + my $uc_type = uc($type); + if (@detail) { + for (@detail, $detail[-1]) { # don't forget the last one if ($_ > $last+1 || $_ == $last) { - if ($min == $last) { - push @canon, $last; - } else { - push @canon, "$min-$last"; - } + push @canon, ($min == $last) ? $last : "$min-$last"; $min = $_; } $last = $_; } local $" = ", "; - push @result, "FAILED tests @canon\n"; + push @result, "$uc_type tests @canon\n"; $canon = join ' ', @canon; - } else { - push @result, "FAILED test $last\n"; + } + else { + push @result, "$uc_type test $last\n"; $canon = $last; } - push @result, "\tFailed $failed/$max tests, "; - push @result, sprintf("%.2f",100*(1-$failed/$max)), "% okay"; + return (join("", @result), $canon) + if $type=~/todo/i; + push @result, "\t$type $detail/$max tests, "; + if ($max) { + push @result, sprintf("%.2f",100*(1-$detail/$max)), "% okay"; + } + else { + push @result, "?% okay"; + } my $ender = 's' x ($skipped > 1); - my $good = $max - $failed - $skipped; - my $goodper = sprintf("%.2f",100*($good/$max)); - push @result, " (-$skipped skipped test$ender: $good okay, ". - "$goodper%)" - if $skipped; + if ($skipped) { + my $good = $max - $detail - $skipped; + my $skipmsg = " (less $skipped skipped test$ender: $good okay, "; + if ($max) { + my $goodper = sprintf("%.2f",100*($good/$max)); + $skipmsg .= "$goodper%)"; + } + else { + $skipmsg .= "?%)"; + } + push @result, $skipmsg; + } push @result, "\n"; my $txt = join "", @result; - ($txt, $canon); + return ($txt, $canon); } -=end _private - -=back - -=cut - - 1; __END__ =head1 EXPORT -C<&runtests> is exported by Test::Harness per default. - -C<$verbose> and C<$switches> are exported upon request. +C<&runtests> is exported by Test::Harness by default. +C<&execute_tests>, C<$verbose>, C<$switches> and C<$debug> are +exported upon request. =head1 DIAGNOSTICS @@ -1068,22 +889,33 @@ and C<$?> are printed in a message similar to the above. If not all tests were successful, the script dies with one of the above messages. -=item C +=item C If a single subtest decides that further testing will not make sense, the script dies with this message. =back -=head1 ENVIRONMENT +=head1 ENVIRONMENT VARIABLES THAT TEST::HARNESS SETS + +Test::Harness sets these before executing the individual tests. =over 4 =item C -Harness sets this before executing the individual tests. This allows -the tests to determine if they are being executed through the harness -or by any other means. +This is set to a true value. It allows the tests to determine if they +are being executed through the harness or by any other means. + +=item C + +This is the version of Test::Harness. + +=back + +=head1 ENVIRONMENT VARIABLES THAT AFFECT TEST::HARNESS + +=over 4 =item C @@ -1100,6 +932,13 @@ C before running it. B This currently only works when sitting in the perl source directory! +=item C + +If true, Test::Harness will print debugging information about itself as +it runs the tests. This is different from C, which prints +the output from the test being run. Setting C<$Test::Harness::Debug> will +override this, or you can use the C<-d> switch in the F utility. + =item C When set to the name of a directory, harness will check after each @@ -1111,10 +950,6 @@ If relative, directory name is with respect to the current directory at the moment runtests() was called. Putting absolute path into C may give more predictable results. -=item C - -Makes harness ignore the exit status of child processes when defined. - =item C When set to a true value, forces it to behave as though STDOUT were @@ -1123,16 +958,31 @@ output more frequent progress messages using carriage returns. Some consoles may not handle carriage returns properly (which results in a somewhat messy output). +=item C + +Usually your tests will be run by C<$^X>, the currently-executing Perl. +However, you may want to have it run by a different executable, such as +a threading perl, or a different version. + +If you're using the F utility, you can use the C<--perl> switch. + =item C Its value will be prepended to the switches used to invoke perl on each test. For example, setting C to C<-W> will run all tests with all warnings enabled. +=item C + +Setting this to true will make the harness display the number of +milliseconds each test took. You can also use F's C<--timer> +switch. + =item C If true, Test::Harness will output the verbose results of running -its tests. Setting $Test::Harness::verbose will override this. +its tests. Setting C<$Test::Harness::verbose> will override this, +or you can use the C<-v> switch in the F utility. =back @@ -1153,20 +1003,11 @@ Here's how Test::Harness tests itself =head1 SEE ALSO +The included F utility for running test scripts from the command line, L and L for writing test scripts, L for -the underlying timing routines, L to generate core -dumps from failed tests and L for test coverage +the underlying timing routines, and L for test coverage analysis. -=head1 AUTHORS - -Either Tim Bunce or Andreas Koenig, we don't know. What we know for -sure is, that it was inspired by Larry Wall's TEST script that came -with perl distributions for ages. Numerous anonymous contributors -exist. Andreas Koenig held the torch for many years. - -Current maintainer is Michael G Schwern Eschwern@pobox.comE - =head1 TODO Provide a way of running tests quietly (ie. no printing) for automated @@ -1174,25 +1015,107 @@ validation of tests. This will probably take the form of a version of runtests() which rather than printing its output returns raw data on the state of the tests. (Partially done in Test::Harness::Straps) +Document the format. + Fix HARNESS_COMPILE_TEST without breaking its core usage. Figure a way to report test names in the failure summary. Rework the test summary so long test names are not truncated as badly. - -Deal with VMS's "not \nok 4\n" mistake. +(Partially done with new skip test styles) Add option for coverage analysis. -=for _private +Trap STDERR. + +Implement Straps total_results() + +Remember exit code + +Completely redo the print summary code. + +Implement Straps callbacks. (experimentally implemented) + +Straps->analyze_file() not taint clean, don't know if it can be + +Fix that damned VMS nit. + +Add a test for verbose. + +Change internal list of test results to a hash. + +Fix stats display when there's an overrun. + +Fix so perls with spaces in the filename work. + Keeping whittling away at _run_all_tests() -=for _private Clean up how the summary is printed. Get rid of those damned formats. =head1 BUGS -HARNESS_COMPILE_TEST currently assumes it's run from the Perl source -directory. +Please report any bugs or feature requests to +C, or through the web interface at +L. +I will be notified, and then you'll automatically be notified of progress on +your bug as I make changes. + +=head1 SUPPORT + +You can find documentation for this module with the F command. + + perldoc Test::Harness + +You can get docs for F with + + prove --man + +You can also look for information at: + +=over 4 + +=item * AnnoCPAN: Annotated CPAN documentation + +L + +=item * CPAN Ratings + +L + +=item * RT: CPAN's request tracker + +L + +=item * Search CPAN + +L + +=back + +=head1 SOURCE CODE + +The source code repository for Test::Harness is at +L. + +=head1 AUTHORS + +Either Tim Bunce or Andreas Koenig, we don't know. What we know for +sure is, that it was inspired by Larry Wall's F script that came +with perl distributions for ages. Numerous anonymous contributors +exist. Andreas Koenig held the torch for many years, and then +Michael G Schwern. + +Current maintainer is Andy Lester C<< >>. + +=head1 COPYRIGHT + +Copyright 2002-2006 +by Michael G Schwern C<< >>, +Andy Lester C<< >>. + +This program is free software; you can redistribute it and/or +modify it under the same terms as Perl itself. + +See L. =cut