Test::Run::Core_GplArt - GPL/Artistic-licensed code of Test::Run::Core.
use Test::Run::Obj; my $tester = Test::Run::Obj->new({'test_files' => \@test_files}); $tester->runtests();
STOP! If all you want to do is write a test script, consider using Test::Simple. Test::Run::Core is the module that reads the output from Test::Simple, Test::More and other modules based on Test::Builder. You don't need to know about Test::Run::Core to use those modules.
Test::Run::Core runs tests and expects output from the test in a certain format. That format is called TAP, the Test Anything Protocol. It is defined in Test::Harness::TAP.
$tester-runtests()> runs all the testscripts named as arguments and checks standard output for the expected strings in TAP format.
$tester-
Test::Run::Obj is an applicative derived class of Test::Run::Core that provides a programmer API for running and analyzing the output of TAP files. For calling from the command line, look at Test::Run::CmdLine.
Test::Run will honor the -T or -t in the #! line on your test files. So if you begin a test with:
-T
-t
#!perl -T
the test will be run with taint mode on.
When tests fail, analyze the summary report:
t/base..............ok t/nonumbers.........ok t/ok................ok t/test-harness......ok t/waterloo..........dubious Test returned status 3 (wstat 768, 0x300) DIED. FAILED tests 1, 3, 5, 7, 9, 11, 13, 15, 17, 19 Failed 10/20 tests, 50.00% okay Failed Test Stat Wstat Total Fail Failed List of Failed ----------------------------------------------------------------------- t/waterloo.t 3 768 20 10 50.00% 1 3 5 7 9 11 13 15 17 19 Failed 1/5 test scripts, 80.00% okay. 10/44 subtests failed, 77.27% okay.
Everything passed but t/waterloo.t. It failed 10 of 20 tests and exited with non-zero status indicating something dubious happened.
The columns in the summary report mean:
The test file which failed.
If the test exited with non-zero, this is its exit status.
The wait status of the test.
Total number of tests expected to run.
Number which failed, either from "not ok" or because they never ran.
Percentage of the total tests which failed.
A list of the tests which failed. Successive failures may be abbreviated (ie. 15-20 to indicate that tests 15, 16, 17, 18, 19 and 20 failed).
Test::Run currently only has one interface function, here it is.
my $allok = $self->runtests();
This runs all the given @test_files and divines whether they passed or failed based on their output to STDOUT (details above). It prints out each individual test which failed along with a summary report and a how long it all took.
It returns true if everything was ok. Otherwise it will die() with one of the messages in the DIAGNOSTICS section.
die()
None.
All tests successful.\nFiles=%d, Tests=%d, %s
If all tests are successful some statistics about the performance are printed.
FAILED tests %s\n\tFailed %d/%d tests, %.2f%% okay.
For any single script that has failing subtests statistics like the above are printed.
Test returned status %d (wstat %d)
Scripts that return a non-zero exit status, both $? >> 8 and $? are printed in a message similar to the above.
$? >> 8
$?
Failed 1 test, %.2f%% okay. %s
Failed %d/%d tests, %.2f%% okay. %s
If not all tests were successful, the script dies with one of the above messages.
FAILED--Further testing stopped: %s
If a single subtest decides that further testing will not make sense, the script dies with this message.
Test::Run sets these before executing the individual tests.
HARNESS_ACTIVE
This is set to a true value. It allows the tests to determine if they are being executed through the harness or by any other means.
HARNESS_VERSION
This is the version of Test::Run.
TODO: FIXME
Here's how Test::Run tests itself
$ cd ~/src/devel/Test-Harness $ perl -Mblib -e 'use Test::Run qw(&runtests $verbose); $verbose=0; runtests @ARGV;' t/*.t Using /home/schwern/src/devel/Test-Harness/blib t/base..............ok t/nonumbers.........ok t/ok................ok t/test-harness......ok All tests successful. Files=4, Tests=24, 2 wallclock secs ( 0.61 cusr + 0.41 csys = 1.02 CPU)
The included prove utility for running test scripts from the command line, Test and Test::Simple for writing test scripts, Benchmark for the underlying timing routines, and Devel::Cover for test coverage analysis.
Provide a way of running tests quietly (ie. no printing) for automated validation of tests. This will probably take the form of a version of runtests() which rather than printing its output returns raw data on the state of the tests. (Partially done in Test::Run::Straps)
Document the format.
Fix HARNESS_COMPILE_TEST without breaking its core usage.
Figure a way to report test names in the failure summary.
Rework the test summary so long test names are not truncated as badly. (Partially done with new skip test styles)
Add option for coverage analysis.
Trap STDERR.
Implement Straps total_results()
Remember exit code
Completely redo the print summary code.
Implement Straps callbacks. (experimentally implemented)
Straps->analyze_file() not taint clean, don't know if it can be
Fix that damned VMS nit.
HARNESS_TODOFAIL to display TODO failures
Add a test for verbose.
Change internal list of test results to a hash.
Fix stats display when there's an overrun.
Fix so perls with spaces in the filename work.
Keeping whittling away at _run_all_tests()
Clean up how the summary is printed. Get rid of those damned formats.
HARNESS_COMPILE_TEST currently assumes it's run from the Perl source directory.
Please use the CPAN bug ticketing system at http://rt.cpan.org/. You can also mail bugs, fixes and enhancements to <bug-test-harness at rt.cpan.org>.
<bug-test-harness
rt.cpan.org>
Copyright 2002-2005 by Michael G Schwern <schwern at pobox.com>, Andy Lester <andy at petdance.com>.
<schwern at pobox.com>
<andy at petdance.com>
This program is free software; you can redistribute it and/or modify it under the same terms as Perl itself.
See http://www.perl.com/perl/misc/Artistic.html.
To install Test::Run, copy and paste the appropriate command in to your terminal.
cpanm
cpanm Test::Run
CPAN shell
perl -MCPAN -e shell install Test::Run
For more information on module installation, please visit the detailed CPAN module installation guide.