Difference between revisions of "Eiffeltest"

From Liberty Eiffel Wiki
Jump to navigation Jump to search
Line 1: Line 1:
[[Category:Smarteiffel]]
 
 
This page specifies a tool that is currently being written.
 
Check the [http://gforge.inria.fr/plugins/scmsvn/viewcvs.php/trunk/?root=smarteiffel Subversion repository] if you want premium access to the tool.
 
 
 
<code>eiffeltest</code> is a tool that runs a suite of tests.
 
<code>eiffeltest</code> is a tool that runs a suite of tests.
   
 
Users can use it to run a test-suite of their own, which is a useful tool for their project robustness.
 
Users can use it to run a test-suite of their own, which is a useful tool for their project robustness.
   
It is also used by SmartEiffel itself to validate, with the provided test-suite, the compiler and the libraries.
+
It is also used by LibertyEiffel itself to validate, with the provided test-suite, the compiler and the libraries.
   
 
== Synopsis ==
 
== Synopsis ==

Revision as of 20:42, 5 September 2014

eiffeltest is a tool that runs a suite of tests.

Users can use it to run a test-suite of their own, which is a useful tool for their project robustness.

It is also used by LibertyEiffel itself to validate, with the provided test-suite, the compiler and the libraries.

Synopsis

  • se test directory
  • se test source_file.e [source_file2.e...]

File handling

The tools recursively iterates over the directory given as a command-line parameter, looking for test files or over the given test file names.

Test files are Eiffel source files with special names:

  • test_*.e: valid source file that should be compilable and runnable without causing an error
  • bad_*.e: invalid source file that should trigger a given compiler error message

For each test file, there can be optional files that have the same name but a different extension. These optional files can be used to provide:

  • output that the compiler is expected to provide when compiling bad_*.e files. This mandatory file for invalid tests has the extension .msg.
  • input to be fed to the program. This optional file (allowed only for valid tests) has the extension .in.
  • output that the program is expected to provide when run. This optional file (allowed only for valid tests) has the extension .out
  • glu_*.c: ***
  • cecil_*.se: ***
  • extra_options_*.se: ***

In each test directory a file named loadpath.se may exist, which is passed to the compiler as loadpath.

Why not remove the 'test_' and 'bad_' substrings from the test names? That reduces clutter and leaves more space for readable, self-documenting filenames.

The test harness can look for response files that match the test prefix. You already have two suffixes: *.out (for output) and *.msg (for a compiler error message). Another useful one is *.match (for a regular expression match on the output, because the output might contain things like date/time/version-number that are not relevant to the test).

That is how I'm doing testing for the Amber project, and it seems to work well. --Roger Browne 2005-11-26

That's a good point to raise, if only to mention that the current test suite (pre-eiffeltest) does a "fuzzy" matching of error messages (using SmartEiffel/tools/commands/same_msg.e).--FM 21:42, 19 Dec 2005 (CET)

Modes that are checked

Eiffeltest checks the following modes (presuming they are not disabled by the excluded.lst file)

  • -boost
  • -flat_check -debug_check
  • -no_check
  • -flat_check -require_check
  • -flat_check -ensure_check
  • -flat_check -invariant_check
  • -flat_check -loop_check
  • -flat_check -all_check
  • -no_gc -boost
  • -no_gc -flat_check -debug_check
  • -no_gc -no_check
  • -no_gc -flat_check -require_check
  • -no_gc -ensure_check
  • -no_gc -flat_check -invariant_check
  • -no_gc -flat_check -loop_check
  • -no_gc -flat_check -all_check
  • -debug_check
  • -require_check
  • -ensure_check
  • -invariant_check
  • -loop_check
  • -all_check
  • -no_gc -debug_check
  • -no_gc -require_check
  • -no_gc -ensure_check
  • -no_gc -invariant_check
  • -no_gc -loop_check
  • -no_gc -all_check
  • -profile -boost
  • -profile -debug_check
  • -profile -no_check
  • -profile -require_check
  • -profile -ensure_check
  • -profile -invariant_check
  • -profile -loop_check
  • -profile -all_check
  • -profile -no_gc -boost
  • -profile -no_gc -debug_check
  • -profile -no_gc -no_check
  • -profile -no_gc -require_check
  • -profile -no_gc -ensure_check
  • -profile -no_gc -invariant_check
  • -profile -no_gc -loop_check
  • -profile -no_gc -all_check
  • -profile -flat_check -debug_check
  • -profile -flat_check -require_check
  • -profile -flat_check -ensure_check
  • -profile -flat_check -invariant_check
  • -profile -flat_check -loop_check
  • -profile -flat_check -all_check
  • -no_split -boost
  • -no_split -no_gc -boost


Open issues

  • How can we use .in and .out files on platforms that do not allow redirecting the standard input and output of programs?

Easy: just finish the "exec" cluster ;-) --Cyril 08:16, 10 nov 2005 (CET)

Sounds good. Can you give a brief listing of what is to be done ? --pini 23:36, 25 nov 2005 (CET)

The win32 port (at least) is to be written. Ah, and testing ;-) --Cyril 15:39, 28 Nov 2005 (CET)

  • Which compilation modes are used for testing? There are many options and running all possibilities is probably too much.

Flags I can remember from: -no_gc, -flat_check, -no_split, (-debug_check | -all_check | -loop_check | -invariant_check | -ensure_check | -require_check | -no_check | -boost) This means compiling and running about 64 times (minus incompatibilities between -flat_check and boost/no_check).

  • How to test some specific capabilities: options -profile, -no_main, -c_mode, -sedb, -cecil, or testing with c2j?
  • What about tests specific to some compiler mode? For example GC tests should not be run without GC, optimizer tests are invalid in modes other than -boost...

Perhaps we could use indexing/notes clauses with tags indicating which modes/options are relevant for the test? --Dmoisset 22:44, 26 Nov 2005 (CET)

Nice idea. I would improve it by proposing the opposite: list the modes/options to exclude. This means that all modes will be checked by default. If the test is incompatible with some mode, then it has to be specified. If some new mode is added, this mode will be tested with all tests. If some tests are not compatible with this mode, they have to be excluded. Otherway, you would have to edit about every test to add this mode. -- Philippe Ribet