Hi everyone,

As a QA engineer, I had a few thoughts I would put into the discussion as well.  Please feel free to take them or leave them (or to throw them a parade and give them the key to the city) ;-)

On Mon, Apr 6, 2009 at 3:17 PM, Kristian Nielsen <knielsen@knielsen-hq.org> wrote:
Michael Widenius <michael.widenius@gmail.com> writes:

> To go trough the things that are wrong with mysqltest:

In my opinion, the single biggest problem with mysqltest is that it is a
seperate program from the test suite driver mysql-test-run.pl. It's even
written in a different language (C vs. Perl), and test cases themselves are
written in a third language (the mysqltest language) + a mixture of SQL and
even some shell scripts. That's _5_ different languages in a single
application :-(

This is a disaster for the structure of mysql-test-run. Such a large part of
the code (and complexity) is just trying to handle communication between the
various parts, which is really hard to do well due to different languages.

The first step in any major improvement (in my opinion) should be to
re-implement mysqltest in Perl inside mysql-test-run.pl. But that would be a
project of several months, not something I'm eager to start just now ...

I agree that more 'moving parts' = more overhead.  I can't say that I agree with re-implementing in this fashion, but now would be a good time to see how we can make things leaner / more efficient / more to-the-point (ie we can focus on using the tool to test and not 'keeping the parts in harmony') 


> I like the philosophy behind it:  Write a test in pure SQL and then
> ensure that the generated result doesn't change.  It's easy enough to
> allow any developers to generate a test case.

Yes. This is a very strong point for testing that correct results are produced
by a query or similar. Keeping something like this is very important if
thinking about extending or replacing the test framework.

The weakness with mysqltest / mysql-test-run is not that it does this, it is
that it does not provide additional/separate functinality for other kinds of
tests that do not fit this simple framework well.

I very much like the straight-up SQL-> result paradigm *for certain tests*.  This is GREAT for a basic, functional suite (does SELECT work.  Are INSERT DEFAULT and INSERT () interchangeable <in certain scenarios>).  However, getting into more 'scenario-based' testing, there are problems - coordination between 'users' in a test, tools for greybox-manipulation (tweaking the software mechanics beyond just INPUT/OUTPUT testing), etc.  

It would be great to keep the simplicity for certain scenarios, but to have extensible power for more complicated tests when needed.  One issue I have seen with certain proposed replacements is that they allow complex testing, but the overhead of that power results in even the most simple tests becoming a monster (the overhead of defining a test).

> In particular, it's good that tests and results are different files as
> it makes it easier to understand what are wrong when you get a failure.

I am trying to understand what precisely you have in mind here...

One big problem with other test suites I have seen is that if you want to
debug a failure, you first have to spend a lot of time debugging and
understanding the test case code just to understand what the problem actually
is. Getting a diff of queries run and results produced like mysqltest does
makes this step trivial (if the problem is wrong results from a query).

Also the --record functionality for generating .result files can be really
good to ease the production of test case and also make it more likely the
developer will not be tempted to insufficiently check the results (everything
is checked by default). It does have the big risk of not properly checking
that the committed .result is in fact correct, something I have seen more than
once.

This does not really have to mean using separate files. I have in mind
something like a Perl-based suite where I could do:

   query <<SQL, <<RESULT
   SELECT a, b
     FROM t1 JOIN t2 ON (a = b)
    ORDER BY a, b;
   SQL
   RECORDME
   RESULT

and `mysql-test-run --record` would rewrite this as

   query <<SQL, <<RESULT
   SELECT a, b
     FROM t1 JOIN t2 ON (a = b)
    ORDER BY a, b;
   SQL
    a  b
    1  1
    2  2
    4  4
   RESULT

And in case of different result, it could output a diff of just the RESULT
section of the failing query, along with the query. It could also give the
correct line number in the .test file, avoiding the need to search the .test
file for the query that produced the diff. I think that might make it even
easier than separate result file.

Something like that would of course allow putting arbitrary Perl code
in-between running queries (with lots of useful modules available) where this
is more appropriate. This would eliminate the need for the mysqltest language
for new test cases, which is one less language again.

And it would be quite easy to have a backwards-compatible emulation of the old
.test and .result files (just a small wrapper that reads the .test and .result
files and passes the contents to the query() function).

I just gave a MySQL University session on some of these problems.  One of the issues I have seen is that people don't expend the extra energy to make a test informative (through comments explaining the test, defining what a 'real' <not in setup / teardown, but in the behavior being tested> failure would look like / what 'next-steps' seem logical for diagnosing a failure, etc.
http://forge.mysql.com/wiki/How_to_Create_a_Test_Case

I do like the thought of defining query / result portions a LOT.  One of the biggest time-sinks we have is sifting through more complicated tests and piecing together what is going on.

I think a properly revised test-suite would help if it presented the ability to capture much more information:
Test purpose
Defining key portions of the test - maybe defining code blocks of a test as setup / teardown / and testcode.  Each testcode portion could have things like  "on_failure - If the test fails here, it is a problem with xyz."  Each 'real' portion of the test could define this.  Failures in setup / teardown are still bad, but they might point to a less-serious failure and could be a Build host issue (full filesystem, etc).

Probably the biggest wishes for a testing tool I have are:
Power
Simplicity / Elegance (make it easy to do anything, and allow complex 'thoughts' to be spelled out in an easily understood way)
Ease of comprehending a test (opening it up and understanding what is being tested, how it is being done, and why a method has been chosen)
Speed of failure analysis. (seeing a failure and being able to quickly discern what failed and what kind of a problem it is pointing to).

Some extra wishlist items:
Ease of cataloging / cross-referencing tests (ie run all 'Innodb' related tests and not having this rely solely on suite organization / naming patterns)
Execution speed (we are likely to expand the test suite, so ensuring a lean run-time is always nice)

Anyway, I saw the topic and couldn't resist chiming in.

Thanks,
Patrick.



>   One benefit is that you can usually run a test by just piping it
>   trough the normal MySQL client, which simplifies testing and
>   debugging a lot!

Yes, this should be kept for simple stuff, very important.

 - Kristian.

_______________________________________________
Mailing list: https://launchpad.net/~maria-developers
Post to     : maria-developers@lists.launchpad.net
Unsubscribe : https://launchpad.net/~maria-developers
More help   : https://help.launchpad.net/ListHelp