[TIP] Meta-test methods...
Scott David Daniels
Scott.Daniels at Acm.Org
Sun Apr 26 07:42:51 PDT 2009
Douglas Philips wrote:
> Ok, enough background... :)
> In the context of one test method with hundreds of scenarios,
> unless every scenario passes, we lose information about the nature,
> severity, and extent of the problems/failures. Maybe the first few
> devices we tested had dozens of scenario failures in a particular
> test. A new version of the device might fix some issues and have fewer
> (scenario) failures. However, that test's results won't look any
> different. Worse case is if only some of original issues are fixed and
> new issues are introduced, and so even the total number of scenario
> failures might remain the same. For these kinds of tests we have to
> turn on tracing and manually check the results. :(
>
Sounds like the "test" is really a test class, while the scenario is
a test case. A test should test one thing and be "independent."
Naming the scenarios separately, so they can be reported separately,
seems part of the issue here. Is there a way to run only a few of the
scenario elements without the entire set?
--Scott David Daniels
Scott.Daniels at Acm.Org
More information about the testing-in-python
mailing list