On 1/31/06, Mark Stosberg <[EMAIL PROTECTED]> wrote:
> Here's my test-first TODO test management paradox:
>
> If I write a failing test and share it through the central repo,
> the smoke bot fails and keeps sending us e-mail until it is fixed,
> which can be annoying when these are un-implemented features and not
> bugs. The effect can be quit paying attention to the smoke bot.
>
> If I mark the test TODO, the smokebot succeeds and the test disappears
> from the radar of tests that should be fixed soon.
>
> What's a good way to manage TODO tests so that they continue to be
> noticed and worked on, but without being annoying?
>
> Partly I wish that the reporting tools provided more detail about TODO
> tests. Rather than just telling me that X TODO tests passed, I'd like
> to know exactly what they were and where they were located so I can go
> work on them.
>
> I also realize I have another class of TODO tests, it's the:
>
> Ill-get-to-it-eventually,-maybe-next-year class of TODO tests.
>
> These are things that I've noted I'd like to have an automated test for,
> but the tests are long term because they are expensive, difficult to
> setup, or well, I'm imperfect.
>
> Maybe being able to add a "due date" to tests would help. :)
>
> The TODO tests would pass before the due date, but if they aren't
> addressed in flow of work, they start failing to bring attention to
> themselves.
>
> And then there could be a "snooze" button too...

While only indirectly related.....

Lately when ive built blead I see stuff like "10 TODO tests
unexpectedly passed!"

So far ive not found a way to make harness tell me which they are...

So i too would like a report of TODO tests, both those passing and
those failing.

yves

--
perl -Mre=debug -e "/just|another|perl|hacker/"

Reply via email to