On 1/31/06, Mark Stosberg <[EMAIL PROTECTED]> wrote: > Here's my test-first TODO test management paradox: > > If I write a failing test and share it through the central repo, > the smoke bot fails and keeps sending us e-mail until it is fixed, > which can be annoying when these are un-implemented features and not > bugs. The effect can be quit paying attention to the smoke bot. > > If I mark the test TODO, the smokebot succeeds and the test disappears > from the radar of tests that should be fixed soon. > > What's a good way to manage TODO tests so that they continue to be > noticed and worked on, but without being annoying? > > Partly I wish that the reporting tools provided more detail about TODO > tests. Rather than just telling me that X TODO tests passed, I'd like > to know exactly what they were and where they were located so I can go > work on them. > > I also realize I have another class of TODO tests, it's the: > > Ill-get-to-it-eventually,-maybe-next-year class of TODO tests. > > These are things that I've noted I'd like to have an automated test for, > but the tests are long term because they are expensive, difficult to > setup, or well, I'm imperfect. > > Maybe being able to add a "due date" to tests would help. :) > > The TODO tests would pass before the due date, but if they aren't > addressed in flow of work, they start failing to bring attention to > themselves. > > And then there could be a "snooze" button too...
While only indirectly related..... Lately when ive built blead I see stuff like "10 TODO tests unexpectedly passed!" So far ive not found a way to make harness tell me which they are... So i too would like a report of TODO tests, both those passing and those failing. yves -- perl -Mre=debug -e "/just|another|perl|hacker/"