On Sat, Apr 11, 2009 at 2:02 PM, Ashley Moran <ashley.mo...@patchspace.co.uk> wrote: > > I was just idly thinking, could a code-coverage based system could be > combined with some sort of failure (fragility) history to balance the time > cost of heavy feature runs with the benefits of having something run > end-to-end? We've had reverse-modification-time spec ordering for ages > which is a useful start.
I've had it in my head for a while now that someday (yes, that mythical 'someday') I want to write a better autotest. Maybe this is heresy, but I am a huge fan of the _idea_ behind autotest and totally annoyed by its implementation. It's extensible only in strange ways (hence wrappers like autospec), and its fundamental strategy is too static. I once lost most of a day trying to fix merb_cucumber so the features would run when they should, and was ready to hurl cats when I realized autotest's idea of context chaining was to make you list them all in the classname in alphabetical order. Look at the files in the Cucumber gem's 'lib/autotest' directory and you'll see what I mean. A proper design would let you plug in your own file-change discovery strategy, plug in multiple runners (RSpec, Cucumber, yadda yadda) with true modularity, specify lists of observers on directories or files, and allow different output views. An _ideal_ design would also let you set priority rules like you're describing here, so you get instant feedback only on the stuff you're working with, and do end-to-end runs in the background. Right now this is just a pipe dream, but I don't think it would be _hard._ It's just finding the time to do it vs. actual public-facing applications that's the challenge. If anybody wants to have a conversation about this, maybe get some collaboration going, feel free to drop me a line. > On a more ranty note - I have very little time for these "XXX > BDD/development technique is always bad, don't do it" articles. (But hey, > maybe I was guilty of this myself and have forgotten since...) "Declaring absolutes is always bad, don't do it?" >8-> Oh -- one other thought I had from reflecting upon your e-mail. This is totally unrelated to the above, but since we're being Big Thinkers I might as well write it down before I forget. You mentioned fragility/failure history, in relation to coverage, and I started thinking... I wonder if everyone's going about test coverage from the wrong direction, simply trying to _anticipate_ failure? What if we extended something like exception_notifier or Hoptoad as well, and brought real exceptions from the application's staging and production environments into our test tools? We know from the stack traces where failures occur, so it'd be pretty straightforward to write an RCov-like utility that nagged you: "You dingbat, your specs totally fail to cover line 119 of hamburgers.rb. It threw a MeatNotFound exception last Tuesday. Gonna test for that? Ever?" What do you think? Decent idea? Or does something like this already exist and I don't know about it? -- Have Fun, Steve Eley (sfe...@gmail.com) ESCAPE POD - The Science Fiction Podcast Magazine http://www.escapepod.org _______________________________________________ rspec-users mailing list rspec-users@rubyforge.org http://rubyforge.org/mailman/listinfo/rspec-users