On Sat, Apr 11, 2009 at 2:02 PM, Ashley Moran
<ashley.mo...@patchspace.co.uk> wrote:
>
> I was just idly thinking, could a code-coverage based system could be
> combined with some sort of failure (fragility) history to balance the time
> cost of heavy feature runs with the benefits of having something run
> end-to-end?  We've had reverse-modification-time spec ordering for ages
> which is a useful start.

I've had it in my head for a while now that someday (yes, that
mythical 'someday') I want to write a better autotest.  Maybe this is
heresy, but I am a huge fan of the _idea_ behind autotest and totally
annoyed by its implementation.  It's extensible only in strange ways
(hence wrappers like autospec), and its fundamental strategy is too
static.  I once lost most of a day trying to fix merb_cucumber so the
features would run when they should, and was ready to hurl cats when I
realized autotest's idea of context chaining was to make you list them
all in the classname in alphabetical order.  Look at the files in the
Cucumber gem's 'lib/autotest' directory and you'll see what I mean.

A proper design would let you plug in your own file-change discovery
strategy, plug in multiple runners (RSpec, Cucumber, yadda yadda) with
true modularity, specify lists of observers on directories or files,
and allow different output views.  An _ideal_ design would also let
you set priority rules like you're describing here, so you get instant
feedback only on the stuff you're working with, and do end-to-end runs
in the background.

Right now this is just a pipe dream, but I don't think it would be
_hard._  It's just finding the time to do it vs. actual public-facing
applications that's the challenge.  If anybody wants to have a
conversation about this, maybe get some collaboration going, feel free
to drop me a line.


> On a more ranty note - I have very little time for these "XXX
> BDD/development technique is always bad, don't do it" articles.  (But hey,
> maybe I was guilty of this myself and have forgotten since...)

"Declaring absolutes is always bad, don't do it?"  >8->

Oh -- one other thought I had from reflecting upon your e-mail.  This
is totally unrelated to the above, but since we're being Big Thinkers
I might as well write it down before I forget.  You mentioned
fragility/failure history, in relation to coverage, and I started
thinking...  I wonder if everyone's going about test coverage from the
wrong direction, simply trying to _anticipate_ failure?  What if we
extended something like exception_notifier or Hoptoad as well, and
brought real exceptions from the application's staging and production
environments into our test tools?  We know from the stack traces where
failures occur, so it'd be pretty straightforward to write an
RCov-like utility that nagged you: "You dingbat, your specs totally
fail to cover line 119 of hamburgers.rb.  It threw a MeatNotFound
exception last Tuesday.  Gonna test for that?  Ever?"

What do you think?  Decent idea?  Or does something like this already
exist and I don't know about it?


-- 
Have Fun,
   Steve Eley (sfe...@gmail.com)
   ESCAPE POD - The Science Fiction Podcast Magazine
   http://www.escapepod.org
_______________________________________________
rspec-users mailing list
rspec-users@rubyforge.org
http://rubyforge.org/mailman/listinfo/rspec-users

Reply via email to