I'm curious to hear what people think on this. I'm a really big fan of spinning up a minicluster instance to do some "more real" testing of software as I write it.
With 1.5.0, it's a bit more painful because I have to add a bunch more dependencies to my project (which previously would only have to depend on the accumulo-minicluster artifact). The list includes, but is likely not limited to, commons-io, commons-configuration, hadoop-client, zookeeper, log4j, slf4j-api, slf4j-log4j12. Best as I understand it, the intent of this was that Hadoop will typically provide these artifacts at runtime, and therefore Accumulo doesn't need to re-bundle them itself which I'd agree with (not getting into that whole issue about the Hadoop "ecosystem"). However, I would think that the minicluster should have non-provided scope dependencies declared on these, as there is no Hadoop installation -- there's just the minicluster. As such, this would alleviate users from having to dig into our dependency management or trial&error to figure out what "extra" dependencies they have to include in their project to actually make it work Thoughts? - Josh