Sorry, for replying the private email to the mailing list, but I strongly
believe in leaving the next guy something to google ;)
Anyway, as you seem to be knowledgeable about sorting, one question:
Does hadoop provide all key/value tuples for a given key in one batch to the
reducer, or not?
For me, I had to upgrade to 0.17.0, which made this problem go away magically.
No idea if that will solve your problem.
Andreas
On Thursday 12 June 2008 23:04:17 Rob Collins wrote:
In a previous life, I had no problems setting up a small cluster. Now I
have managed to mess it up. I see reports
Thanks for all advices. :)
Edward
On Fri, Jun 13, 2008 at 3:22 PM, lohit [EMAIL PROTECTED] wrote:
Check RandomWriter.java
look for reporter.setStatus(wrote record + itemCount + ..
- Original Message
From: Edward J. Yoon [EMAIL PROTECTED]
To:
Usually hadoop programs are not used interactively since what they excel at
is batch operations on very large collections of data.
It is quite reasonable to store resulting data in hadoop and access those
results using hadoop. The cleanest way to do that is to have a presentation
layer web
Yes. It does.
This can be the source of some imbalanced load in the reducer, but is
essential to correct functioning of the map-reduce model.
Sometimes it is good to actually add some additional salt to the key just so
that the large reduce lists get split up a bit for more equitable