Hello folks,
I have a lot of input splits (10k-50k - 128 mb blocks) which contains text
files. I need to process those line by line, then copy the result into
roughly equal size of shards.
So i generate a random key (from a range of [0:numberOfShards]) which is
used to route the map output to
Hi,
I am trying to install Hbql on pseudo distributed node. I am not sure how
to build the *hbase-trx-0.90.0-DEV-2.jar* from hbase-transactional package
which was downloaded from *
https://github.com/hbase-trx/hbase-transactional-tableindexed*
Appreciate your help on the same.
--
Thanks
Hi Mete
A custom Paritioner class can control the flow of keys to the desired reducer.
It gives you more control on which key to which reducer.
Regards
Bejoy KS
Sent from handheld, please excuse typos.
-Original Message-
From: mete efk...@gmail.com
Date: Fri, 27 Apr 2012 09:19:21
Unfortunately in 1.x the format command's prompt is case-sensitive
(Fixed in 2.x):
You had:
Re-format filesystem in /app/hadoop/name ? (Y or N) y
Format aborted in /app/hadoop/name
Answer with a capital Y instead and it won't abort.
On Fri, Apr 27, 2012 at 3:07 PM, Mathias Schnydrig
Hi,
I am pretty a newbie and i am following the quick start guide for single
node set up on windows using cygwin.
In this step,
$ bin/hadoop fs -put conf input
I am getting the following errors.
I have got no files
under /user/EXT0125622/input/conf/capacity-scheduler.xml. That might be a
Can you run a regular 'hadoop fs' (put orls or get) command?
If yes, how about a wordcount example?
'path/hadoop jar pathhadoop-*examples*.jar wordcount input output'
-Original Message-
From: Mohit Anchlia mohitanch...@gmail.com
Reply-To: common-user@hadoop.apache.org
After all the jobs fail I can't run anything. Once I restart the cluster I
am able to run other jobs with no problems, hadoop fs and other io
intensive jobs run just fine.
On Fri, Apr 27, 2012 at 3:12 PM, John George john...@yahoo-inc.com wrote:
Can you run a regular 'hadoop fs' (put orls or
Hi all,
I am a newbie in Hadoop and I like the system. I have one question: Is
there a node-wide combiner or something similar in Hadoop? I think it
can reduce the number of intermediate results in further. Any hint?
Thanks a lot!
Superymk