Thanks Harsh.

But will it also sort the data as Partitioner does.

On Sun, Feb 19, 2012 at 10:54 PM, Harsh J <ha...@cloudera.com> wrote:

> Hi,
>
> You would find it easier to use the Java API's MultipleOutputs (and/or
> MultipleOutputFormat, which directly works on a configured key field),
> to write each key-partition out in its own file.
>
> On Mon, Feb 20, 2012 at 7:38 AM, Piyush Kansal <piyush.kan...@gmail.com>
> wrote:
> > Hi Friends,
> >
> > I have to sort huge amount of data in minimum possible time probably
> using
> > partitioning. The key is composed of 3 fields(partition, text and
> number).
> > This is how partition is defined:
> >
> > Partition "1" for range 1-10
> > Partition "2" for range 11-20
> > Partition "3" for range 21-30
> >
> > I/P file format: partition[tab]text[tab]range-start[tab]range-end
> >
> > [cloudera@localhost kMer2]$ cat input1
> >
> > 1 chr1 1 10
> > 1 chr1 2 8
> > 2 chr1 11 18
> >
> > [cloudera@localhost kMer2]$ cat input2
> >
> > 1 chr1 3 7
> > 2 chr1 12 19
> >
> > [cloudera@localhost kMer2]$ cat input3
> >
> > 3 chr1 22 30
> >
> > [cloudera@localhost kMer2]$ cat input4
> >
> > 3 chr1 22 30
> > 1 chr1 9 10
> > 2 chr1 15 16
> >
> > Then I ran following command:
> >
> > hadoop jar
> > /usr/lib/hadoop/contrib/streaming/hadoop-streaming-0.20.2-cdh3u2.jar \
> > -D stream.map.output.field.separator='\t' \
> > -D stream.num.map.output.key.fields=3 \
> > -D map.output.key.field.separator='\t' \
> > -D mapred.text.key.partitioner.options=-k1 \
> > -D mapred.reduce.tasks=3 \
> > -input /usr/pkansal/kMer2/ip \
> > -output /usr/pkansal/kMer2/op \
> > -mapper /home/cloudera/kMer2/kMer2Map.py \
> > -file /home/cloudera/kMer2/kMer2Map.py \
> > -reducer /home/cloudera/kMer2/kMer2Red.py \
> > -file /home/cloudera/kMer2/kMer2Red.py
> >
> > Both mapper and reducer scripts just contain one line of code:
> >
> > for line in sys.stdin:
> >     line = line.strip()
> >     print "%s" % (line)
> >
> > Following is the o/p:
> >
> > [cloudera@localhost kMer2]$ hadoop dfs -cat
> /usr/pkansal/kMer2/op/part-00000
> >
> > 2 chr1 12 19
> > 2 chr1 15 16
> > 3 chr1 22 30
> > 3 chr1 22 30
> >
> > [cloudera@localhost kMer2]$ hadoop dfs -cat
> /usr/pkansal/kMer2/op/part-00001
> >
> > 1 chr1 2 8
> > 1 chr1 3 7
> > 1 chr1 9 10
> > 2 chr1 11 18
> >
> > [cloudera@localhost kMer2]$ hadoop dfs -cat
> /usr/pkansal/kMer2/op/part-00002
> >
> > 1 chr1 1 10
> > 3 chr1 22 29
> >
> > This is not the o/p which I expected. I expected all records with:
> >
> > partition 1 in one single file eg part-m-00000
> > partition 2 in one single file eg part-m-00001
> > partition 3 in one single file eg part-m-00002
> >
> > Can you please suggest if I am doing it in a right way?
> >
> > --
> > Regards,
> > Piyush Kansal
> >
>
>
>
> --
> Harsh J
> Customer Ops. Engineer
> Cloudera | http://tiny.cloudera.com/about
>



-- 
Regards,
Piyush Kansal

Reply via email to