Doing it in the pig script is not feasible because pig doesn't have any
notion of sequentiality - to maintain it, you'd need to have access to
state that's shared globally by all the mappers and reducers. One way I can
think of doing this is to have a UDF that maintains state - perhaps it can
maintain a file that's NFS mounted/or in HDFS so that it's available on all
the task nodes; then any call to the UDF can update that file (atomically)
and return a 'row number' that you could associate with your current tuple.
Something like:
B = FOREACH A GENERATE $0, $1, $2, $3, MyUDFs.GETROWNUM() as rownum;

However, AFAIK, you'd be better off doing it in HBase - perhaps at the time
of record insert, you could also add a 'row number' into the record?

On 22 May 2012 12:43, Mohammad Tariq <donta...@gmail.com> wrote:

> Hi Abhinav,
>
>   Thanks a lot for the valuable response..Actually I was thinking of
> doing the same thing, but being new to Pig I thought of asking it on
> the mailing list first..As far as the data is concerned, second column
> will always be in ascending order.But I don't think it will be of any
> help..I think whatever you have suggested here would be the
> appropriate solution..Although I would like to ask you one thing..Is
> it feasible to add that first column having count in my pig script or
> do I have to change the data in my Hbase table itself???If yes then
> how can I achieve it in my script??Many thanks.
>
> Regards,
>     Mohammad Tariq
>
>
> On Tue, May 22, 2012 at 1:16 AM, Abhinav Neelam <abhinavroc...@gmail.com>
> wrote:
> > Hey Mohammad,
> >
> > You need to have sorting requirements when you say 'top 5' records.
> Because
> > relations/bags in Pig are unordered, it's natural to ask: 'top 5 by what
> > parameter?' I'm unfamiliar with HBase, but if your data in HBase has an
> > implicit ordering with say an auto-increment primary key, or an explicit
> > one, you could include that field in your input to Pig and then apply TOP
> > on that field.
> >
> > Having said that, if I understand your problem correctly, you don't need
> > TOP at all - you just want to process your input in groups of 5 tuples
> at a
> > time. Again, I can't think of a way of doing this without modifying your
> > input. For example, if your input included an extra field like this:
> > 1 18.98   2000             1.21   193.46  2.64        58.17
> > 1 52.49   2000.5   4.32           947.11  2.74        64.45
> > 1 115.24  2001             16.8   878.58  2.66        94.49
> > 1 55.55   2001.5   33.03  656.56  2.82        60.76
> > 1 156.14  2002             35.52  83.75   2.6         59.57
> > 2 138.77  2002.5   21.51  105.76  2.62        85.89
> > 2 71.89   2003             27.79  709.01  2.63        85.44
> > 2 59.84   2003.5   32.1           444.82  2.72        70.8
> > 2 103.18  2004             4.09   413.15  2.8         54.37
> >
> > you could do a group on that field and proceed. Even if you had a field
> > like 'line number' or 'record number' in your input, you could still
> > manipulate that field (say through integer division by 5) to use it for
> > grouping. In any case, you need something to let Pig bring together your
> 5
> > tuple groups.
> >
> > B = group A by $0;
> > C = FOREACH B { <do some processing on your 5 tuple bag A> ...
> >
> > Thanks,
> > Abhinav
> >
> > On 21 May 2012 23:03, Mohammad Tariq <donta...@gmail.com> wrote:
> >
> >> Hi Ruslan,
> >>
> >>    Thanks for the response.I think I have made a mistake.Actually I
> >> just want the top 5 records each time.I don't have any sorting
> >> requirements.
> >>
> >> Regards,
> >>     Mohammad Tariq
> >>
> >>
> >> On Mon, May 21, 2012 at 9:31 PM, Ruslan Al-fakikh
> >> <ruslan.al-fak...@jalent.ru> wrote:
> >> > Hey Mohammad,
> >> >
> >> > Here
> >> > c = TOP(5,3,a);
> >> > you say: take 5 records out of a that have the biggest values in the
> >> third
> >> > column. Do you really need that sorting by the third column?
> >> >
> >> > -----Original Message-----
> >> > From: Mohammad Tariq [mailto:donta...@gmail.com]
> >> > Sent: Monday, May 21, 2012 3:54 PM
> >> > To: user@pig.apache.org
> >> > Subject: How to use TOP?
> >> >
> >> > Hello list,
> >> >
> >> >  I have an Hdfs file that has 6 columns that contain some data stored
> in
> >> an
> >> > Hbase table.the data looks like this -
> >> >
> >> > 18.98   2000             1.21   193.46  2.64        58.17
> >> > 52.49   2000.5   4.32           947.11  2.74        64.45
> >> > 115.24  2001             16.8   878.58  2.66        94.49
> >> > 55.55   2001.5   33.03  656.56  2.82        60.76
> >> > 156.14  2002             35.52  83.75   2.6         59.57
> >> > 138.77  2002.5   21.51  105.76  2.62        85.89
> >> > 71.89   2003             27.79  709.01  2.63        85.44
> >> > 59.84   2003.5   32.1           444.82  2.72        70.8
> >> > 103.18  2004             4.09   413.15  2.8         54.37
> >> >
> >> > Now I have to take each record along with its next 4 records and do
> some
> >> > processing(for example, in the first shot I have to take records 1-5,
> in
> >> the
> >> > next shot I have to take 2-6 and so on)..I am trying to use TOP for
> this,
> >> > but getting the following error -
> >> >
> >> > 2012-05-21 17:04:30,328 [main] ERROR org.apache.pig.tools.grunt.Grunt
> >> > - ERROR 1200: Pig script failed to parse:
> >> > <line 6, column 37> Invalid scalar projection: parameters : A column
> >> needs
> >> > to be projected from a relation for it to be used as a scalar Details
> at
> >> > logfile: /home/mohammad/pig-0.9.2/logs/pig_1337599211281.log
> >> >
> >> > I am using following commands -
> >> >
> >> > grunt> a = load 'hbase://logdata'
> >> >>> using org.apache.pig.backend.hadoop.hbase.HBaseStorage(
> >> >>> 'cf:DGR cf:HD cf:POR cf:RES cf:RHOB cf:SON', '-loadKey true') as
> (id,
> >> >>> DGR, HD, POR, RES, RHOB, SON);
> >> > grunt> b = foreach a { c = TOP(5,3,a);
> >> >>> generate flatten(c);
> >> >>> }
> >> >
> >> > Could anyone tell me how to achieve that????Many thanks.
> >> >
> >> > Regards,
> >> >     Mohammad Tariq
> >> >
> >>
> >
> >
> >
> > --
> > Hacking is, and always has been, the Holy
> > Grail of computer science.
>



-- 
Hacking is, and always has been, the Holy
Grail of computer science.

Reply via email to