Re: Converting timestamp to date format

2012-07-20 Thread Techy Teck
Thanks for the suggestion. I will use BIGINT now. On Fri, Jul 20, 2012 at 3:48 PM, kulkarni.swar...@gmail.com < kulkarni.swar...@gmail.com> wrote: > BIGINT is 8 bytes whereas INT is 4 bytes. Timestamps are usually of "long" > type. To avoid loss of precision, I would recommend BIGINT. > > > On

Re: Invalid Table Alias

2012-07-20 Thread Techy Teck
In hive we cannot calculate the difference between dates in minutes? On Fri, Jul 20, 2012 at 6:06 PM, Techy Teck wrote: > Whenever I am running the below query > > SELECT TT.BUYER_ID , COUNT(*) FROM > (SELECT testingtable1.buyer_id, testingtable1.item_id, > testingtable1.created_time from (se

Re: Both Left and Right Aliases Encountered in Join

2012-07-20 Thread Navis류승우
Can you try it with, *UNIX_TIMESTAMP(table1.**created_time) = **prod.timestamps + (15 * **60* 1000)* 2012/7/21 Techy Teck > My full query is like this- The thing that is in red color is giving me > exception. And I already have where clause after the JOIN. Not sure how to > make this thing work

Re: Both Left and Right Aliases Encountered in Join

2012-07-20 Thread Techy Teck
My full query is like this- The thing that is in red color is giving me exception. And I already have where clause after the JOIN. Not sure how to make this thing work. *SELECT TT.BUYER_ID , COUNT(*) FROM* *(SELECT table1.buyer_id, table1.item_id, table1.created_time from prod RIGHT OUTER JOIN tab

Re: Both Left and Right Aliases Encountered in Join

2012-07-20 Thread Bejoy KS
Hi AFAIK hive supports only equi joins. Can you check whether changing that <= to = , removes the semantic error? You can give conditions other than equality in a where clause post join. Regards Bejoy KS Sent from handheld, please excuse typos. -Original Message- From: Techy Teck Dat

Re: Converting timestamp to date format

2012-07-20 Thread kulkarni.swar...@gmail.com
BIGINT is 8 bytes whereas INT is 4 bytes. Timestamps are usually of "long" type. To avoid loss of precision, I would recommend BIGINT. On Fri, Jul 20, 2012 at 4:52 PM, Tech RJ wrote: > What is the difference between these two? Trying to convert timestamps to > full date format. The only differen

Re: Error while reading from task log url

2012-07-20 Thread Raihan Jamal
Yup, Thanks it worked. *Raihan Jamal* On Fri, Jul 20, 2012 at 1:40 PM, Bejoy KS wrote: > ** > Raihan > > To see the failed task logs in hadoop, the easiest approach is > drilling down the jobtracker web UI. > > Go to the job url (which you'll get in the beginning of a job on your > console,

Both Left and Right Aliases Encountered in Join

2012-07-20 Thread Techy Teck
I have a join query on which I am joining like this below. And it is working fine ON (table1.item = prod.product *AND* table1.buyer = prod.user *AND* UNIX_TIMESTAMP(table1.tt_time) = prod.timestamps) But when I tried to modify the above same thing. The thing that is in red color will work or not

Re: Hive 0.10 release date

2012-07-20 Thread kulkarni.swar...@gmail.com
Thanks for the advice Edward. That makes sense to me. As a side note, while doing some searching, I stumbled upon your blog[1] regarding the release which made me even more curious. :) [1] http://www.edwardcapriolo.com/roller/edwardcapriolo/entry/when_hive_0_10_release On Fri, Jul 20, 2012 at 4:

Re: Hive 0.10 release date

2012-07-20 Thread Edward Capriolo
Usually it is ok to build a trunk, we do not do anything extra special for releases other then some basic sanity testing and cut an svn tag. (we run the full unit tests every commit). The only time this advice is not true is if there are some metastore changes, however the scripts to handle the upg

Hive 0.10 release date

2012-07-20 Thread kulkarni.swar...@gmail.com
Hello, I totally understand that usually open source projects to not have a fixed date for release but I was just curious if something was chalked out for releasing hive 0.10 out in the wild. There are some really interesting additions that I am looking forward to. Thanks, -- Swarnim

Re: Error while reading from task log url

2012-07-20 Thread Bejoy KS
Raihan To see the failed task logs in hadoop, the easiest approach is drilling down the jobtracker web UI. Go to the job url (which you'll get in the beginning of a job on your console, by the name Tracking url) http://ares-jt.vip.ebay.com:50030/jobdetails.jsp?jobid=job_201207172005_14407 Brow

Re: Error while reading from task log url

2012-07-20 Thread Raihan Jamal
I tried opening the below URL, and nothing got opened, I got page cannot be displayed. Why is that so? *Raihan Jamal* On Fri, Jul 20, 2012 at 12:39 PM, Sriram Krishnan wrote: > What version of Hadoop and Hive are you using? We have seen errors like > this in the past – and you can actually

Re: Error while reading from task log url

2012-07-20 Thread Raihan Jamal
After setting this in Hive- hive> SET hive.exec.show.job.failure.debug.info=false; I can see the logs on my console itself? Or I need to go somewhere to see the actual logs and what is causing the problem? *Raihan Jamal* On Fri, Jul 20, 2012 at 12:28 PM, kulkarni.swar...@gmail.com < kulkarni

Re: Error while reading from task log url

2012-07-20 Thread Sriram Krishnan
What version of Hadoop and Hive are you using? We have seen errors like this in the past – and you can actually replace taskid with attemptid to fetch your logs. So try this: http://lvsaishdc3dn0857.lvs.ebay.com:50060/tasklog?attemptid=attempt_201207172005_14407_r_00_1&all=true But yes, th

Error while reading from task log url

2012-07-20 Thread Raihan Jamal
Whenever I run the below query- * * *SELECT buyer_id, item_id, ranknew(buyer_id, item_id), created_time* *FROM (* *SELECT buyer_id, item_id, created_time* *FROM testingtable1* *DISTRIBUTE BY buyer_id, item_id* *SORT BY buyer_id, item_id, created_time desc* *) a* *WHERE ranknew(buyer

Re: Disc quota exceeded

2012-07-20 Thread kulkarni.swar...@gmail.com
*rpool/tmp 10G10G 0K 100%/tmp* * * This might be the source of your problem as I mentioned earlier. Try freeing some space here and then try again. On Fri, Jul 20, 2012 at 11:34 AM, comptech geeky wrote: > After trying "df -kh". I got below result. > > *bash-3.00$ df -

Re: Disc quota exceeded

2012-07-20 Thread comptech geeky
After trying "df -kh". I got below result. *bash-3.00$ df -kh* *Filesystem size used avail capacity Mounted on* *rpool/ROOT/sol10 916G30G 668G 5%/* */devices 0K 0K 0K 0%/devices* *ctfs 0K 0K 0K 0%

Re: What will be the Rank UDF for this particular scenario?

2012-07-20 Thread comptech geeky
Can anyone help me on this particular problem? I am still stuck on this from so long. I have modified my Rank function again for my scenario. But not sure whether this will work or not. Any suggestions? *public final class RankAnother extends UDF{* *private int counter;* *private String l