Thanks for the suggestion. I will use BIGINT now.
On Fri, Jul 20, 2012 at 3:48 PM, kulkarni.swar...@gmail.com <
kulkarni.swar...@gmail.com> wrote:
> BIGINT is 8 bytes whereas INT is 4 bytes. Timestamps are usually of "long"
> type. To avoid loss of precision, I would recommend BIGINT.
>
>
> On
In hive we cannot calculate the difference between dates in minutes?
On Fri, Jul 20, 2012 at 6:06 PM, Techy Teck wrote:
> Whenever I am running the below query
>
> SELECT TT.BUYER_ID , COUNT(*) FROM
> (SELECT testingtable1.buyer_id, testingtable1.item_id,
> testingtable1.created_time from (se
Can you try it with,
*UNIX_TIMESTAMP(table1.**created_time) = **prod.timestamps + (15 * **60*
1000)*
2012/7/21 Techy Teck
> My full query is like this- The thing that is in red color is giving me
> exception. And I already have where clause after the JOIN. Not sure how to
> make this thing work
My full query is like this- The thing that is in red color is giving me
exception. And I already have where clause after the JOIN. Not sure how to
make this thing work.
*SELECT TT.BUYER_ID , COUNT(*) FROM*
*(SELECT table1.buyer_id, table1.item_id, table1.created_time from prod
RIGHT OUTER JOIN tab
Hi
AFAIK hive supports only equi joins. Can you check whether changing that <= to
= , removes the semantic error?
You can give conditions other than equality in a where clause post join.
Regards
Bejoy KS
Sent from handheld, please excuse typos.
-Original Message-
From: Techy Teck
Dat
BIGINT is 8 bytes whereas INT is 4 bytes. Timestamps are usually of "long"
type. To avoid loss of precision, I would recommend BIGINT.
On Fri, Jul 20, 2012 at 4:52 PM, Tech RJ wrote:
> What is the difference between these two? Trying to convert timestamps to
> full date format. The only differen
Yup, Thanks it worked.
*Raihan Jamal*
On Fri, Jul 20, 2012 at 1:40 PM, Bejoy KS wrote:
> **
> Raihan
>
> To see the failed task logs in hadoop, the easiest approach is
> drilling down the jobtracker web UI.
>
> Go to the job url (which you'll get in the beginning of a job on your
> console,
I have a join query on which I am joining like this below. And it is
working fine
ON (table1.item = prod.product *AND* table1.buyer = prod.user *AND*
UNIX_TIMESTAMP(table1.tt_time)
= prod.timestamps)
But when I tried to modify the above same thing.
The thing that is in red color will work or not
Thanks for the advice Edward. That makes sense to me.
As a side note, while doing some searching, I stumbled upon your blog[1]
regarding the release which made me even more curious. :)
[1]
http://www.edwardcapriolo.com/roller/edwardcapriolo/entry/when_hive_0_10_release
On Fri, Jul 20, 2012 at 4:
Usually it is ok to build a trunk, we do not do anything extra special
for releases other then some basic sanity testing and cut an svn tag.
(we run the full unit tests every commit). The only time this advice
is not true is if there are some metastore changes, however the
scripts to handle the upg
Hello,
I totally understand that usually open source projects to not have a fixed
date for release but I was just curious if something was chalked out for
releasing hive 0.10 out in the wild. There are some really interesting
additions that I am looking forward to.
Thanks,
--
Swarnim
Raihan
To see the failed task logs in hadoop, the easiest approach is
drilling down the jobtracker web UI.
Go to the job url (which you'll get in the beginning of a job on your console,
by the name Tracking url)
http://ares-jt.vip.ebay.com:50030/jobdetails.jsp?jobid=job_201207172005_14407
Brow
I tried opening the below URL, and nothing got opened, I got page cannot be
displayed. Why is that so?
*Raihan Jamal*
On Fri, Jul 20, 2012 at 12:39 PM, Sriram Krishnan wrote:
> What version of Hadoop and Hive are you using? We have seen errors like
> this in the past – and you can actually
After setting this in Hive-
hive> SET hive.exec.show.job.failure.debug.info=false;
I can see the logs on my console itself? Or I need to go somewhere to see
the actual logs and what is causing the problem?
*Raihan Jamal*
On Fri, Jul 20, 2012 at 12:28 PM, kulkarni.swar...@gmail.com <
kulkarni
What version of Hadoop and Hive are you using? We have seen errors like this in
the past – and you can actually replace taskid with attemptid to fetch your
logs.
So try this:
http://lvsaishdc3dn0857.lvs.ebay.com:50060/tasklog?attemptid=attempt_201207172005_14407_r_00_1&all=true
But yes, th
Whenever I run the below query-
*
*
*SELECT buyer_id, item_id, ranknew(buyer_id, item_id), created_time*
*FROM (*
*SELECT buyer_id, item_id, created_time*
*FROM testingtable1*
*DISTRIBUTE BY buyer_id, item_id*
*SORT BY buyer_id, item_id, created_time desc*
*) a*
*WHERE ranknew(buyer
*rpool/tmp 10G10G 0K 100%/tmp*
*
*
This might be the source of your problem as I mentioned earlier. Try
freeing some space here and then try again.
On Fri, Jul 20, 2012 at 11:34 AM, comptech geeky wrote:
> After trying "df -kh". I got below result.
>
> *bash-3.00$ df -
After trying "df -kh". I got below result.
*bash-3.00$ df -kh*
*Filesystem size used avail capacity Mounted on*
*rpool/ROOT/sol10 916G30G 668G 5%/*
*/devices 0K 0K 0K 0%/devices*
*ctfs 0K 0K 0K 0%
Can anyone help me on this particular problem? I am still stuck on this
from so long. I have modified my Rank function again for my scenario. But
not sure whether this will work or not. Any suggestions?
*public final class RankAnother extends UDF{*
*private int counter;*
*private String l
19 matches
Mail list logo