try rk in upper select statement as well
On Tue, Jul 10, 2012 at 12:12 PM, Raihan Jamal wrote:
> Thanks for commenting out. Yes I figured that out, its a UDF. So now I
> have created a new UDF Rank and added to classpath also. But when I am
> again running the below query-
>
> SELECT buyer_id, i
Thanks for commenting out. Yes I figured that out, its a UDF. So now I have
created a new UDF Rank and added to classpath also. But when I am again
running the below query-
SELECT buyer_id, item_id, created_time
FROM (
SELECT buyer_id, item_id, Rank(buyer_id) as rk, created_time
FROM testi
hive does not support update as a query
its a costly operation to update records in hive
as a hadoop based system mainly its meant like write once read many times
On Tue, Jul 10, 2012 at 11:36 AM, VanHuy Pham wrote:
> Hi All,
>Is it possible to insert and/or update some data in a hive table
Hi All,
Is it possible to insert and/or update some data in a hive table not
through any local file?
To be specific, I look to insert and update some data (not large amount)
to a hive table through queries (like insert into...) in a traditional SQL
style, where you specify the new data in the
hive has no built-in rank function. you'd need to use a user-defined
function (UDF) to simulate it. there are a few custom implementations
on the net that you can leverage.
On Mon, Jul 9, 2012 at 10:40 PM, Raihan Jamal wrote:
> What's wrong with the below query.
>
>
> SELECT buyer_id, item_id, cr
This is my below requirement. I need- *Find `TOP 10` data for each
`BUYER_ID, *So I cannot use LIMIT 10 here in this case.
This is the below table
CREATE TABLE IF NOT EXISTS TestingTable1
(
BUYER_ID BIGINT,
ITEM_ID BIGINT,
CREATED_TIME STRING
)
And this is the below data
hello, you can use "desc" and "limit 10" to filter the top 10.
2012-07-10
Best Regards
Andes
发件人:Raihan Jamal
发送时间:2012-07-10 10:31
主题:Find TOP 10 using HiveQL
收件人:"user"
抄送:
When I run this query,
SELECT TOP 10 FROM TestingTable1 WHERE ORDER BY buyer_id, created_time DESC;
I always
yes you need an input format. I have one that turns key plus value to a
pairwritable
On Monday, July 9, 2012, Daning Wang wrote:
> Thanks Navis.
>
> The problem is that the key and value are custom classes. e.g, key is
class A, and value is class B. we want to show certain data in Class A and
B f
Thanks Navis.
The problem is that the key and value are custom classes. e.g, key is class
A, and value is class B. we want to show certain data in Class A and B from
getters. so we do need a SerDe to deserialize the data.
Since SerDe interface can only pass one parameter, so I need to write a
Inp
Yup that worked for me. I figure that out after reading the docs, INNER
JOIN means JOIN in HiveQL.
*Raihan Jamal*
On Mon, Jul 9, 2012 at 2:48 PM, Roberto Sanabria wrote:
> Did you try just using "join" instead of "inner join"?
>
>
> On Mon, Jul 9, 2012 at 11:42 AM, Raihan Jamal wrote:
>
>>
You can try 'STORED AS SEQUENCEFILE' when creating table.
https://cwiki.apache.org/Hive/languagemanual-ddl.html#LanguageManualDDL-Create%252FDropTable
2012/7/10 Daning Wang
> Hi all,
>
> New to here. I have sequence files and I want to read in Hive, I did some
> search on the net. looks the w
Hi all,
New to here. I have sequence files and I want to read in Hive, I did some
search on the net. looks the way I need to do
1. Write new FileInputFormat so Hive can read both key and value.
(ref http://osdir.com/ml/hive-user-hadoop-apache/2009-10/msg00034.html)
2. Write a SerDe to deseriali
Did you try just using "join" instead of "inner join"?
On Mon, Jul 9, 2012 at 11:42 AM, Raihan Jamal wrote:
> I wrote this query but everytime I get the below error.
>
> select * from (select t2.buyer_id, t2.item_id, t2.created_time as
> created_time, subq.user_id, subq.product_id, subq.time
I wrote this query but everytime I get the below error.
select * from (select t2.buyer_id, t2.item_id, t2.created_time as
created_time, subq.user_id, subq.product_id, subq.timestamps as timestamps
from
(select user_id, prod_and_ts.product_id as product_id,
prod_and_ts.timestamps as tim
No, there is no way to re-attach. Because hive works with
side-directories and moves data at the last step the job is not
salvageable. Your should kill the job from the hadoop job tracker.
On Mon, Jul 9, 2012 at 11:30 AM, Martin Meyer wrote:
> Hi all -
>
> I've just had a hive CLI process die wil
Hi all -
I've just had a hive CLI process die wile a job was running. Even
though the CLI has terminated, the hadoop job is still being
processed. Is there any way to either:
a) reattach the hive command to an already-running job, or
b) read the query results out of HDFS when the job actually fi
16 matches
Mail list logo