Hi,
I downloaded this book someday back and started reading it. This book is giving
me the programing style using hive. But when I tried to visualize
how the hive will connect with hadoop cluster,
how the hive will get the request,
how the hive will process the request,
after analysis ,where the
>
> Hi,
>
> This is my xml file
>
>
>
> 100
> ranjini
> IT1
> 123456
> nextlevel1
>
> Chennai1
> Navallur1
>
>
>
> 1001
> ranjinikumar
> IT
> 1234516
> nextlevel
>
> Chennai
> Navallur
>
>
>
>
>
> In hive shell , i use xpath
>
> hive> select xpath(str,'/Company/Employee/ename/text()')
Welcome back, Sanjay.
-- Lefty
On Sun, Jan 12, 2014 at 10:33 PM, Subramanian, Sanjay (HQP) <
sanjay.subraman...@roberthalf.com> wrote:
> Greetings Hive-ians
>
>
> Great to be writing to this group once more.
>
>
> After using Cloudera CDH distributions in production for almost 2 years
> , I
If I modify 4to5 columns by replace columns the rest of the columns will be
deleted. I just want to change some column names and types, rest of the
columns as it is.
On Sun, Jan 12, 2014 at 6:45 PM, Swagatika Tripathy wrote:
> Yes; u can do so... 5 individual alter table replace column command
Thanks Prashant, Definitely i shall go through that if needed. But from my
experience, what i have faced is that user will have some integration
problem with HADOOP 2.
Hi Vikas
>
> Welcome to the world of Hive !
>
> The first book u should read is by Capriolo , Wampler, Rutherglen
> Programmin
Hi there,
To analyze the data you need to load the data file into hive first.this
will be done by creating a table(CREATE TABLE) and loading the file
data(LOAD DATA) into the table according to your requirement.
Actually the data is still stored in HDFS itself, you can go to
user/hive/warehouse d
The best way to answer your queries is,
1) set up a single node hadoop VM (there are readily available images from
hortonworks and cloudera)
2) try to load data and see where it is stored (hive is a data access
framework .. it does not store any data, information related to data is
stored in metas
I don't work for IBM, but found their training material helpful:
http://bigdatauniversity.com
There is a bit of biased toward IBM's stack, but they do a good job of
teaching Hive in general.
On Mon, Jan 13, 2014 at 3:01 AM, Nitin Pawar wrote:
> The best way to answer your queries is,
>
> 1) set
Thanks Lefty
Thanks
Warm Regards
Sanjay
From: Lefty Leverenz mailto:leftylever...@gmail.com>>
Reply-To: mailto:user@hive.apache.org>>
Date: Mon, 13 Jan 2014 01:50:24 -0800
To: mailto:user@hive.apache.org>>
Subject: Re: Hive on Amazon
Welcome back, Sanjay.
-- Lefty
On Sun, Jan 12, 2014 at 10
Hi,
I have a table that is of the following format
create table t1 ( f1 int, f2 array> );
Now I have a custom script that does some computation and generates
the value for f2
like so
from (
from randomtable r
map r.g1, r.g2, r.g3
using '/bin/cat' as g1, g2, g3
cluster by g1 ) m
In my try, it worked (and should be).
CREATE EXTERNAL TABLE MasterTable (
column1 STRING, column2 STRING)
LOCATION 'hdfs://localhost:9000/home/navis/my_location';
CREATE VIEW IF NOT EXISTS View1 (column1, column2) AS SELECT column1,
column2 FROM MasterTable WHERE column1<30;
CREATE VIEW IF N
Hello!
When i run hive job, it creates large temporary files on
/tmp/hive-ben/hive_2014-01-14_15-01-39_521_3015861149916225685-1/
somewhere around 300GB. this number tends to get larger if i use lower
number of reducers.
with 5 reducers the size goes up to 1TB
my input files are total 1GB, but
12 matches
Mail list logo