Hi,
I downloaded this book someday back and started reading it. This book is giving
me the programing style using hive. But when I tried to visualize
how the hive will connect with hadoop cluster,
how the hive will get the request,
how the hive will process the request,
after analysis ,where
Hi,
This is my xml file
Company
Employee
id100/id
enameranjini/ename
deptIT1/dept
sal123456/sal
locationnextlevel1/location
Address
HomeChennai1/Home
OfficeNavallur1/Office
/Address
/Employee
Employee
id1001/id
enameranjinikumar/ename
deptIT/dept
sal1234516/sal
Welcome back, Sanjay.
-- Lefty
On Sun, Jan 12, 2014 at 10:33 PM, Subramanian, Sanjay (HQP)
sanjay.subraman...@roberthalf.com wrote:
Greetings Hive-ians
Great to be writing to this group once more.
After using Cloudera CDH distributions in production for almost 2 years
, I am using
If I modify 4to5 columns by replace columns the rest of the columns will be
deleted. I just want to change some column names and types, rest of the
columns as it is.
On Sun, Jan 12, 2014 at 6:45 PM, Swagatika Tripathy swagatikat...@gmail.com
wrote:
Yes; u can do so... 5 individual alter table
Thanks Prashant, Definitely i shall go through that if needed. But from my
experience, what i have faced is that user will have some integration
problem with HADOOP 2.
Hi Vikas
Welcome to the world of Hive !
The first book u should read is by Capriolo , Wampler, Rutherglen
Programming
Hi there,
To analyze the data you need to load the data file into hive first.this
will be done by creating a table(CREATE TABLE) and loading the file
data(LOAD DATA) into the table according to your requirement.
Actually the data is still stored in HDFS itself, you can go to
user/hive/warehouse
The best way to answer your queries is,
1) set up a single node hadoop VM (there are readily available images from
hortonworks and cloudera)
2) try to load data and see where it is stored (hive is a data access
framework .. it does not store any data, information related to data is
stored in
I don't work for IBM, but found their training material helpful:
http://bigdatauniversity.com
There is a bit of biased toward IBM's stack, but they do a good job of
teaching Hive in general.
On Mon, Jan 13, 2014 at 3:01 AM, Nitin Pawar nitinpawar...@gmail.comwrote:
The best way to answer your
Thanks Lefty
Thanks
Warm Regards
Sanjay
From: Lefty Leverenz leftylever...@gmail.commailto:leftylever...@gmail.com
Reply-To: user@hive.apache.orgmailto:user@hive.apache.org
Date: Mon, 13 Jan 2014 01:50:24 -0800
To: user@hive.apache.orgmailto:user@hive.apache.org
Subject: Re: Hive on Amazon
Hi,
I have a table that is of the following format
create table t1 ( f1 int, f2 arraystructa1:int, a2:int );
Now I have a custom script that does some computation and generates
the value for f2
like so
from (
from randomtable r
map r.g1, r.g2, r.g3
using '/bin/cat' as g1, g2, g3
In my try, it worked (and should be).
CREATE EXTERNAL TABLE MasterTable (
column1 STRING, column2 STRING)
LOCATION 'hdfs://localhost:9000/home/navis/my_location';
CREATE VIEW IF NOT EXISTS View1 (column1, column2) AS SELECT column1,
column2 FROM MasterTable WHERE column130;
CREATE VIEW IF
Hello!
When i run hive job, it creates large temporary files on
/tmp/hive-ben/hive_2014-01-14_15-01-39_521_3015861149916225685-1/
somewhere around 300GB. this number tends to get larger if i use lower
number of reducers.
with 5 reducers the size goes up to 1TB
my input files are total 1GB,
12 matches
Mail list logo