RE: One information about the Hive

2014-01-13 Thread Prashant Kumar - ERS, HCL Tech
Hi, I downloaded this book someday back and started reading it. This book is giving me the programing style using hive. But when I tried to visualize how the hive will connect with hadoop cluster, how the hive will get the request, how the hive will process the request, after analysis ,where

Re: passing Xml using hive

2014-01-13 Thread Ranjini Rathinam
Hi, This is my xml file Company Employee id100/id enameranjini/ename deptIT1/dept sal123456/sal locationnextlevel1/location Address HomeChennai1/Home OfficeNavallur1/Office /Address /Employee Employee id1001/id enameranjinikumar/ename deptIT/dept sal1234516/sal

Re: Hive on Amazon

2014-01-13 Thread Lefty Leverenz
Welcome back, Sanjay. -- Lefty On Sun, Jan 12, 2014 at 10:33 PM, Subramanian, Sanjay (HQP) sanjay.subraman...@roberthalf.com wrote: Greetings Hive-ians Great to be writing to this group once more. After using Cloudera CDH distributions in production for almost 2 years , I am using

Re: Deleting a column from internally managed table

2014-01-13 Thread Kishore kumar
If I modify 4to5 columns by replace columns the rest of the columns will be deleted. I just want to change some column names and types, rest of the columns as it is. On Sun, Jan 12, 2014 at 6:45 PM, Swagatika Tripathy swagatikat...@gmail.com wrote: Yes; u can do so... 5 individual alter table

Re: One information about the Hive

2014-01-13 Thread Vikas Parashar
Thanks Prashant, Definitely i shall go through that if needed. But from my experience, what i have faced is that user will have some integration problem with HADOOP 2. Hi Vikas Welcome to the world of Hive ! The first book u should read is by Capriolo , Wampler, Rutherglen Programming

Re: One information about the Hive

2014-01-13 Thread prakash gujjar
Hi there, To analyze the data you need to load the data file into hive first.this will be done by creating a table(CREATE TABLE) and loading the file data(LOAD DATA) into the table according to your requirement. Actually the data is still stored in HDFS itself, you can go to user/hive/warehouse

Re: One information about the Hive

2014-01-13 Thread Nitin Pawar
The best way to answer your queries is, 1) set up a single node hadoop VM (there are readily available images from hortonworks and cloudera) 2) try to load data and see where it is stored (hive is a data access framework .. it does not store any data, information related to data is stored in

Re: One information about the Hive

2014-01-13 Thread Peyman Mohajerian
I don't work for IBM, but found their training material helpful: http://bigdatauniversity.com There is a bit of biased toward IBM's stack, but they do a good job of teaching Hive in general. On Mon, Jan 13, 2014 at 3:01 AM, Nitin Pawar nitinpawar...@gmail.comwrote: The best way to answer your

Re: Hive on Amazon

2014-01-13 Thread Subramanian, Sanjay (HQP)
Thanks Lefty Thanks Warm Regards Sanjay From: Lefty Leverenz leftylever...@gmail.commailto:leftylever...@gmail.com Reply-To: user@hive.apache.orgmailto:user@hive.apache.org Date: Mon, 13 Jan 2014 01:50:24 -0800 To: user@hive.apache.orgmailto:user@hive.apache.org Subject: Re: Hive on Amazon

casting complex data types for outputs of custom scripts

2014-01-13 Thread rohan monga
Hi, I have a table that is of the following format create table t1 ( f1 int, f2 arraystructa1:int, a2:int ); Now I have a custom script that does some computation and generates the value for f2 like so from ( from randomtable r map r.g1, r.g2, r.g3 using '/bin/cat' as g1, g2, g3

Re: Pointing multiple external tables to the same location

2014-01-13 Thread Navis류승우
In my try, it worked (and should be). CREATE EXTERNAL TABLE MasterTable ( column1 STRING, column2 STRING) LOCATION 'hdfs://localhost:9000/home/navis/my_location'; CREATE VIEW IF NOT EXISTS View1 (column1, column2) AS SELECT column1, column2 FROM MasterTable WHERE column130; CREATE VIEW IF

HIve temp file

2014-01-13 Thread Ben Kim
Hello! When i run hive job, it creates large temporary files on /tmp/hive-ben/hive_2014-01-14_15-01-39_521_3015861149916225685-1/ somewhere around 300GB. this number tends to get larger if i use lower number of reducers. with 5 reducers the size goes up to 1TB my input files are total 1GB,