Hive metatool JVM Heap error

2013-11-11 Thread Mark Schramm
Hive Metatool experts, Trying to use the hive "metatool" to update the locations for all of our tables after a configuration change did not do this automatically. The tool gets a heap error even though we upped the JVM to 4 GB. Does anyone know of a workaround? Thanks, Mark

Re: How to load a web log file (text format) to Hive with compression

2013-11-11 Thread Vijay
Hive has built-in support for gzipped text files so you don't need to do anything special with them (meaning you just load them like regular files). However, there is one important limitation with this approach these files are not splittable. Depending on the size of your log files and how many of

Re: Developing a GenericUDAF

2013-11-11 Thread Ross Levin
Thanks Navis, that got me past this exception! Ross On Mon, Nov 11, 2013 at 6:03 PM, Navis류승우 wrote: > in handling PARTIAL1, > > inputOI = (StandardListObjectInspector) parameters[0]; > return ObjectInspectorFactory.getStandardListObjectInspector(inputOI); > > 1. > inputOI is not guaranteed to

Hive run ERROR org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException as:jyfx (auth:SIMPLE) cause:java.io.IOException: cannot find class

2013-11-11 Thread asword
Hi all, I am using Hadoop cdh3u6 and apache hive 0.10. I run the summary sql as follow : SELECT pt, count(1), sum(flow_byte), count(distinct t.user_id), count(distinct t.phone_no), max(t.user_ip), min(t.user_ip), max(t.phone_no),

Re: [ANNOUNCE] New Hive Committer - Prasad Mujumdar

2013-11-11 Thread Gunther Hagleitner
Awesome! Congrats, Prasad! On Mon, Nov 11, 2013 at 4:00 PM, Zhang Xiaoyu wrote: > Congrats, Prasad! > > Johnny > > > On Mon, Nov 11, 2013 at 10:01 AM, Prasanth Jayachandran < > pjayachand...@hortonworks.com> wrote: > >> Congrats Prasad! >> >> Thanks >> Prasanth Jayachandran >> >> On Nov 10, 2013

How to load a web log file (text format) to Hive with compression

2013-11-11 Thread Raj Hadoop
Hi, I have a web log files (text format). I want to load these files to a Hive table in compressed format. How do I do it ? Should I compress the text file (using any Linux utilities) and then create the Hive table? Can any one provide me the Hive syntax for loading the compressed file? Thank

Re: Developing a GenericUDAF

2013-11-11 Thread Navis류승우
in handling PARTIAL1, inputOI = (StandardListObjectInspector) parameters[0]; return ObjectInspectorFactory.getStandardListObjectInspector(inputOI); 1. inputOI is not guaranteed to be a StandardListObjectInspector. Use ListObjectInspector instead. 2. ObjectInspectorFactory.getStandardListObjectIn

Re: Seeking Help configuring log4j for sqoop import into hive

2013-11-11 Thread David Morel
On 12 Nov 2013, at 0:01, Sunita Arvind wrote: Just in case this acts as a workaround for someone: The issue is resolved if I eliminate the "where" clause in the query (just keep "where $CONDITIONS"). So 2 workarounds I can think of now are: 1. Create views in Oracle and query without the where

Re: Seeking Help configuring log4j for sqoop import into hive

2013-11-11 Thread Jarek Jarcec Cecho
Hi Sunita, Sqoop specific questions are better asked on Sqoop user mailing list u...@sqoop.apache.org. You can find instructions how to subscribe to that at [1]. I would suggest to take a look into the failed map task log as that log usually contain entire exception including all the chained ex

Re: [ANNOUNCE] New Hive Committer - Prasad Mujumdar

2013-11-11 Thread Zhang Xiaoyu
Congrats, Prasad! Johnny On Mon, Nov 11, 2013 at 10:01 AM, Prasanth Jayachandran < pjayachand...@hortonworks.com> wrote: > Congrats Prasad! > > Thanks > Prasanth Jayachandran > > On Nov 10, 2013, at 10:16 PM, Vaibhav Gumashta > wrote: > > > Congrats Prasad! > > > > > > On Sun, Nov 10, 2013 at

Re: Seeking Help configuring log4j for sqoop import into hive

2013-11-11 Thread Sunita Arvind
Just in case this acts as a workaround for someone: The issue is resolved if I eliminate the "where" clause in the query (just keep "where $CONDITIONS"). So 2 workarounds I can think of now are: 1. Create views in Oracle and query without the where clause in the sqoop import command 2. Import every

Re: What is pfile in pom.xml

2013-11-11 Thread Abdelrahman Shettia
Hi Jie, Seems that there is an error related to: No FileSystem for scheme: pfile Have you tried to replace the line? pfile:// With: /test So that it would work in this format: CREATE TABLE rename_partition_table (key STRING, value STRING) PARTITIONED BY (part STRING) STORED AS RCFILE L

Seeking Help configuring log4j for sqoop import into hive

2013-11-11 Thread Sunita Arvind
Hello, I am using sqoop to import data from oracle into hive. Below is my SQL: nohup sqoop import --connect "jdbc:oracle:thin:@(DESCRIPTION = (ADDRESS = (PROTOCOL = TCP)(HOST = xxx)(PORT = )) (CONNECT_DATA = (SERVER = DEDICATED) (SERVICE_NAME = CDWQ.tms.toyota.com) (FAILOVER_MODE= (TYPE=s

Developing a GenericUDAF

2013-11-11 Thread Ross Levin
Hello, I'm writing a generic UDAF function that closely resembles SUM() with the main difference being that it accepts an array datatype parameter and returns an array datatype. I've already done this for a GenericUDF successfully. I believe I am having difficulty coding the proper ObjectInspecto

Re: [ANNOUNCE] New Hive Committer - Prasad Mujumdar

2013-11-11 Thread Prasanth Jayachandran
Congrats Prasad! Thanks Prasanth Jayachandran On Nov 10, 2013, at 10:16 PM, Vaibhav Gumashta wrote: > Congrats Prasad! > > > On Sun, Nov 10, 2013 at 8:17 PM, Lefty Leverenz > wrote: > >> Congratulations Prasad! >> >> -- Lefty >> >> >> On Sun, Nov 10, 2013 at 11:04 PM, Brock Noland wrot

Re: RLE in hive ORC

2013-11-11 Thread Prasanth Jayachandran
As Owen noted, max run for version 0.11 is 130. 3 is minimum run for RLE to be used. So max value that can be interpreted from 7 bits is 130. Thanks Prasanth Jayachandran On Nov 11, 2013, at 9:51 AM, Owen O'Malley wrote: > Hi, > The RLE in ORC is a tradeoff (as is all compression) between t

Re: RLE in hive ORC

2013-11-11 Thread Owen O'Malley
Hi, The RLE in ORC is a tradeoff (as is all compression) between tight representations for commonly occurring patterns and longer representations for rarely occurring patterns. The question at hand is how to use the bits available to reduce the average size of the column. In Hive 0.12, ORC gained

Re: RLE in hive ORC

2013-11-11 Thread Prasanth Jayachandran
Runs of 1M is not common case. I am not sure how vertica stores the run lengths. It seems like variable length integers are used. ORC does not use variable length integers for storing run length. Using variable length integer has advantage of storing much longer runs but for repeating shorter r

RLE in hive ORC

2013-11-11 Thread qihua wu
In vertica, if I have a column sorted, and the same value repeat 1M times, it only used very small storage as it only stores (value, 1M). But in ORC, looks like the max length is less than 200 ( not very sure, but at about the same level of hundreds), why restrict the max run length?