Re: Reducers Failing

2014-05-21 Thread Shengjun Xin
Are datanodes dead?


On Thu, May 22, 2014 at 2:23 PM, Sreenath  wrote:

> Hi All,
>
> We are running a hadoop cluster and many of our hive queries are failing
> in the reduce phase with the following error
>
> java.io.IOException: All datanodes *.*.*.*:50230 are bad. Aborting...
>   at 
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.processDatanodeError(DFSClient.java:3290)
>   at 
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2200(DFSClient.java:2783)
>   at 
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2987)
>
> Any IDead why this might be happening ?
> --
> Sreenath S Kamath
> Bangalore
> Ph No:+91-9590989106
>



-- 
Regards
Shengjun


Re: Connecting hive to SAP BO

2014-05-19 Thread Shengjun Xin
hive –auxpath /path-to-/csvserde.jar


On Tue, May 20, 2014 at 12:59 PM, Chhaya Vishwakarma <
chhaya.vishwaka...@lntinfotech.com> wrote:

>
>
> Hi,
>
> I have connected SAP BO to Hive using a ODBC driver. I am able to see the
> database and table in hive. but when i fetch data from hive it gives error
> as
>
> org.apache.hadoop.hìve.serde2.SerDeExceptio SerDe
> com,bizohive.serde.csv.CSVSerde does not exist
>
> Can ayone suggest where i should put csvserde jar in SAP BO
>
>
>
>
>
> Regards,
>
> Chhaya Vishwakarma
>
>
>
> --
> The contents of this e-mail and any attachment(s) may contain confidential
> or privileged information for the intended recipient(s). Unintended
> recipients are prohibited from taking action on the basis of information in
> this e-mail and using or disseminating the information, and must notify the
> sender and delete it from their system. L&T Infotech will not accept
> responsibility or liability for the accuracy or completeness of, or the
> presence of any virus or disabling code in this e-mail"
>



-- 
Regards
Shengjun


Re: Hive 0.13.0 Memory Leak

2014-05-18 Thread Shengjun Xin
We found a hive jdbc connection leak
https://issues.apache.org/jira/browse/HIVE-6866, is it similar to your
issue?


On Thu, May 15, 2014 at 9:29 PM, Bryan Jeffrey wrote:

> All,
>
> We are running Hadoop 2.2.0 and Hive 0.13.0.  One typical application is
> to load data (as text), and then convert that data to ORC to decrease query
> time.  When running these processes we are seeing significant memory leaks
> (leaking 4 GB in about 5 days).
>
> We're running HiveServer2 with the following options:
>
> root 28136 1 51 May14 ?09:51:09 /usr/java/latest/bin/java
> -Xmx2048m -XX:NewRatio=12 -Xms10m -XX:MaxHeapFreeRatio=40
> -XX:MinHeapFreeRatio=15 -XX:-UseGCOverheadLimit -XX:MaxPermSize=1024m
> -XX:NewRatio=12 -Xms10m -XX:MaxHeapFreeRatio=40 -XX:MinHeapFreeRatio=15
> -XX:-UseGCOverheadLimit -XX:MaxPermSize=1024m
> -Dhadoop.log.dir=/opt/hadoop/latest-hadoop/logs
> -Dhadoop.log.file=hadoop.log -Dhadoop.home.dir=/opt/hadoop/latest-hadoop
> -Dhadoop.id.str= -Dhadoop.root.logger=INFO,console
> -Djava.library.path=/opt/hadoop/latest-hadoop/lib/native
> -Dhadoop.policy.file=hadoop-policy.xml -Djava.net.preferIPv4Stack=true
> -Dhadoop.security.logger=INFO,NullAppender org.apache.hadoop.util.RunJar
> /opt/hadoop/latest-hive/lib/hive-service-0.13.0.jar
> org.apache.hive.service.server.HiveServer2
>
> Typical ORC conversion query looks like the following:
>
> HIVE -u jdbc:hive2://hive_server:10002/db -n root --hiveconf
> hive.exec.dynamic.partition.mode=nonstrict --hiveconf
> hive.enforce.sorting=true --hiveconf $SET_QUEUE -e "insert into table
> orc_table partition (range) select * from loading_text_table; "
>
> I saw a couple of tickets for memory leaks, but they seemed to deal with
> failed queries.  The memory usage increases in a linear fashion.  Jobs all
> succeed until memory limit exceeded.
>
> Is there an open bug for memory leaks associated with successful jobs in
> HS2?  Is there a fix for this issue?
>
> Regards,
>
> Bryan Jeffrey
>



-- 
Regards
Shengjun


Re: issue about Job running in-process (local Hadoop)

2014-05-12 Thread Shengjun Xin
According to the log, you configured a local hadoop, you need to check you
configuration


On Mon, May 12, 2014 at 3:46 PM, ch huang  wrote:

> hi,maillist:
>i try hive 0.12,but it always run job in local mode ,why? i install
> hive on a seperate hox ,and install hadoop client on it ,and configure the
> client connect to my hadoop cluster
>
> hive> select count(*) from media_visit_info;
> Total MapReduce jobs = 1
> Launching Job 1 out of 1
> Number of reduce tasks determined at compile time: 1
> In order to change the average load for a reducer (in bytes):
>   set hive.exec.reducers.bytes.per.reducer=
> In order to limit the maximum number of reducers:
>   set hive.exec.reducers.max=
> In order to set a constant number of reducers:
>   set mapred.reduce.tasks=
> 14/05/12 15:36:52 WARN conf.Configuration:
> file:/tmp/root/hive_2014-05-12_15-36-48_130_3729255905600309633-1/-local-10002/jobconf.xml:an
> attempt to override final parameter:
> mapreduce.job.end-notification.max.retry.interval;  Ignoring.
> 14/05/12 15:36:52 WARN conf.Configuration:
> file:/tmp/root/hive_2014-05-12_15-36-48_130_3729255905600309633-1/-local-10002/jobconf.xml:an
> attempt to override final parameter:
> mapreduce.job.end-notification.max.attempts;  Ignoring.
> 14/05/12 15:36:52 INFO Configuration.deprecation:
> mapred.input.dir.recursive is deprecated. Instead, use
> mapreduce.input.fileinputformat.input.dir.recursive
> 14/05/12 15:36:52 INFO Configuration.deprecation: mapred.max.split.size is
> deprecated. Instead, use mapreduce.input.fileinputformat.split.maxsize
> 14/05/12 15:36:52 INFO Configuration.deprecation: mapred.min.split.size is
> deprecated. Instead, use mapreduce.input.fileinputformat.split.minsize
> 14/05/12 15:36:52 INFO Configuration.deprecation:
> mapred.min.split.size.per.rack is deprecated. Instead, use
> mapreduce.input.fileinputformat.split.minsize.per.rack
> 14/05/12 15:36:52 INFO Configuration.deprecation:
> mapred.min.split.size.per.node is deprecated. Instead, use
> mapreduce.input.fileinputformat.split.minsize.per.node
> 14/05/12 15:36:52 INFO Configuration.deprecation: mapred.reduce.tasks is
> deprecated. Instead, use mapreduce.job.reduces
> 14/05/12 15:36:52 INFO Configuration.deprecation:
> mapred.reduce.tasks.speculative.execution is deprecated. Instead, use
> mapreduce.reduce.speculative
> Execution log at:
> /tmp/root/root_20140512153636_d3088311-cdbe-44c4-8622-e47dce685f34.log
> Job running in-process (local Hadoop)
> 2014-05-12 15:36:56,257 null map = 0%,  reduce = 0%
> 2014-05-12 15:36:59,275 null map = 100%,  reduce = 0%
> 2014-05-12 15:37:01,314 null map = 100%,  reduce = 100%
> Ended Job = job_local1940121222_0001
> Execution completed successfully
> MapredLocal task succeeded
> OK
> 3093138
> Time taken: 14.136 seconds, Fetched: 1 row(s)
>



-- 
Regards
Shengjun


Re: java.lang.IncompatibleClassChangeError in hive

2014-05-05 Thread Shengjun Xin
Check the hive-env.sh and environment variables, whether the HADOOP_HOME is
what you want?


On Tue, May 6, 2014 at 11:33 AM, chandra kant <
chandralakshmikan...@gmail.com> wrote:

>
> Hi,
> When running bin/hive , i get the following error :-
> Exception in thread "main" java.lang.IncompatibleClassChangeError: Found
> interface jline.Terminal, but class was expected
>  at jline.ConsoleReader.(ConsoleReader.java:191)
> at jline.ConsoleReader.(ConsoleReader.java:186)
>  at jline.ConsoleReader.(ConsoleReader.java:174)
> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:719)
>  at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>  at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>  at java.lang.reflect.Method.invoke(Method.java:597)
> at org.apache.hadoop.util.RunJar.main(RunJar.java:160)
>
> I think this comes when there is version mismatching between hadoop
> version and hive version. In my case it is - hadoop-1.2.1 and hive-0.11.0 ,
> which were working perfectly fine till yesterday. And yes, I unsuccessfully
> tried to run hadoop-2 on my system and later reverted back to hadoop-1. And
> , it occurred  ...
>
> Any help in debugging..?
>
> --
> Chandra
>
>


-- 
Regards
Shengjun


Re: Number of hive-server2 threads increments after jdbc connection

2014-05-04 Thread Shengjun Xin
Is it same with https://issues.apache.org/jira/browse/HIVE-6866 ?


On Tue, Apr 29, 2014 at 8:26 PM, Dima Fadeyev  wrote:

>  Hi, Chinna. Thanks for your reply.
>
> Yes, modifying code solves the problem. This is what my code looks like (a
> piece of it):
>
> Connection con =
> DriverManager.getConnection("jdbc:hive2://localhost:1/default", "hive",
> "hive");
> Statement stmt = con.createStatement();
> String tableName = "testHiveDriverTable";
> stmt.execute("drop table if exists " + tableName);
> //stmt.close();
>
> When I uncomment the last line, the number of hive-server2 threads doesn't
> keep incrementing to infinity. However I'm investigating the issue where
> the code is not really my code. Is there a way to correct this behavior
> from within hive-server2 without changing the client's code?
>
> El 29/04/14 14:05, Chinna Rao Lalam escribió:
>
> Hi,
>
>   In your code if more connections and statements are created?. If so,
> closed those connections?
>  After use close unused connections and statements.
>
>
>  Hope It Helps,
> Chinna
>
>
> On Tue, Apr 29, 2014 at 3:47 PM, Dima Fadeyev wrote:
>
>>  Hello everyone,
>>
>> When I run a jdbc example from
>> https://cwiki.apache.org/confluence/display/Hive/HiveServer2+Clients#HiveServer2Clients-JDBCClientSampleCodeagainst
>>  my hive server, the number of hive-server2 threads increments. If I
>> execute it long enough I either start seeing exceptions
>>
>> Exception in thread "main" java.sql.SQLException:
>> org.apache.thrift.TApplicationException: Internal error processing
>> ExecuteStatement
>> at org.apache.hive.jdbc.HiveStatement.execute(HiveStatement.java:203)
>> at HiveJdbcClient.main(HiveJdbcClient.java:24)
>> Caused by: org.apache.thrift.TApplicationException: Internal error
>> processing ExecuteStatement
>> at
>> org.apache.thrift.TApplicationException.read(TApplicationException.java:108)
>> at
>> org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:71)
>> at
>> org.apache.hive.service.cli.thrift.TCLIService$Client.recv_ExecuteStatement(TCLIService.java:213)
>> at
>> org.apache.hive.service.cli.thrift.TCLIService$Client.ExecuteStatement(TCLIService.java:200)
>> at org.apache.hive.jdbc.HiveStatement.execute(HiveStatement.java:197)
>> ... 1 more
>>
>> or I bump into a limit of connections with Zookeeper (each hive-server2
>> thread maintains a connection with Zookeeper. I have hive.support.concurrency
>> enabled).
>>
>> In either case I can't connect to hive server after that.
>>
>> I've tried this on hive 0.10 (cdh 4.4) and hive 0.12 (cdh 5.0 and hdp
>> 2.0.6) with same results.
>>
>> Please, could anyone help me resolve this.
>> Thanks in advance.
>>
>>
>>
>
>
>  --
> Hope It Helps,
> Chinna
>
>
>


-- 
Regards
Shengjun


Re: [ANNOUNCE] New Hive Committers - Prasanth J and Vaibhav Gumashta

2014-04-24 Thread Shengjun Xin
Congratulations ~~


On Fri, Apr 25, 2014 at 10:33 AM, Carl Steinbach wrote:

> + Prasanth's correct email address
>
>
> On Thu, Apr 24, 2014 at 7:31 PM, Xuefu Zhang  wrote:
>
>> Congratulations to Prasanth and Vaibhav!
>>
>> --Xuefu
>>
>>
>> On Thu, Apr 24, 2014 at 7:26 PM, Carl Steinbach  wrote:
>>
>> > The Apache Hive PMC has voted to make Prasanth J and Vaibhav
>> > Gumashta committers on the Apache Hive Project.
>> >
>> > Please join me in congratulating Prasanth and Vaibhav!
>> >
>> > Thanks.
>> >
>> > - Carl
>> >
>>
>
>


-- 
Regards
Shengjun


Re: create table question

2014-04-22 Thread Shengjun Xin
in the ql, you set relative path tmp/states, according to the error
message, you need to set absolute path


On Wed, Apr 23, 2014 at 11:23 AM, EdwardKing  wrote:

>  I use hadoop 2.2.0 and hive 0.13.0, I want to create a table from an
> existing file, states.hql is follows:
> CREATE EXTERNAL TABLE states(abbreviation string, full_name
> string)
> ROW FORMAT DELIMITED
> FIELDS TERMINATED BY '\t'
> LOCATION 'tmp/states' ;
>
>
> [hadoop@master ~]$ hadoop fs -ls
> 14/04/22 20:17:32 WARN util.NativeCodeLoader: Unable to load native-hadoop
> library for your platform... using builtin-java classes where applicable
> Found 1 items
> drwxr-xr-x   - hadoop supergroup  0 2014-04-22 20:02 tmp
>  [hadoop@master ~]$ hadoop fs -put states.txt tmp/states
> [hadoop@master ~]$ hadoop fs -ls tmp/states
> 14/04/22 20:17:19 WARN util.NativeCodeLoader: Unable to load native-hadoop
> library for your platform... using builtin-java classes where applicable
> Found 1 items
> -rw-r--r--   2 hadoop supergroup654 2014-04-22 20:02
> tmp/states/states.txt
>
>
> Then I execute states.hql
> [hadoop@master ~]$ hive -f states.hql
> 14/04/22 20:11:47 INFO Configuration.deprecation: mapred.reduce.tasks is
> deprecated. Instead, use mapreduce.job.reduces
> 14/04/22 20:11:47 INFO Configuration.deprecation: mapred.min.split.size is
> deprecated. Instead, use mapreduce.input.fileinputformat.split.minsize
> 14/04/22 20:11:47 INFO Configuration.deprecation:
> mapred.reduce.tasks.speculative.execution is deprecated. Instead, use
> mapreduce.reduce.speculative
> 14/04/22 20:11:47 INFO Configuration.deprecation:
> mapred.min.split.size.per.node is deprecated. Instead, use
> mapreduce.input.fileinputformat.split.minsize.per.node
> 14/04/22 20:11:47 INFO Configuration.deprecation:
> mapred.input.dir.recursive is deprecated. Instead, use
> mapreduce.input.fileinputformat.input.dir.recursive
> 14/04/22 20:11:47 INFO Configuration.deprecation:
> mapred.min.split.size.per.rack is deprecated. Instead, use
> mapreduce.input.fileinputformat.split.minsize.per.rack
> 14/04/22 20:11:47 INFO Configuration.deprecation: mapred.max.split.size is
> deprecated. Instead, use mapreduce.input.fileinputformat.split.maxsize
> 14/04/22 20:11:47 INFO Configuration.deprecation:
> mapred.committer.job.setup.cleanup.needed is deprecated. Instead, use
> mapreduce.job.committer.setup.cleanup.needed
> Logging initialized using configuration in
> jar:file:/home/software/apache-hive-0.13.0-bin/lib/hive-common-0.13.0.jar!/hive-log4j.properties
> FAILED: Execution Error, return code 1 from
> org.apache.hadoop.hive.ql.exec.DDLTask.
> MetaException(message:java.lang.IllegalArgumentException:
> java.net.URISyntaxException: Relative path in absolute URI:
> hdfs://master:9000./tmp/states)
>
>
> It raise following error,why? How to correct it?
> 2014-04-22 20:12:03,907 INFO  [main]: exec.DDLTask
> (DDLTask.java:createTable(4074)) - Default to LazySimpleSerDe for table
> states
> 2014-04-22 20:12:05,147 INFO  [main]: metastore.HiveMetaStore
> (HiveMetaStore.java:logInfo(624)) - 0: create_table:
> Table(tableName:states, dbName:default, owner:hadoop,
> createTime:1398222724, lastAccessTime:0, retention:0,
> sd:StorageDescriptor(cols:[FieldSchema(name:abbreviation, type:string,
> comment:null), FieldSchema(name:full_name, type:string, comment:null)],
> location:tmp/states, inputFormat:org.apache.hadoop.mapred.TextInputFormat,
> outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat,
> compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null,
> serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe,
> parameters:{serialization.format= , field.delim= }), bucketCols:[],
> sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[],
> skewedColValues:[], skewedColValueLocationMaps:{}),
> storedAsSubDirectories:false), partitionKeys:[],
> parameters:{EXTERNAL=TRUE}, viewOriginalText:null, viewExpandedText:null,
> tableType:EXTERNAL_TABLE)
> 2014-04-22 20:12:05,147 INFO  [main]: HiveMetaStore.audit
> (HiveMetaStore.java:logAuditEvent(306)) -
> ugi=hadoop ip=unknown-ip-addr cmd=create_table: Table(tableName:states,
> dbName:default, owner:hadoop, createTime:1398222724, lastAccessTime:0,
> retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:abbreviation,
> type:string, comment:null), FieldSchema(name:full_name, type:string,
> comment:null)], location:tmp/states,
> inputFormat:org.apache.hadoop.mapred.TextInputFormat,
> outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat,
> compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null,
> serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe,
> parameters:{serialization.format= , field.delim= }), bucketCols:[],
> sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[],
> skewedColValues:[], skewedColValueLocationMaps:{}),
> storedAsSubDirectories:false), partitionKeys:[],
> parameters:{EXTERNAL=TRUE}, viewOriginalText:null, viewExpandedText:nu

Re: question about hive sql

2014-04-21 Thread Shengjun Xin
You need to check the container log for the details


On Tue, Apr 22, 2014 at 10:27 AM, EdwardKing  wrote:

>  I use hive under hadoop 2.2.0, first I start hive
> [hadoop@master sbin]$ hive
> 14/04/21 19:06:32 INFO Configuration.deprecation:
> mapred.input.dir.recursive is deprecated. Instead, use
> mapreduce.input.fileinputformat.input.dir.recursive
> 14/04/21 19:06:32 INFO Configuration.deprecation: mapred.max.split.size is
> deprecated. Instead, use mapreduce.input.fileinputformat.split.maxsize
> 14/04/21 19:06:32 INFO Configuration.deprecation: mapred.min.split.size is
> deprecated. Instead, use mapreduce.input.fileinputformat.split.minsize
> 14/04/21 19:06:32 INFO Configuration.deprecation:
> mapred.min.split.size.per.rack is deprecated. Instead, use
> mapreduce.input.fileinputformat.split.minsize.per.rack
> 14/04/21 19:06:32 INFO Configuration.deprecation:
> mapred.min.split.size.per.node is deprecated. Instead, use
> mapreduce.input.fileinputformat.split.minsize.per.node
> 14/04/21 19:06:32 INFO Configuration.deprecation: mapred.reduce.tasks is
> deprecated. Instead, use mapreduce.job.reduces
> 14/04/21 19:06:32 INFO Configuration.deprecation:
> mapred.reduce.tasks.speculative.execution is deprecated. Instead, use
> mapreduce.reduce.speculative
> 14/04/21 19:06:32 WARN conf.Configuration:
> org.apache.hadoop.hive.conf.LoopingByteArrayInputStream@2128d0:an attempt
> to override final parameter:
> mapreduce.job.end-notification.max.retry.interval;  Ignoring.
> 14/04/21 19:06:32 WARN conf.Configuration:
> org.apache.hadoop.hive.conf.LoopingByteArrayInputStream@2128d0:an attempt
> to override final parameter: mapreduce.job.end-notification.max.attempts;
> Ignoring.
> Logging initialized using configuration in
> jar:file:/home/software/hive-0.11.0/lib/hive-common-0.11.0.jar!/hive-log4j.properties
> Hive history
> file=/tmp/hadoop/hive_job_log_hadoop_7623@master_201404211906_2069310090.txt
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in
> [jar:file:/home/software/hadoop-2.2.0/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in
> [jar:file:/home/software/hive-0.11.0/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
> explanation.
> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> Then I creat a table
> hive> create table test(id STRING);
> OK
> Time taken: 17.277 seconds
> Then  I insert some date into test
> hive> load data inpath 'a.txt' overwrite into table test;
> Loading data to table default.test
> rmr: DEPRECATED: Please use 'rm -r' instead.
> Deleted /user/hive/warehouse/test
> Table default.test stats: [num_partitions: 0, num_files: 1, num_rows: 0,
> total_size: 19, raw_data_size: 0]
> OK
> Time taken: 1.855 seconds
>
> hive> select * from test;
> OK
> China
> US
> Australia
> Time taken: 0.526 seconds, Fetched: 3 row(s)
> Now I use count command, I expected the result value is 3, but it runs
> failure!  Why? Where is wrong? I am puzzled with it for several days.
> Anyone could tell me how to correct it?
> hive> select count(*) from test;
> Total MapReduce jobs = 1
> Launching Job 1 out of 1
> Number of reduce tasks determined at compile time: 1
> In order to change the average load for a reducer (in bytes):
>   set hive.exec.reducers.bytes.per.reducer=
> In order to limit the maximum number of reducers:
>   set hive.exec.reducers.max=
> In order to set a constant number of reducers:
>   set mapred.reduce.tasks=
> Starting Job = job_1398132272370_0001, Tracking URL =
> http://master:8088/proxy/application_1398132272370_0001/
> Kill Command = /home/software/hadoop-2.2.0/bin/hadoop job  -kill
> job_1398132272370_0001
> Hadoop job information for Stage-1: number of mappers: 0; number of
> reducers: 0
> 2014-04-21 19:15:56,684 Stage-1 map = 0%,  reduce = 0%
> Ended Job = job_1398132272370_0001 with errors
> Error during job, obtaining debugging information...
> FAILED: Execution Error, return code 2 from
> org.apache.hadoop.hive.ql.exec.MapRedTask
> MapReduce Jobs Launched:
> Job 0:  HDFS Read: 0 HDFS Write: 0 FAIL
> Total MapReduce CPU Time Spent: 0 msec
> hive>
>
> Error information under
> http://172.11.12.6:8088/cluster/app/application_1398132272370_0001
> User:  hadoop
> Name:  select count(*) from test(Stage-1)
> Application Type:  MAPREDUCE
> State:  FAILED
> FinalStatus:  FAILED
> Started:  21-Apr-2014 19:14:55
> Elapsed:  57sec
> Tracking URL:  History
> Diagnostics:
> Application application_1398132272370_0001 failed 2 times due to AM
> Container for appattempt_1398132272370_0001_02 exited with exitCode: 1
> due to: Exception from container-launch:
> org.apache.hadoop.util.Shell$ExitCodeException:
> at org.apache.hadoop.util.Shell.runCommand(Shell.java:464)
> at org.apache.hadoop.util.Shell.run(Shell.java:379)
> at
> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.jav

Re: hive hbase integration

2014-04-17 Thread Shengjun Xin
For the first question, it's hive-hbase-handler, for the second question,
you can refer to
https://cwiki.apache.org/confluence/display/Hive/HBaseIntegration


On Fri, Apr 18, 2014 at 1:05 AM, Shushant Arora
wrote:

> what is hive storage handlers?
> What are the best practices for hive hbase integration?
>
>


-- 
Regards
Shengjun


Re: Hive install under hadoop

2014-04-14 Thread Shengjun Xin
Do you install hadoop correctly?


On Mon, Apr 14, 2014 at 4:22 PM, EdwardKing  wrote:

>  I want to use hive in hadoop2.2.0, so I execute following steps:
>
> [hadoop@master /]$ tar -xzf  hive-0.11.0.tar.gz
> [hadoop@master /]$ export HIVE_HOME=/home/software/hive
> [hadoop@master /]$ export PATH=${HIVE_HOME}/bin:${PATH}
> [hadoop@master /]$ hadoop fs -mkdir /tmp
> [hadoop@master /]$ hadoop fs -mkdir /user/hive/warehouse
> [hadoop@master /]$ hadoop fs -chmod g+w /tmp
> [hadoop@master /]$ hadoop fs -chmod g+w /user/hive/warehouse
> [hadoop@master /]$ hive
> Error creating temp dir in hadoop.tmp.dir file:/home/software/temp due to
> Permission denied
>
> How to make hive install success? Thanks.
>
>
>
>
> ---
> Confidentiality Notice: The information contained in this e-mail and any
> accompanying attachment(s)
> is intended only for the use of the intended recipient and may be
> confidential and/or privileged of
> Neusoft Corporation, its subsidiaries and/or its affiliates. If any reader
> of this communication is
> not the intended recipient, unauthorized use, forwarding, printing,
> storing, disclosure or copying
> is strictly prohibited, and may be unlawful.If you have received this
> communication in error,please
> immediately notify the sender by return e-mail, and delete the original
> message and all copies from
> your system. Thank you.
>
> ---
>
>


-- 
Regards
Shengjun


Re: pig,hive install over hadoop

2014-04-01 Thread Shengjun Xin
Install hive and pig completely and configure hive and pig to use the
existing hadoop cluster


On Tue, Apr 1, 2014 at 6:17 PM, Chinna Rao Lalam  wrote:

> Hi,
>
> You need to install hive completely because both are different
> processes(Installing hive is very simple and it require few
> configurations). Hive process to work, it required Hadoop processes.  I
> think same with the pig also(I did't try)...
>
>
> Hope It Helps,
> Chinna..
>
>
> On Tue, Apr 1, 2014 at 3:34 PM, Rahul Singh wrote:
>
>> Hi,
>>I have installed and configured hadoop. Now, I want to install hive
>> and pig, As per my understanding pig and hive internally uses hadoop. So is
>> there a way i can just install bare minimum hive or pig and take advantage
>> of already installed hadoop or i need to separately install and configure
>> complete hive and pig.
>>
>> Thanks,
>> -Rahul Singh
>>
>
>
>
> --
> Hope It Helps,
> Chinna
>



-- 
Regards
Shengjun