Re: Reducers Failing
Are datanodes dead? On Thu, May 22, 2014 at 2:23 PM, Sreenath wrote: > Hi All, > > We are running a hadoop cluster and many of our hive queries are failing > in the reduce phase with the following error > > java.io.IOException: All datanodes *.*.*.*:50230 are bad. Aborting... > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.processDatanodeError(DFSClient.java:3290) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2200(DFSClient.java:2783) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2987) > > Any IDead why this might be happening ? > -- > Sreenath S Kamath > Bangalore > Ph No:+91-9590989106 > -- Regards Shengjun
Re: Connecting hive to SAP BO
hive –auxpath /path-to-/csvserde.jar On Tue, May 20, 2014 at 12:59 PM, Chhaya Vishwakarma < chhaya.vishwaka...@lntinfotech.com> wrote: > > > Hi, > > I have connected SAP BO to Hive using a ODBC driver. I am able to see the > database and table in hive. but when i fetch data from hive it gives error > as > > org.apache.hadoop.hìve.serde2.SerDeExceptio SerDe > com,bizohive.serde.csv.CSVSerde does not exist > > Can ayone suggest where i should put csvserde jar in SAP BO > > > > > > Regards, > > Chhaya Vishwakarma > > > > -- > The contents of this e-mail and any attachment(s) may contain confidential > or privileged information for the intended recipient(s). Unintended > recipients are prohibited from taking action on the basis of information in > this e-mail and using or disseminating the information, and must notify the > sender and delete it from their system. L&T Infotech will not accept > responsibility or liability for the accuracy or completeness of, or the > presence of any virus or disabling code in this e-mail" > -- Regards Shengjun
Re: Hive 0.13.0 Memory Leak
We found a hive jdbc connection leak https://issues.apache.org/jira/browse/HIVE-6866, is it similar to your issue? On Thu, May 15, 2014 at 9:29 PM, Bryan Jeffrey wrote: > All, > > We are running Hadoop 2.2.0 and Hive 0.13.0. One typical application is > to load data (as text), and then convert that data to ORC to decrease query > time. When running these processes we are seeing significant memory leaks > (leaking 4 GB in about 5 days). > > We're running HiveServer2 with the following options: > > root 28136 1 51 May14 ?09:51:09 /usr/java/latest/bin/java > -Xmx2048m -XX:NewRatio=12 -Xms10m -XX:MaxHeapFreeRatio=40 > -XX:MinHeapFreeRatio=15 -XX:-UseGCOverheadLimit -XX:MaxPermSize=1024m > -XX:NewRatio=12 -Xms10m -XX:MaxHeapFreeRatio=40 -XX:MinHeapFreeRatio=15 > -XX:-UseGCOverheadLimit -XX:MaxPermSize=1024m > -Dhadoop.log.dir=/opt/hadoop/latest-hadoop/logs > -Dhadoop.log.file=hadoop.log -Dhadoop.home.dir=/opt/hadoop/latest-hadoop > -Dhadoop.id.str= -Dhadoop.root.logger=INFO,console > -Djava.library.path=/opt/hadoop/latest-hadoop/lib/native > -Dhadoop.policy.file=hadoop-policy.xml -Djava.net.preferIPv4Stack=true > -Dhadoop.security.logger=INFO,NullAppender org.apache.hadoop.util.RunJar > /opt/hadoop/latest-hive/lib/hive-service-0.13.0.jar > org.apache.hive.service.server.HiveServer2 > > Typical ORC conversion query looks like the following: > > HIVE -u jdbc:hive2://hive_server:10002/db -n root --hiveconf > hive.exec.dynamic.partition.mode=nonstrict --hiveconf > hive.enforce.sorting=true --hiveconf $SET_QUEUE -e "insert into table > orc_table partition (range) select * from loading_text_table; " > > I saw a couple of tickets for memory leaks, but they seemed to deal with > failed queries. The memory usage increases in a linear fashion. Jobs all > succeed until memory limit exceeded. > > Is there an open bug for memory leaks associated with successful jobs in > HS2? Is there a fix for this issue? > > Regards, > > Bryan Jeffrey > -- Regards Shengjun
Re: issue about Job running in-process (local Hadoop)
According to the log, you configured a local hadoop, you need to check you configuration On Mon, May 12, 2014 at 3:46 PM, ch huang wrote: > hi,maillist: >i try hive 0.12,but it always run job in local mode ,why? i install > hive on a seperate hox ,and install hadoop client on it ,and configure the > client connect to my hadoop cluster > > hive> select count(*) from media_visit_info; > Total MapReduce jobs = 1 > Launching Job 1 out of 1 > Number of reduce tasks determined at compile time: 1 > In order to change the average load for a reducer (in bytes): > set hive.exec.reducers.bytes.per.reducer= > In order to limit the maximum number of reducers: > set hive.exec.reducers.max= > In order to set a constant number of reducers: > set mapred.reduce.tasks= > 14/05/12 15:36:52 WARN conf.Configuration: > file:/tmp/root/hive_2014-05-12_15-36-48_130_3729255905600309633-1/-local-10002/jobconf.xml:an > attempt to override final parameter: > mapreduce.job.end-notification.max.retry.interval; Ignoring. > 14/05/12 15:36:52 WARN conf.Configuration: > file:/tmp/root/hive_2014-05-12_15-36-48_130_3729255905600309633-1/-local-10002/jobconf.xml:an > attempt to override final parameter: > mapreduce.job.end-notification.max.attempts; Ignoring. > 14/05/12 15:36:52 INFO Configuration.deprecation: > mapred.input.dir.recursive is deprecated. Instead, use > mapreduce.input.fileinputformat.input.dir.recursive > 14/05/12 15:36:52 INFO Configuration.deprecation: mapred.max.split.size is > deprecated. Instead, use mapreduce.input.fileinputformat.split.maxsize > 14/05/12 15:36:52 INFO Configuration.deprecation: mapred.min.split.size is > deprecated. Instead, use mapreduce.input.fileinputformat.split.minsize > 14/05/12 15:36:52 INFO Configuration.deprecation: > mapred.min.split.size.per.rack is deprecated. Instead, use > mapreduce.input.fileinputformat.split.minsize.per.rack > 14/05/12 15:36:52 INFO Configuration.deprecation: > mapred.min.split.size.per.node is deprecated. Instead, use > mapreduce.input.fileinputformat.split.minsize.per.node > 14/05/12 15:36:52 INFO Configuration.deprecation: mapred.reduce.tasks is > deprecated. Instead, use mapreduce.job.reduces > 14/05/12 15:36:52 INFO Configuration.deprecation: > mapred.reduce.tasks.speculative.execution is deprecated. Instead, use > mapreduce.reduce.speculative > Execution log at: > /tmp/root/root_20140512153636_d3088311-cdbe-44c4-8622-e47dce685f34.log > Job running in-process (local Hadoop) > 2014-05-12 15:36:56,257 null map = 0%, reduce = 0% > 2014-05-12 15:36:59,275 null map = 100%, reduce = 0% > 2014-05-12 15:37:01,314 null map = 100%, reduce = 100% > Ended Job = job_local1940121222_0001 > Execution completed successfully > MapredLocal task succeeded > OK > 3093138 > Time taken: 14.136 seconds, Fetched: 1 row(s) > -- Regards Shengjun
Re: java.lang.IncompatibleClassChangeError in hive
Check the hive-env.sh and environment variables, whether the HADOOP_HOME is what you want? On Tue, May 6, 2014 at 11:33 AM, chandra kant < chandralakshmikan...@gmail.com> wrote: > > Hi, > When running bin/hive , i get the following error :- > Exception in thread "main" java.lang.IncompatibleClassChangeError: Found > interface jline.Terminal, but class was expected > at jline.ConsoleReader.(ConsoleReader.java:191) > at jline.ConsoleReader.(ConsoleReader.java:186) > at jline.ConsoleReader.(ConsoleReader.java:174) > at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:719) > at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > at java.lang.reflect.Method.invoke(Method.java:597) > at org.apache.hadoop.util.RunJar.main(RunJar.java:160) > > I think this comes when there is version mismatching between hadoop > version and hive version. In my case it is - hadoop-1.2.1 and hive-0.11.0 , > which were working perfectly fine till yesterday. And yes, I unsuccessfully > tried to run hadoop-2 on my system and later reverted back to hadoop-1. And > , it occurred ... > > Any help in debugging..? > > -- > Chandra > > -- Regards Shengjun
Re: Number of hive-server2 threads increments after jdbc connection
Is it same with https://issues.apache.org/jira/browse/HIVE-6866 ? On Tue, Apr 29, 2014 at 8:26 PM, Dima Fadeyev wrote: > Hi, Chinna. Thanks for your reply. > > Yes, modifying code solves the problem. This is what my code looks like (a > piece of it): > > Connection con = > DriverManager.getConnection("jdbc:hive2://localhost:1/default", "hive", > "hive"); > Statement stmt = con.createStatement(); > String tableName = "testHiveDriverTable"; > stmt.execute("drop table if exists " + tableName); > //stmt.close(); > > When I uncomment the last line, the number of hive-server2 threads doesn't > keep incrementing to infinity. However I'm investigating the issue where > the code is not really my code. Is there a way to correct this behavior > from within hive-server2 without changing the client's code? > > El 29/04/14 14:05, Chinna Rao Lalam escribió: > > Hi, > > In your code if more connections and statements are created?. If so, > closed those connections? > After use close unused connections and statements. > > > Hope It Helps, > Chinna > > > On Tue, Apr 29, 2014 at 3:47 PM, Dima Fadeyev wrote: > >> Hello everyone, >> >> When I run a jdbc example from >> https://cwiki.apache.org/confluence/display/Hive/HiveServer2+Clients#HiveServer2Clients-JDBCClientSampleCodeagainst >> my hive server, the number of hive-server2 threads increments. If I >> execute it long enough I either start seeing exceptions >> >> Exception in thread "main" java.sql.SQLException: >> org.apache.thrift.TApplicationException: Internal error processing >> ExecuteStatement >> at org.apache.hive.jdbc.HiveStatement.execute(HiveStatement.java:203) >> at HiveJdbcClient.main(HiveJdbcClient.java:24) >> Caused by: org.apache.thrift.TApplicationException: Internal error >> processing ExecuteStatement >> at >> org.apache.thrift.TApplicationException.read(TApplicationException.java:108) >> at >> org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:71) >> at >> org.apache.hive.service.cli.thrift.TCLIService$Client.recv_ExecuteStatement(TCLIService.java:213) >> at >> org.apache.hive.service.cli.thrift.TCLIService$Client.ExecuteStatement(TCLIService.java:200) >> at org.apache.hive.jdbc.HiveStatement.execute(HiveStatement.java:197) >> ... 1 more >> >> or I bump into a limit of connections with Zookeeper (each hive-server2 >> thread maintains a connection with Zookeeper. I have hive.support.concurrency >> enabled). >> >> In either case I can't connect to hive server after that. >> >> I've tried this on hive 0.10 (cdh 4.4) and hive 0.12 (cdh 5.0 and hdp >> 2.0.6) with same results. >> >> Please, could anyone help me resolve this. >> Thanks in advance. >> >> >> > > > -- > Hope It Helps, > Chinna > > > -- Regards Shengjun
Re: [ANNOUNCE] New Hive Committers - Prasanth J and Vaibhav Gumashta
Congratulations ~~ On Fri, Apr 25, 2014 at 10:33 AM, Carl Steinbach wrote: > + Prasanth's correct email address > > > On Thu, Apr 24, 2014 at 7:31 PM, Xuefu Zhang wrote: > >> Congratulations to Prasanth and Vaibhav! >> >> --Xuefu >> >> >> On Thu, Apr 24, 2014 at 7:26 PM, Carl Steinbach wrote: >> >> > The Apache Hive PMC has voted to make Prasanth J and Vaibhav >> > Gumashta committers on the Apache Hive Project. >> > >> > Please join me in congratulating Prasanth and Vaibhav! >> > >> > Thanks. >> > >> > - Carl >> > >> > > -- Regards Shengjun
Re: create table question
in the ql, you set relative path tmp/states, according to the error message, you need to set absolute path On Wed, Apr 23, 2014 at 11:23 AM, EdwardKing wrote: > I use hadoop 2.2.0 and hive 0.13.0, I want to create a table from an > existing file, states.hql is follows: > CREATE EXTERNAL TABLE states(abbreviation string, full_name > string) > ROW FORMAT DELIMITED > FIELDS TERMINATED BY '\t' > LOCATION 'tmp/states' ; > > > [hadoop@master ~]$ hadoop fs -ls > 14/04/22 20:17:32 WARN util.NativeCodeLoader: Unable to load native-hadoop > library for your platform... using builtin-java classes where applicable > Found 1 items > drwxr-xr-x - hadoop supergroup 0 2014-04-22 20:02 tmp > [hadoop@master ~]$ hadoop fs -put states.txt tmp/states > [hadoop@master ~]$ hadoop fs -ls tmp/states > 14/04/22 20:17:19 WARN util.NativeCodeLoader: Unable to load native-hadoop > library for your platform... using builtin-java classes where applicable > Found 1 items > -rw-r--r-- 2 hadoop supergroup654 2014-04-22 20:02 > tmp/states/states.txt > > > Then I execute states.hql > [hadoop@master ~]$ hive -f states.hql > 14/04/22 20:11:47 INFO Configuration.deprecation: mapred.reduce.tasks is > deprecated. Instead, use mapreduce.job.reduces > 14/04/22 20:11:47 INFO Configuration.deprecation: mapred.min.split.size is > deprecated. Instead, use mapreduce.input.fileinputformat.split.minsize > 14/04/22 20:11:47 INFO Configuration.deprecation: > mapred.reduce.tasks.speculative.execution is deprecated. Instead, use > mapreduce.reduce.speculative > 14/04/22 20:11:47 INFO Configuration.deprecation: > mapred.min.split.size.per.node is deprecated. Instead, use > mapreduce.input.fileinputformat.split.minsize.per.node > 14/04/22 20:11:47 INFO Configuration.deprecation: > mapred.input.dir.recursive is deprecated. Instead, use > mapreduce.input.fileinputformat.input.dir.recursive > 14/04/22 20:11:47 INFO Configuration.deprecation: > mapred.min.split.size.per.rack is deprecated. Instead, use > mapreduce.input.fileinputformat.split.minsize.per.rack > 14/04/22 20:11:47 INFO Configuration.deprecation: mapred.max.split.size is > deprecated. Instead, use mapreduce.input.fileinputformat.split.maxsize > 14/04/22 20:11:47 INFO Configuration.deprecation: > mapred.committer.job.setup.cleanup.needed is deprecated. Instead, use > mapreduce.job.committer.setup.cleanup.needed > Logging initialized using configuration in > jar:file:/home/software/apache-hive-0.13.0-bin/lib/hive-common-0.13.0.jar!/hive-log4j.properties > FAILED: Execution Error, return code 1 from > org.apache.hadoop.hive.ql.exec.DDLTask. > MetaException(message:java.lang.IllegalArgumentException: > java.net.URISyntaxException: Relative path in absolute URI: > hdfs://master:9000./tmp/states) > > > It raise following error,why? How to correct it? > 2014-04-22 20:12:03,907 INFO [main]: exec.DDLTask > (DDLTask.java:createTable(4074)) - Default to LazySimpleSerDe for table > states > 2014-04-22 20:12:05,147 INFO [main]: metastore.HiveMetaStore > (HiveMetaStore.java:logInfo(624)) - 0: create_table: > Table(tableName:states, dbName:default, owner:hadoop, > createTime:1398222724, lastAccessTime:0, retention:0, > sd:StorageDescriptor(cols:[FieldSchema(name:abbreviation, type:string, > comment:null), FieldSchema(name:full_name, type:string, comment:null)], > location:tmp/states, inputFormat:org.apache.hadoop.mapred.TextInputFormat, > outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, > compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, > serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, > parameters:{serialization.format= , field.delim= }), bucketCols:[], > sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], > skewedColValues:[], skewedColValueLocationMaps:{}), > storedAsSubDirectories:false), partitionKeys:[], > parameters:{EXTERNAL=TRUE}, viewOriginalText:null, viewExpandedText:null, > tableType:EXTERNAL_TABLE) > 2014-04-22 20:12:05,147 INFO [main]: HiveMetaStore.audit > (HiveMetaStore.java:logAuditEvent(306)) - > ugi=hadoop ip=unknown-ip-addr cmd=create_table: Table(tableName:states, > dbName:default, owner:hadoop, createTime:1398222724, lastAccessTime:0, > retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:abbreviation, > type:string, comment:null), FieldSchema(name:full_name, type:string, > comment:null)], location:tmp/states, > inputFormat:org.apache.hadoop.mapred.TextInputFormat, > outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, > compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, > serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, > parameters:{serialization.format= , field.delim= }), bucketCols:[], > sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], > skewedColValues:[], skewedColValueLocationMaps:{}), > storedAsSubDirectories:false), partitionKeys:[], > parameters:{EXTERNAL=TRUE}, viewOriginalText:null, viewExpandedText:nu
Re: question about hive sql
You need to check the container log for the details On Tue, Apr 22, 2014 at 10:27 AM, EdwardKing wrote: > I use hive under hadoop 2.2.0, first I start hive > [hadoop@master sbin]$ hive > 14/04/21 19:06:32 INFO Configuration.deprecation: > mapred.input.dir.recursive is deprecated. Instead, use > mapreduce.input.fileinputformat.input.dir.recursive > 14/04/21 19:06:32 INFO Configuration.deprecation: mapred.max.split.size is > deprecated. Instead, use mapreduce.input.fileinputformat.split.maxsize > 14/04/21 19:06:32 INFO Configuration.deprecation: mapred.min.split.size is > deprecated. Instead, use mapreduce.input.fileinputformat.split.minsize > 14/04/21 19:06:32 INFO Configuration.deprecation: > mapred.min.split.size.per.rack is deprecated. Instead, use > mapreduce.input.fileinputformat.split.minsize.per.rack > 14/04/21 19:06:32 INFO Configuration.deprecation: > mapred.min.split.size.per.node is deprecated. Instead, use > mapreduce.input.fileinputformat.split.minsize.per.node > 14/04/21 19:06:32 INFO Configuration.deprecation: mapred.reduce.tasks is > deprecated. Instead, use mapreduce.job.reduces > 14/04/21 19:06:32 INFO Configuration.deprecation: > mapred.reduce.tasks.speculative.execution is deprecated. Instead, use > mapreduce.reduce.speculative > 14/04/21 19:06:32 WARN conf.Configuration: > org.apache.hadoop.hive.conf.LoopingByteArrayInputStream@2128d0:an attempt > to override final parameter: > mapreduce.job.end-notification.max.retry.interval; Ignoring. > 14/04/21 19:06:32 WARN conf.Configuration: > org.apache.hadoop.hive.conf.LoopingByteArrayInputStream@2128d0:an attempt > to override final parameter: mapreduce.job.end-notification.max.attempts; > Ignoring. > Logging initialized using configuration in > jar:file:/home/software/hive-0.11.0/lib/hive-common-0.11.0.jar!/hive-log4j.properties > Hive history > file=/tmp/hadoop/hive_job_log_hadoop_7623@master_201404211906_2069310090.txt > SLF4J: Class path contains multiple SLF4J bindings. > SLF4J: Found binding in > [jar:file:/home/software/hadoop-2.2.0/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in > [jar:file:/home/software/hive-0.11.0/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an > explanation. > SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] > Then I creat a table > hive> create table test(id STRING); > OK > Time taken: 17.277 seconds > Then I insert some date into test > hive> load data inpath 'a.txt' overwrite into table test; > Loading data to table default.test > rmr: DEPRECATED: Please use 'rm -r' instead. > Deleted /user/hive/warehouse/test > Table default.test stats: [num_partitions: 0, num_files: 1, num_rows: 0, > total_size: 19, raw_data_size: 0] > OK > Time taken: 1.855 seconds > > hive> select * from test; > OK > China > US > Australia > Time taken: 0.526 seconds, Fetched: 3 row(s) > Now I use count command, I expected the result value is 3, but it runs > failure! Why? Where is wrong? I am puzzled with it for several days. > Anyone could tell me how to correct it? > hive> select count(*) from test; > Total MapReduce jobs = 1 > Launching Job 1 out of 1 > Number of reduce tasks determined at compile time: 1 > In order to change the average load for a reducer (in bytes): > set hive.exec.reducers.bytes.per.reducer= > In order to limit the maximum number of reducers: > set hive.exec.reducers.max= > In order to set a constant number of reducers: > set mapred.reduce.tasks= > Starting Job = job_1398132272370_0001, Tracking URL = > http://master:8088/proxy/application_1398132272370_0001/ > Kill Command = /home/software/hadoop-2.2.0/bin/hadoop job -kill > job_1398132272370_0001 > Hadoop job information for Stage-1: number of mappers: 0; number of > reducers: 0 > 2014-04-21 19:15:56,684 Stage-1 map = 0%, reduce = 0% > Ended Job = job_1398132272370_0001 with errors > Error during job, obtaining debugging information... > FAILED: Execution Error, return code 2 from > org.apache.hadoop.hive.ql.exec.MapRedTask > MapReduce Jobs Launched: > Job 0: HDFS Read: 0 HDFS Write: 0 FAIL > Total MapReduce CPU Time Spent: 0 msec > hive> > > Error information under > http://172.11.12.6:8088/cluster/app/application_1398132272370_0001 > User: hadoop > Name: select count(*) from test(Stage-1) > Application Type: MAPREDUCE > State: FAILED > FinalStatus: FAILED > Started: 21-Apr-2014 19:14:55 > Elapsed: 57sec > Tracking URL: History > Diagnostics: > Application application_1398132272370_0001 failed 2 times due to AM > Container for appattempt_1398132272370_0001_02 exited with exitCode: 1 > due to: Exception from container-launch: > org.apache.hadoop.util.Shell$ExitCodeException: > at org.apache.hadoop.util.Shell.runCommand(Shell.java:464) > at org.apache.hadoop.util.Shell.run(Shell.java:379) > at > org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.jav
Re: hive hbase integration
For the first question, it's hive-hbase-handler, for the second question, you can refer to https://cwiki.apache.org/confluence/display/Hive/HBaseIntegration On Fri, Apr 18, 2014 at 1:05 AM, Shushant Arora wrote: > what is hive storage handlers? > What are the best practices for hive hbase integration? > > -- Regards Shengjun
Re: Hive install under hadoop
Do you install hadoop correctly? On Mon, Apr 14, 2014 at 4:22 PM, EdwardKing wrote: > I want to use hive in hadoop2.2.0, so I execute following steps: > > [hadoop@master /]$ tar -xzf hive-0.11.0.tar.gz > [hadoop@master /]$ export HIVE_HOME=/home/software/hive > [hadoop@master /]$ export PATH=${HIVE_HOME}/bin:${PATH} > [hadoop@master /]$ hadoop fs -mkdir /tmp > [hadoop@master /]$ hadoop fs -mkdir /user/hive/warehouse > [hadoop@master /]$ hadoop fs -chmod g+w /tmp > [hadoop@master /]$ hadoop fs -chmod g+w /user/hive/warehouse > [hadoop@master /]$ hive > Error creating temp dir in hadoop.tmp.dir file:/home/software/temp due to > Permission denied > > How to make hive install success? Thanks. > > > > > --- > Confidentiality Notice: The information contained in this e-mail and any > accompanying attachment(s) > is intended only for the use of the intended recipient and may be > confidential and/or privileged of > Neusoft Corporation, its subsidiaries and/or its affiliates. If any reader > of this communication is > not the intended recipient, unauthorized use, forwarding, printing, > storing, disclosure or copying > is strictly prohibited, and may be unlawful.If you have received this > communication in error,please > immediately notify the sender by return e-mail, and delete the original > message and all copies from > your system. Thank you. > > --- > > -- Regards Shengjun
Re: pig,hive install over hadoop
Install hive and pig completely and configure hive and pig to use the existing hadoop cluster On Tue, Apr 1, 2014 at 6:17 PM, Chinna Rao Lalam wrote: > Hi, > > You need to install hive completely because both are different > processes(Installing hive is very simple and it require few > configurations). Hive process to work, it required Hadoop processes. I > think same with the pig also(I did't try)... > > > Hope It Helps, > Chinna.. > > > On Tue, Apr 1, 2014 at 3:34 PM, Rahul Singh wrote: > >> Hi, >>I have installed and configured hadoop. Now, I want to install hive >> and pig, As per my understanding pig and hive internally uses hadoop. So is >> there a way i can just install bare minimum hive or pig and take advantage >> of already installed hadoop or i need to separately install and configure >> complete hive and pig. >> >> Thanks, >> -Rahul Singh >> > > > > -- > Hope It Helps, > Chinna > -- Regards Shengjun