Hi All
When I am trying to run a select count(*) from table_name;  getting an 
error as given below..what could be the reason  , pls help.

hive> select count(*) from u_data ;
Total MapReduce jobs = 1
Launching Job 1 out of 1
Number of reduce tasks determined at compile time: 1
In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
  set mapred.reduce.tasks=<number>
java.net.UnknownHostException: 01hw357381.tcsgegdc.com: 
01hw357381.tcsgegdc.com: unknown error
        at java.net.InetAddress.getLocalHost(InetAddress.java:1484)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:439)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1296)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1293)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:422)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1293)
        at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:562)
        at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:557)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:422)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
        at 
org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:557)
        at 
org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:548)
        at 
org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:425)
        at 
org.apache.hadoop.hive.ql.exec.mr.MapRedTask.execute(MapRedTask.java:144)
        at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:151)
        at 
org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:65)
        at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1414)
        at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1192)
        at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1020)
        at org.apache.hadoop.hive.ql.Driver.run(Driver.java:888)
        at 
org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:259)
        at 
org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:216)
        at 
org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:413)
        at 
org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:781)
        at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675)
        at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:497)
        at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
        at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
Caused by: java.net.UnknownHostException: 01hw357381.tcsgegdc.com: unknown 
error
        at java.net.Inet4AddressImpl.lookupAllHostAddr(Native Method)
        at java.net.InetAddress$2.lookupAllHostAddr(InetAddress.java:907)
        at 
java.net.InetAddress.getAddressesFromNameService(InetAddress.java:1302)
        at java.net.InetAddress.getLocalHost(InetAddress.java:1479)
        ... 34 more
Job Submission failed with exception 
'java.net.UnknownHostException(01hw357381.tcsgegdc.com: 
01hw357381.tcsgegdc.com: unknown error)'
FAILED: Execution Error, return code 1 from 
org.apache.hadoop.hive.ql.exec.mr.MapRedTask
hive>



Regards
Ajeet Ojha
Tata Consultancy Services
Ph:- 0120 6744953
Buzz:- 412-4953
Cell:- 9811220828
Mailto: ajee...@tcs.com
Website: http://www.tcs.com
____________________________________________
Experience certainty.   IT Services
                        Business Solutions
                        Consulting
____________________________________________




From:   "Lonikar, Kiran" <kloni...@informatica.com>
To:     "user@hive.apache.org" <user@hive.apache.org>
Date:   07/09/2015 02:36 PM
Subject:        RE: Add file/jar location for jdbc clients



Never mind... It works for HDFS. I can copy to HDFS and specify the HDFS 
file path as hdfs://<namenode>/path/to/file.

-Kiran

-----Original Message-----
From: Lonikar, Kiran [mailto:kloni...@informatica.com] 
Sent: Wednesday, July 08, 2015 9:01 PM
To: user@hive.apache.org
Subject: Add file/jar location for jdbc clients

Hi,

The add jar/file syntax stated on hive language manual is as below:

ADD { FILE[S] | JAR[S] | ARCHIVE[S] } <filepath1> [<filepath2>]*

I have an application which remotely connects to hiveserver2 over jdbc. It 
needs to add local resources before execution of any hive queries. When it 
tries to execute the above add jar or add file commands, with the filepath 
being from the jdbc client's local filesystem, it gets an error. But when 
it gives path local to hiveserver2 filesystem, it works.

What are my options here? One obvious one is to do scp to hiveserver's 
filesystem to say /tmp and then do add jar.

The hive 1.2.0 allows ivy urls. I dont think i can use that since my 
application generates the files and i don't want to setup ivy repo where i 
will need to copy the files before running add jar.

Isn't there any way to specify client's filesystem paths?

-Kiran

=====-----=====-----=====
Notice: The information contained in this e-mail
message and/or attachments to it may contain 
confidential or privileged information. If you are 
not the intended recipient, any dissemination, use, 
review, distribution, printing or copying of the 
information contained in this e-mail message 
and/or attachments to it are strictly prohibited. If 
you have received this communication in error, 
please notify us by reply e-mail or telephone and 
immediately and permanently delete the message 
and any attachments. Thank you


Reply via email to