Hi, if i copied htrace-1.50.jar in map-reduce job. I got below error..
java.io.IOException: java.lang.reflect.InvocationTargetException at org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:383) at org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:360) at org.apache.hadoop.hbase.client.HConnectionManager.getConnection(HConnectionManager.java:244) at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:187) at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:164) at org.apache.hadoop.hive.hbase.HiveHBaseTableInputFormat.getSplits(HiveHBaseTableInputFormat.java:442) at org.apache.hadoop.hive.ql.io.HiveInputFormat.getSplits(HiveInputFormat.java:294) at org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getSplits(CombineHiveInputFormat.java:303) at org.apache.hadoop.mapreduce.JobSubmitter.writeOldSplits(JobSubmitter.java:518) at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:510) at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:392) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1268) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1265) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491) at org.apache.hadoop.mapreduce.Job.submit(Job.java:1265) at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:562) at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:557) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491) at org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:557) at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:548) at org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:425) at org.apache.hadoop.hive.ql.exec.mr.MapRedTask.execute(MapRedTask.java:136) at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:151) at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:65) at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1437) at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1215) at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1043) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:911) at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:259) at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:216) at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:413) at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:781) at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675) at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at org.apache.hadoop.util.RunJar.main(RunJar.java:212) Caused by: java.lang.reflect.InvocationTargetException at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27) at java.lang.reflect.Constructor.newInstance(Constructor.java:513) at org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:381) ... 42 more Caused by: java.lang.NoSuchMethodError: org.cloudera.htrace.Trace.startSpan(Ljava/lang/String;)Lorg/cloudera/htrace/TraceScope; at org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.exists(RecoverableZooKeeper.java:196) at org.apache.hadoop.hbase.zookeeper.ZKUtil.checkExists(ZKUtil.java:479) at org.apache.hadoop.hbase.zookeeper.ZKClusterId.readClusterIdZNode(ZKClusterId.java:65) at org.apache.hadoop.hbase.client.ZooKeeperRegistry.getClusterId(ZooKeeperRegistry.java:83) at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.retrieveClusterId(HConnectionManager.java:794) at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.<init>(HConnectionManager.java:627) ... 47 more On Thu, Dec 26, 2013 at 6:26 PM, Ted Yu <yuzhih...@gmail.com> wrote: > The error was due to htrace jar missing in class path of the map reduce > task. > > Cheers > > On Dec 26, 2013, at 4:05 AM, Vikas Parashar < > vikas.paras...@fosteringlinux.com> wrote: > > > Hi, > > > > I am integrating hive(0.12) with hbase(0.96). Everything is working fine > > there but get stuck between two quires. > > > > When i create table or select * from table then it's working fine . > > but in case of select count(*) from table it give me below error. > > > > > > 2013-12-26 13:25:01,864 ERROR ql.Driver > (SessionState.java:printError(419)) > > - FAILED: Execution Error, return code 2 from > > org.apache.hadoop.hive.ql.exec.mr.MapRedTask > > 2013-12-26 13:25:01,869 WARN mapreduce.Counters > > (AbstractCounters.java:getGroup(234)) - Group FileSystemCounters is > > deprecated. Use org.apache.hadoop.mapreduce.FileSystemCounter instead > > 2013-12-26 14:25:44,119 WARN mapreduce.JobSubmitter > > (JobSubmitter.java:copyAndConfigureFiles(149)) - Hadoop command-line > option > > parsing not performed. Implement the Tool interface and execute your > > application with ToolRunner to remedy this. > > 2013-12-26 14:26:14,677 WARN mapreduce.Counters > > (AbstractCounters.java:getGroup(234)) - Group > > org.apache.hadoop.mapred.Task$Counter is deprecated. Use > > org.apache.hadoop.mapreduce.TaskCounter instead > > 2013-12-26 14:26:33,613 WARN mapreduce.Counters > > (AbstractCounters.java:getGroup(234)) - Group > > org.apache.hadoop.mapred.Task$Counter is deprecated. Use > > org.apache.hadoop.mapreduce.TaskCounter instead > > 2013-12-26 14:27:30,355 WARN mapreduce.Counters > > (AbstractCounters.java:getGroup(234)) - Group > > org.apache.hadoop.mapred.Task$Counter is deprecated. Use > > org.apache.hadoop.mapreduce.TaskCounter instead > > 2013-12-26 14:27:32,479 WARN mapreduce.Counters > > (AbstractCounters.java:getGroup(234)) - Group > > org.apache.hadoop.mapred.Task$Counter is deprecated. Use > > org.apache.hadoop.mapreduce.TaskCounter instead > > 2013-12-26 14:27:32,528 ERROR exec.Task > (SessionState.java:printError(419)) > > - Ended Job = job_1388037394132_0013 with errors > > 2013-12-26 14:27:32,530 ERROR exec.Task > (SessionState.java:printError(419)) > > - Error during job, obtaining debugging information... > > 2013-12-26 14:27:32,538 ERROR exec.Task > (SessionState.java:printError(419)) > > - Examining task ID: task_1388037394132_0013_m_000000 (and more) from job > > job_1388037394132_0013 > > 2013-12-26 14:27:32,539 WARN shims.HadoopShimsSecure > > (Hadoop23Shims.java:getTaskAttemptLogUrl(72)) - Can't fetch tasklog: > > TaskLogServlet is not supported in MR2 mode. > > 2013-12-26 14:27:32,593 WARN shims.HadoopShimsSecure > > (Hadoop23Shims.java:getTaskAttemptLogUrl(72)) - Can't fetch tasklog: > > TaskLogServlet is not supported in MR2 mode. > > 2013-12-26 14:27:32,596 WARN shims.HadoopShimsSecure > > (Hadoop23Shims.java:getTaskAttemptLogUrl(72)) - Can't fetch tasklog: > > TaskLogServlet is not supported in MR2 mode. > > 2013-12-26 14:27:32,599 WARN shims.HadoopShimsSecure > > (Hadoop23Shims.java:getTaskAttemptLogUrl(72)) - Can't fetch tasklog: > > TaskLogServlet is not supported in MR2 mode. > > 2013-12-26 14:27:32,615 ERROR exec.Task > (SessionState.java:printError(419)) > > - > > Task with the most failures(4): > > ----- > > Task ID: > > task_1388037394132_0013_m_000000 > > > > URL: > > > > > http://ambari1.hadoop.com:8088/taskdetails.jsp?jobid=job_1388037394132_0013&tipid=task_1388037394132_0013_m_000000 > > ----- > > Diagnostic Messages for this Task: > > Error: java.io.IOException: java.io.IOException: > > java.lang.reflect.InvocationTargetException > > at > > > org.apache.hadoop.hive.io.HiveIOExceptionHandlerChain.handleRecordReaderCreationException(HiveIOExceptionHandlerChain.java:97) > > at > > > org.apache.hadoop.hive.io.HiveIOExceptionHandlerUtil.handleRecordReaderCreationException(HiveIOExceptionHandlerUtil.java:57) > > at > > > org.apache.hadoop.hive.ql.io.HiveInputFormat.getRecordReader(HiveInputFormat.java:244) > > at > > > org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getRecordReader(CombineHiveInputFormat.java:538) > > at > > > org.apache.hadoop.mapred.MapTask$TrackedRecordReader.<init>(MapTask.java:167) > > at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:408) > > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) > > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:162) > > at java.security.AccessController.doPrivileged(Native Method) > > at javax.security.auth.Subject.doAs(Subject.java:396) > > at > > > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491) > > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:157) > > Caused by: java.io.IOException: > java.lang.reflect.InvocationTargetException > > at > > > org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:383) > > at > > > org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:360) > > at > > > org.apache.hadoop.hbase.client.HConnectionManager.getConnection(HConnectionManager.java:244) > > at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:187) > > at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:164) > > at > > > org.apache.hadoop.hive.hbase.HiveHBaseTableInputFormat.getRecordReader(HiveHBaseTableInputFormat.java:91) > > at > > > org.apache.hadoop.hive.ql.io.HiveInputFormat.getRecordReader(HiveInputFormat.java:241) > > ... 9 more > > Caused by: java.lang.reflect.InvocationTargetException > > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > > at > > > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39) > > at > > > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27) > > at java.lang.reflect.Constructor.newInstance(Constructor.java:513) > > at > > > org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:381) > > ... 15 more > > Caused by: java.lang.NoClassDefFoundError: org/cloudera/htrace/Trace > > at > > > org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.exists(RecoverableZooKeeper.java:196) > > at org.apache.hadoop.hbase.zookeeper.ZKUtil.checkExists(ZKUtil.java:479) > > at > > > org.apache.hadoop.hbase.zookeeper.ZKClusterId.readClusterIdZNode(ZKClusterId.java:65) > > at > > > org.apache.hadoop.hbase.client.ZooKeeperRegistry.getClusterId(ZooKeeperRegistry.java:83) > > at > > > org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.retrieveClusterId(HConnectionManager.java:794) > > at > > > org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.<init>(HConnectionManager.java:627) > > ... 20 more > > Caused by: java.lang.ClassNotFoundException: org.cloudera.htrace.Trace > > at java.net.URLClassLoader$1.run(URLClassLoader.java:202) > > at java.security.AccessController.doPrivileged(Native Method) > > at java.net.URLClassLoader.findClass(URLClassLoader.java:190) > > at java.lang.ClassLoader.loadClass(ClassLoader.java:306) > > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301) > > at java.lang.ClassLoader.loadClass(ClassLoader.java:247) > > ... 26 more > > > > > > 2013-12-26 14:27:32,870 ERROR ql.Driver > (SessionState.java:printError(419)) > > - FAILED: Execution Error, return code 2 from org.apache. > > > > I think this error is related with mapred job. Whenever my query use the > > map-R then i get error. > > > > Any idea!! > > > > -- > > Thanks & Regards:- > > Vikas Parashar > > Sr. Linux administrator Cum Developer > > Mobile: +91 958 208 8852 > > Email: vikas.paras...@fosteringlinglinux.com > -- Thanks & Regards:- Vikas Parashar Sr. Linux administrator Cum Developer Mobile: +91 958 208 8852 Email: vikas.paras...@fosteringlinglinux.com