[jira] [Commented] (HIVE-7916) Snappy-java error when running hive query on spark [Spark Branch]
[ https://issues.apache.org/jira/browse/HIVE-7916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14183309#comment-14183309 ] Xuefu Zhang commented on HIVE-7916: --- Yeah. I'm on Ubuntu, and I didn't set up anything to HADOOP_OPTS. It just magically works now. :) > Snappy-java error when running hive query on spark [Spark Branch] > - > > Key: HIVE-7916 > URL: https://issues.apache.org/jira/browse/HIVE-7916 > Project: Hive > Issue Type: Bug > Components: Spark >Reporter: Xuefu Zhang > Labels: Spark-M1 > > Recently spark branch upgraded its dependency on Spark to 1.1.0-SNAPSHOT. > While the new version addressed some lib conflicts (such as guava), I'm > afraid that it also introduced new problems. The following might be one, when > I set the master URL to be a spark standalone cluster: > {code} > hive> set hive.execution.engine=spark; > hive> set spark.serializer=org.apache.spark.serializer.KryoSerializer; > hive> set spark.master=spark://xzdt:7077; > hive> select name, avg(value) from dec group by name; > 14/08/28 16:41:52 INFO storage.MemoryStore: Block broadcast_0 stored as > values in memory (estimated size 333.0 KB, free 128.0 MB) > java.lang.reflect.InvocationTargetException > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:601) > at org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317) > at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219) > at org.xerial.snappy.Snappy.(Snappy.java:44) > at org.xerial.snappy.SnappyOutputStream.(SnappyOutputStream.java:79) > at > org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:124) > at > org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:207) > at > org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:83) > at > org.apache.spark.broadcast.TorrentBroadcast.(TorrentBroadcast.scala:68) > at > org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:36) > at > org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:29) > at > org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:62) > at org.apache.spark.SparkContext.broadcast(SparkContext.scala:809) > at org.apache.spark.rdd.HadoopRDD.(HadoopRDD.scala:116) > at org.apache.spark.SparkContext.hadoopRDD(SparkContext.scala:541) > at > org.apache.spark.api.java.JavaSparkContext.hadoopRDD(JavaSparkContext.scala:318) > at > org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generateRDD(SparkPlanGenerator.java:160) > at > org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generate(SparkPlanGenerator.java:88) > at > org.apache.hadoop.hive.ql.exec.spark.SparkClient.execute(SparkClient.java:156) > at > org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.submit(SparkSessionImpl.java:52) > at > org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:77) > at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:161) > at > org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85) > at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1537) > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1304) > at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1116) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:940) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:930) > at > org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:246) > at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:198) > at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:408) > at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:781) > at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675) > at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:601) > at org.apache.hadoop.util.RunJar.main(RunJar.java:212) > Caused by: java.lang.UnsatisfiedLinkError: no snappyjava in java.library.path > at java.lang.ClassLoader.loadLibrary(ClassLoader.java
[jira] [Commented] (HIVE-7916) Snappy-java error when running hive query on spark [Spark Branch]
[ https://issues.apache.org/jira/browse/HIVE-7916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14183290#comment-14183290 ] Suhas Satish commented on HIVE-7916: Not sure what solved it for you, but setting this seems to work for me on a Mac OS X - export HADOOP_OPTS="-Dorg.xerial.snappy.tempdir=/tmp -Dorg.xerial.snappy.lib.name=libsnappyjava.jnilib $HADOOP_OPTS" > Snappy-java error when running hive query on spark [Spark Branch] > - > > Key: HIVE-7916 > URL: https://issues.apache.org/jira/browse/HIVE-7916 > Project: Hive > Issue Type: Bug > Components: Spark >Reporter: Xuefu Zhang > Labels: Spark-M1 > > Recently spark branch upgraded its dependency on Spark to 1.1.0-SNAPSHOT. > While the new version addressed some lib conflicts (such as guava), I'm > afraid that it also introduced new problems. The following might be one, when > I set the master URL to be a spark standalone cluster: > {code} > hive> set hive.execution.engine=spark; > hive> set spark.serializer=org.apache.spark.serializer.KryoSerializer; > hive> set spark.master=spark://xzdt:7077; > hive> select name, avg(value) from dec group by name; > 14/08/28 16:41:52 INFO storage.MemoryStore: Block broadcast_0 stored as > values in memory (estimated size 333.0 KB, free 128.0 MB) > java.lang.reflect.InvocationTargetException > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:601) > at org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317) > at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219) > at org.xerial.snappy.Snappy.(Snappy.java:44) > at org.xerial.snappy.SnappyOutputStream.(SnappyOutputStream.java:79) > at > org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:124) > at > org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:207) > at > org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:83) > at > org.apache.spark.broadcast.TorrentBroadcast.(TorrentBroadcast.scala:68) > at > org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:36) > at > org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:29) > at > org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:62) > at org.apache.spark.SparkContext.broadcast(SparkContext.scala:809) > at org.apache.spark.rdd.HadoopRDD.(HadoopRDD.scala:116) > at org.apache.spark.SparkContext.hadoopRDD(SparkContext.scala:541) > at > org.apache.spark.api.java.JavaSparkContext.hadoopRDD(JavaSparkContext.scala:318) > at > org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generateRDD(SparkPlanGenerator.java:160) > at > org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generate(SparkPlanGenerator.java:88) > at > org.apache.hadoop.hive.ql.exec.spark.SparkClient.execute(SparkClient.java:156) > at > org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.submit(SparkSessionImpl.java:52) > at > org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:77) > at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:161) > at > org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85) > at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1537) > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1304) > at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1116) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:940) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:930) > at > org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:246) > at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:198) > at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:408) > at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:781) > at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675) > at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:601) > at org.apache.hadoop.util.RunJar.main(RunJar.java:212) > Caused by: java.lang.Unsati
[jira] [Commented] (HIVE-7916) Snappy-java error when running hive query on spark [Spark Branch]
[ https://issues.apache.org/jira/browse/HIVE-7916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14175771#comment-14175771 ] Suhas Satish commented on HIVE-7916: Hitting the same problem with snappy 1.1.1.3 as well. Using hive tar ball as of today (fri, oct 17, 2014) with spark.master=local > Snappy-java error when running hive query on spark [Spark Branch] > - > > Key: HIVE-7916 > URL: https://issues.apache.org/jira/browse/HIVE-7916 > Project: Hive > Issue Type: Bug > Components: Spark >Reporter: Xuefu Zhang > Labels: Spark-M1 > > Recently spark branch upgraded its dependency on Spark to 1.1.0-SNAPSHOT. > While the new version addressed some lib conflicts (such as guava), I'm > afraid that it also introduced new problems. The following might be one, when > I set the master URL to be a spark standalone cluster: > {code} > hive> set hive.execution.engine=spark; > hive> set spark.serializer=org.apache.spark.serializer.KryoSerializer; > hive> set spark.master=spark://xzdt:7077; > hive> select name, avg(value) from dec group by name; > 14/08/28 16:41:52 INFO storage.MemoryStore: Block broadcast_0 stored as > values in memory (estimated size 333.0 KB, free 128.0 MB) > java.lang.reflect.InvocationTargetException > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:601) > at org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317) > at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219) > at org.xerial.snappy.Snappy.(Snappy.java:44) > at org.xerial.snappy.SnappyOutputStream.(SnappyOutputStream.java:79) > at > org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:124) > at > org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:207) > at > org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:83) > at > org.apache.spark.broadcast.TorrentBroadcast.(TorrentBroadcast.scala:68) > at > org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:36) > at > org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:29) > at > org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:62) > at org.apache.spark.SparkContext.broadcast(SparkContext.scala:809) > at org.apache.spark.rdd.HadoopRDD.(HadoopRDD.scala:116) > at org.apache.spark.SparkContext.hadoopRDD(SparkContext.scala:541) > at > org.apache.spark.api.java.JavaSparkContext.hadoopRDD(JavaSparkContext.scala:318) > at > org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generateRDD(SparkPlanGenerator.java:160) > at > org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generate(SparkPlanGenerator.java:88) > at > org.apache.hadoop.hive.ql.exec.spark.SparkClient.execute(SparkClient.java:156) > at > org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.submit(SparkSessionImpl.java:52) > at > org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:77) > at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:161) > at > org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85) > at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1537) > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1304) > at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1116) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:940) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:930) > at > org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:246) > at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:198) > at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:408) > at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:781) > at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675) > at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:601) > at org.apache.hadoop.util.RunJar.main(RunJar.java:212) > Caused by: java.lang.UnsatisfiedLinkError: no snappyjava in java.library.path > at java.lang.Class
[jira] [Commented] (HIVE-7916) Snappy-java error when running hive query on spark [Spark Branch]
[ https://issues.apache.org/jira/browse/HIVE-7916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14175712#comment-14175712 ] Suhas Satish commented on HIVE-7916: I also hit the following snappy lib exceptions - I am using snappy snappy-java-1.0.5.jar. Let me try upgrading to snappy 1.1.1.3 2014-10-17 16:18:01,977 ERROR [Executor task launch worker-0]: executor.Executor (Logging.scala:logError(96)) - Exception in task 0.0 in stage 0.0 (TID 0) org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] null at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229) at org.xerial.snappy.Snappy.(Snappy.java:44) at org.xerial.snappy.SnappyOutputStream.(SnappyOutputStream.java:79) at org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:125) at org.apache.spark.storage.BlockManager.wrapForCompression(BlockManager.scala:1083) at org.apache.spark.storage.BlockManager$$anonfun$7.apply(BlockManager.scala:579) at org.apache.spark.storage.BlockManager$$anonfun$7.apply(BlockManager.scala:579) at org.apache.spark.storage.DiskBlockObjectWriter.open(BlockObjectWriter.scala:126) at org.apache.spark.storage.DiskBlockObjectWriter.write(BlockObjectWriter.scala:192) at org.apache.spark.util.collection.ExternalSorter$$anonfun$writePartitionedFile$4$$anonfun$apply$2.apply(ExternalSorter.scala:732) at org.apache.spark.util.collection.ExternalSorter$$anonfun$writePartitionedFile$4$$anonfun$apply$2.apply(ExternalSorter.scala:731) at scala.collection.Iterator$class.foreach(Iterator.scala:727) at org.apache.spark.util.collection.ExternalSorter$IteratorForPartition.foreach(ExternalSorter.scala:789) at org.apache.spark.util.collection.ExternalSorter$$anonfun$writePartitionedFile$4.apply(ExternalSorter.scala:731) at org.apache.spark.util.collection.ExternalSorter$$anonfun$writePartitionedFile$4.apply(ExternalSorter.scala:727) at scala.collection.Iterator$class.foreach(Iterator.scala:727) at scala.collection.AbstractIterator.foreach(Iterator.scala:1157) at org.apache.spark.util.collection.ExternalSorter.writePartitionedFile(ExternalSorter.scala:727) at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:70) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41) at org.apache.spark.scheduler.Task.run(Task.scala:56) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:181) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) -- 2014-10-17 16:18:02,021 INFO [main]: scheduler.DAGScheduler (Logging.scala:logInfo(59)) - Job 0 failed: foreach at SparkPlan.java:80, took 3.389683 s 2014-10-17 16:18:02,021 ERROR [main]: spark.SparkClient (SparkClient.java:execute(166)) - Error executing Spark Plan org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in stage 0.0 (TID 0, localhost): org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] null org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229) org.xerial.snappy.Snappy.(Snappy.java:44) org.xerial.snappy.SnappyOutputStream.(SnappyOutputStream.java:79) org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:125) org.apache.spark.storage.BlockManager.wrapForCompression(BlockManager.scala:1083) org.apache.spark.storage.BlockManager$$anonfun$7.apply(BlockManager.scala:579) org.apache.spark.storage.BlockManager$$anonfun$7.apply(BlockManager.scala:579) org.apache.spark.storage.DiskBlockObjectWriter.open(BlockObjectWriter.scala:126) org.apache.spark.storage.DiskBlockObjectWriter.write(BlockObjectWriter.scala:192) org.apache.spark.util.collection.ExternalSorter$$anonfun$writePartitionedFile$4$$anonfun$apply$2.apply(ExternalSorter.scala:732) org.apache.spark.util.collection.ExternalSorter$$anonfun$writePartitionedFile$4$$anonfun$apply$2.apply(ExternalSorter.scala:731) scala.collection.Iterator$class.foreach(Iterator.scala:727) org.apache.spark.util.collection.ExternalSorter$IteratorForPartition.foreach(ExternalSorter.scala:789) org.apache.spark.util.collection.ExternalSorter$$anonfun$writePartitionedFile$4.apply(ExternalSorter.scala:731) org.apache.spark.util.collection.ExternalSorter$$anonfun$writePartitionedFile$4.apply(ExternalSorter.scala:727) scala.collection.Iterator$class.foreach
[jira] [Commented] (HIVE-7916) Snappy-java error when running hive query on spark [Spark Branch]
[ https://issues.apache.org/jira/browse/HIVE-7916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14119247#comment-14119247 ] Xuefu Zhang commented on HIVE-7916: --- [~lirui], I tried both branch 1.1 and master. Spark 1.1 branch doesn't have the guava lib fix, right? So, using that branch, I hit the guava lib conflict. Using master gives the snappy error. I'm thinking if you can take a look at the guide at https://cwiki.apache.org/confluence/display/Hive/Hive+on+Spark%3A+Getting+Started#HiveonSpark:GettingStarted-HiveonSpark:GettingStarted and update it if necessary. Then, I can try to follow the guide to see if I still get the snappy problem or guava problem. Thanks! > Snappy-java error when running hive query on spark [Spark Branch] > - > > Key: HIVE-7916 > URL: https://issues.apache.org/jira/browse/HIVE-7916 > Project: Hive > Issue Type: Bug > Components: Spark >Reporter: Xuefu Zhang > Labels: Spark-M1 > > Recently spark branch upgraded its dependency on Spark to 1.1.0-SNAPSHOT. > While the new version addressed some lib conflicts (such as guava), I'm > afraid that it also introduced new problems. The following might be one, when > I set the master URL to be a spark standalone cluster: > {code} > hive> set hive.execution.engine=spark; > hive> set spark.serializer=org.apache.spark.serializer.KryoSerializer; > hive> set spark.master=spark://xzdt:7077; > hive> select name, avg(value) from dec group by name; > 14/08/28 16:41:52 INFO storage.MemoryStore: Block broadcast_0 stored as > values in memory (estimated size 333.0 KB, free 128.0 MB) > java.lang.reflect.InvocationTargetException > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:601) > at org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317) > at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219) > at org.xerial.snappy.Snappy.(Snappy.java:44) > at org.xerial.snappy.SnappyOutputStream.(SnappyOutputStream.java:79) > at > org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:124) > at > org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:207) > at > org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:83) > at > org.apache.spark.broadcast.TorrentBroadcast.(TorrentBroadcast.scala:68) > at > org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:36) > at > org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:29) > at > org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:62) > at org.apache.spark.SparkContext.broadcast(SparkContext.scala:809) > at org.apache.spark.rdd.HadoopRDD.(HadoopRDD.scala:116) > at org.apache.spark.SparkContext.hadoopRDD(SparkContext.scala:541) > at > org.apache.spark.api.java.JavaSparkContext.hadoopRDD(JavaSparkContext.scala:318) > at > org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generateRDD(SparkPlanGenerator.java:160) > at > org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generate(SparkPlanGenerator.java:88) > at > org.apache.hadoop.hive.ql.exec.spark.SparkClient.execute(SparkClient.java:156) > at > org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.submit(SparkSessionImpl.java:52) > at > org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:77) > at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:161) > at > org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85) > at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1537) > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1304) > at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1116) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:940) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:930) > at > org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:246) > at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:198) > at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:408) > at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:781) > at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675) > at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.Nat
[jira] [Commented] (HIVE-7916) Snappy-java error when running hive query on spark [Spark Branch]
[ https://issues.apache.org/jira/browse/HIVE-7916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14117911#comment-14117911 ] Rui Li commented on HIVE-7916: -- Hi [~xuefuz], do you use the latest code of spark 1.1 branch? SPARK-2881 is resolved with [#1999|https://github.com/apache/spark/pull/1999] for branch-1.1. You can check the pom file of your spark to verify that, latest code uses snappy-java-1.0.5.3. > Snappy-java error when running hive query on spark [Spark Branch] > - > > Key: HIVE-7916 > URL: https://issues.apache.org/jira/browse/HIVE-7916 > Project: Hive > Issue Type: Bug > Components: Spark >Reporter: Xuefu Zhang > Labels: Spark-M1 > > Recently spark branch upgraded its dependency on Spark to 1.1.0-SNAPSHOT. > While the new version addressed some lib conflicts (such as guava), I'm > afraid that it also introduced new problems. The following might be one, when > I set the master URL to be a spark standalone cluster: > {code} > hive> set hive.execution.engine=spark; > hive> set spark.serializer=org.apache.spark.serializer.KryoSerializer; > hive> set spark.master=spark://xzdt:7077; > hive> select name, avg(value) from dec group by name; > 14/08/28 16:41:52 INFO storage.MemoryStore: Block broadcast_0 stored as > values in memory (estimated size 333.0 KB, free 128.0 MB) > java.lang.reflect.InvocationTargetException > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:601) > at org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317) > at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219) > at org.xerial.snappy.Snappy.(Snappy.java:44) > at org.xerial.snappy.SnappyOutputStream.(SnappyOutputStream.java:79) > at > org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:124) > at > org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:207) > at > org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:83) > at > org.apache.spark.broadcast.TorrentBroadcast.(TorrentBroadcast.scala:68) > at > org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:36) > at > org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:29) > at > org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:62) > at org.apache.spark.SparkContext.broadcast(SparkContext.scala:809) > at org.apache.spark.rdd.HadoopRDD.(HadoopRDD.scala:116) > at org.apache.spark.SparkContext.hadoopRDD(SparkContext.scala:541) > at > org.apache.spark.api.java.JavaSparkContext.hadoopRDD(JavaSparkContext.scala:318) > at > org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generateRDD(SparkPlanGenerator.java:160) > at > org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generate(SparkPlanGenerator.java:88) > at > org.apache.hadoop.hive.ql.exec.spark.SparkClient.execute(SparkClient.java:156) > at > org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.submit(SparkSessionImpl.java:52) > at > org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:77) > at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:161) > at > org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85) > at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1537) > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1304) > at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1116) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:940) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:930) > at > org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:246) > at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:198) > at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:408) > at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:781) > at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675) > at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:601) > at org.apache.hadoop.util.RunJar.main(RunJar.java
[jira] [Commented] (HIVE-7916) Snappy-java error when running hive query on spark [Spark Branch]
[ https://issues.apache.org/jira/browse/HIVE-7916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14117872#comment-14117872 ] Xuefu Zhang commented on HIVE-7916: --- [~lirui] Thanks for looking into this. I had noticed the version difference, and thought for a moment it was the cause. However, I don't see any /tmp/snappy*.so file in my machine. I also built spark-assembly myself, but with slightly different params: {code} mvn -Dhadoop.version=2.3.0-cdh5.0.1 -Phadoop-2.3 -DskipTests clean install. {code} Aslo, I'm not sure if my machine has ever run snappy before. Is there anything that needs to be done in order for it to work? I can try again to reproduce the problem. > Snappy-java error when running hive query on spark [Spark Branch] > - > > Key: HIVE-7916 > URL: https://issues.apache.org/jira/browse/HIVE-7916 > Project: Hive > Issue Type: Bug > Components: Spark >Reporter: Xuefu Zhang > Labels: Spark-M1 > > Recently spark branch upgraded its dependency on Spark to 1.1.0-SNAPSHOT. > While the new version addressed some lib conflicts (such as guava), I'm > afraid that it also introduced new problems. The following might be one, when > I set the master URL to be a spark standalone cluster: > {code} > hive> set hive.execution.engine=spark; > hive> set spark.serializer=org.apache.spark.serializer.KryoSerializer; > hive> set spark.master=spark://xzdt:7077; > hive> select name, avg(value) from dec group by name; > 14/08/28 16:41:52 INFO storage.MemoryStore: Block broadcast_0 stored as > values in memory (estimated size 333.0 KB, free 128.0 MB) > java.lang.reflect.InvocationTargetException > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:601) > at org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317) > at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219) > at org.xerial.snappy.Snappy.(Snappy.java:44) > at org.xerial.snappy.SnappyOutputStream.(SnappyOutputStream.java:79) > at > org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:124) > at > org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:207) > at > org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:83) > at > org.apache.spark.broadcast.TorrentBroadcast.(TorrentBroadcast.scala:68) > at > org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:36) > at > org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:29) > at > org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:62) > at org.apache.spark.SparkContext.broadcast(SparkContext.scala:809) > at org.apache.spark.rdd.HadoopRDD.(HadoopRDD.scala:116) > at org.apache.spark.SparkContext.hadoopRDD(SparkContext.scala:541) > at > org.apache.spark.api.java.JavaSparkContext.hadoopRDD(JavaSparkContext.scala:318) > at > org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generateRDD(SparkPlanGenerator.java:160) > at > org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generate(SparkPlanGenerator.java:88) > at > org.apache.hadoop.hive.ql.exec.spark.SparkClient.execute(SparkClient.java:156) > at > org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.submit(SparkSessionImpl.java:52) > at > org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:77) > at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:161) > at > org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85) > at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1537) > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1304) > at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1116) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:940) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:930) > at > org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:246) > at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:198) > at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:408) > at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:781) > at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675) > at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.N
[jira] [Commented] (HIVE-7916) Snappy-java error when running hive query on spark [Spark Branch]
[ https://issues.apache.org/jira/browse/HIVE-7916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14117057#comment-14117057 ] Rui Li commented on HIVE-7916: -- I noted this may be related to SPARK-2881. Snappy-java is bumped to 1.0.5.3 in the 1.1 branch and to 1.1.1.3 in the master branch. Hadoop-2.4.0 seems to use snappy-java-1.0.4.1. While the snappy-java version is different, I don't see any conflicts on my side. [~xuefuz], I found the following in the description of SPARK-2881: {quote} The issue was that someone else had run with snappy and it created /tmp/snappy-*.so but it had restrictive permissions so I was not able to use it or remove it. This caused my spark job to not start. {quote} Could you check if this is the case in your environment? > Snappy-java error when running hive query on spark [Spark Branch] > - > > Key: HIVE-7916 > URL: https://issues.apache.org/jira/browse/HIVE-7916 > Project: Hive > Issue Type: Bug > Components: Spark >Reporter: Xuefu Zhang > Labels: Spark-M1 > > Recently spark branch upgraded its dependency on Spark to 1.1.0-SNAPSHOT. > While the new version addressed some lib conflicts (such as guava), I'm > afraid that it also introduced new problems. The following might be one, when > I set the master URL to be a spark standalone cluster: > {code} > hive> set hive.execution.engine=spark; > hive> set spark.serializer=org.apache.spark.serializer.KryoSerializer; > hive> set spark.master=spark://xzdt:7077; > hive> select name, avg(value) from dec group by name; > 14/08/28 16:41:52 INFO storage.MemoryStore: Block broadcast_0 stored as > values in memory (estimated size 333.0 KB, free 128.0 MB) > java.lang.reflect.InvocationTargetException > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:601) > at org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317) > at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219) > at org.xerial.snappy.Snappy.(Snappy.java:44) > at org.xerial.snappy.SnappyOutputStream.(SnappyOutputStream.java:79) > at > org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:124) > at > org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:207) > at > org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:83) > at > org.apache.spark.broadcast.TorrentBroadcast.(TorrentBroadcast.scala:68) > at > org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:36) > at > org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:29) > at > org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:62) > at org.apache.spark.SparkContext.broadcast(SparkContext.scala:809) > at org.apache.spark.rdd.HadoopRDD.(HadoopRDD.scala:116) > at org.apache.spark.SparkContext.hadoopRDD(SparkContext.scala:541) > at > org.apache.spark.api.java.JavaSparkContext.hadoopRDD(JavaSparkContext.scala:318) > at > org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generateRDD(SparkPlanGenerator.java:160) > at > org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generate(SparkPlanGenerator.java:88) > at > org.apache.hadoop.hive.ql.exec.spark.SparkClient.execute(SparkClient.java:156) > at > org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.submit(SparkSessionImpl.java:52) > at > org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:77) > at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:161) > at > org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85) > at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1537) > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1304) > at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1116) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:940) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:930) > at > org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:246) > at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:198) > at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:408) > at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:781) > at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675) > at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614) > at sun.reflec
[jira] [Commented] (HIVE-7916) Snappy-java error when running hive query on spark [Spark Branch]
[ https://issues.apache.org/jira/browse/HIVE-7916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14117050#comment-14117050 ] Rui Li commented on HIVE-7916: -- Hi [~xuefuz], I tried on my cluster but cannot reproduce the problem. I removed the spark jars from local maven repo before building hive, so that the jars are downloaded from the AWS server we maintain. After hive is built, I linked the spark-assembly jar to {{lib}} of the hive home directory. The spark-assembly jar is built with {{mvn -Pyarn -Phadoop-2.4 -DskipTests clean package}} of the spark 1.1 branch. Could you provide more info about your environment, e.g. the spark jars you used or if the table is snappy compressed? > Snappy-java error when running hive query on spark [Spark Branch] > - > > Key: HIVE-7916 > URL: https://issues.apache.org/jira/browse/HIVE-7916 > Project: Hive > Issue Type: Bug > Components: Spark >Reporter: Xuefu Zhang > Labels: Spark-M1 > > Recently spark branch upgraded its dependency on Spark to 1.1.0-SNAPSHOT. > While the new version addressed some lib conflicts (such as guava), I'm > afraid that it also introduced new problems. The following might be one, when > I set the master URL to be a spark standalone cluster: > {code} > hive> set hive.execution.engine=spark; > hive> set spark.serializer=org.apache.spark.serializer.KryoSerializer; > hive> set spark.master=spark://xzdt:7077; > hive> select name, avg(value) from dec group by name; > 14/08/28 16:41:52 INFO storage.MemoryStore: Block broadcast_0 stored as > values in memory (estimated size 333.0 KB, free 128.0 MB) > java.lang.reflect.InvocationTargetException > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:601) > at org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317) > at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219) > at org.xerial.snappy.Snappy.(Snappy.java:44) > at org.xerial.snappy.SnappyOutputStream.(SnappyOutputStream.java:79) > at > org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:124) > at > org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:207) > at > org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:83) > at > org.apache.spark.broadcast.TorrentBroadcast.(TorrentBroadcast.scala:68) > at > org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:36) > at > org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:29) > at > org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:62) > at org.apache.spark.SparkContext.broadcast(SparkContext.scala:809) > at org.apache.spark.rdd.HadoopRDD.(HadoopRDD.scala:116) > at org.apache.spark.SparkContext.hadoopRDD(SparkContext.scala:541) > at > org.apache.spark.api.java.JavaSparkContext.hadoopRDD(JavaSparkContext.scala:318) > at > org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generateRDD(SparkPlanGenerator.java:160) > at > org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generate(SparkPlanGenerator.java:88) > at > org.apache.hadoop.hive.ql.exec.spark.SparkClient.execute(SparkClient.java:156) > at > org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.submit(SparkSessionImpl.java:52) > at > org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:77) > at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:161) > at > org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85) > at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1537) > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1304) > at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1116) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:940) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:930) > at > org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:246) > at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:198) > at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:408) > at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:781) > at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675) > at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.re