[ 
https://issues.apache.org/jira/browse/SPARK-4908?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14260490#comment-14260490
 ] 

Michael Armbrust commented on SPARK-4908:
-----------------------------------------

You don't even need the JDBC server to cause the problem.  This seems to fail 
as well:

{code}
(1 to 100).par.map { _ => 
  sql("USE default")
  sql("SHOW TABLES")
}
{code}

> Spark SQL built for Hive 13 fails under concurrent metadata queries
> -------------------------------------------------------------------
>
>                 Key: SPARK-4908
>                 URL: https://issues.apache.org/jira/browse/SPARK-4908
>             Project: Spark
>          Issue Type: Bug
>          Components: SQL
>            Reporter: David Ross
>            Priority: Critical
>
> We are trunk: {{1.3.0-SNAPSHOT}}, as of this commit: 
> https://github.com/apache/spark/commit/3d0c37b8118f6057a663f959321a79b8061132b6
> We are using Spark built for Hive 13, using this option:
> {{-Phive-0.13.1}}
> In single-threaded mode, normal operations look fine. However, under 
> concurrency, with at least 2 concurrent connections, metadata queries fail.
> For example, {{USE some_db}}, {{SHOW TABLES}}, and the implicit {{USE}} 
> statement when you pass a default schema in the JDBC URL, all fail.
> {{SELECT}} queries like {{SELECT * FROM some_table}} do not have this issue.
> Here is some example code:
> {code}
> object main extends App {
>   import java.sql._
>   import scala.concurrent._
>   import scala.concurrent.duration._
>   import scala.concurrent.ExecutionContext.Implicits.global
>   Class.forName("org.apache.hive.jdbc.HiveDriver")
>   val host = "localhost" // update this
>   val url = s"jdbc:hive2://${host}:10511/some_db" // update this
>   val future = Future.traverse(1 to 3) { i =>
>     Future {
>       println("Starting: " + i)
>       try {
>         val conn = DriverManager.getConnection(url)
>       } catch {
>         case e: Throwable => e.printStackTrace()
>         println("Failed: " + i)
>       }
>       println("Finishing: " + i)
>     }
>   }
>   Await.result(future, 2.minutes)
>   println("done!")
> }
> {code}
> Here is the output:
> {code}
> Starting: 1
> Starting: 3
> Starting: 2
> java.sql.SQLException: 
> org.apache.spark.sql.execution.QueryExecutionException: FAILED: Operation 
> cancelled
>       at org.apache.hive.jdbc.Utils.verifySuccess(Utils.java:121)
>       at org.apache.hive.jdbc.Utils.verifySuccessWithInfo(Utils.java:109)
>       at org.apache.hive.jdbc.HiveStatement.execute(HiveStatement.java:231)
>       at 
> org.apache.hive.jdbc.HiveConnection.configureConnection(HiveConnection.java:451)
>       at org.apache.hive.jdbc.HiveConnection.<init>(HiveConnection.java:195)
>       at org.apache.hive.jdbc.HiveDriver.connect(HiveDriver.java:105)
>       at java.sql.DriverManager.getConnection(DriverManager.java:664)
>       at java.sql.DriverManager.getConnection(DriverManager.java:270)
>       at 
> com.atscale.engine.connection.pool.main$$anonfun$30$$anonfun$apply$2.apply$mcV$sp(ConnectionManager.scala:896)
>       at 
> com.atscale.engine.connection.pool.main$$anonfun$30$$anonfun$apply$2.apply(ConnectionManager.scala:893)
>       at 
> com.atscale.engine.connection.pool.main$$anonfun$30$$anonfun$apply$2.apply(ConnectionManager.scala:893)
>       at 
> scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24)
>       at 
> scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24)
>       at 
> scala.concurrent.impl.ExecutionContextImpl$AdaptedForkJoinTask.exec(ExecutionContextImpl.scala:121)
>       at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>       at 
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
>       at 
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>       at 
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> Failed: 3
> Finishing: 3
> java.sql.SQLException: 
> org.apache.spark.sql.execution.QueryExecutionException: FAILED: Operation 
> cancelled
>       at org.apache.hive.jdbc.Utils.verifySuccess(Utils.java:121)
>       at org.apache.hive.jdbc.Utils.verifySuccessWithInfo(Utils.java:109)
>       at org.apache.hive.jdbc.HiveStatement.execute(HiveStatement.java:231)
>       at 
> org.apache.hive.jdbc.HiveConnection.configureConnection(HiveConnection.java:451)
>       at org.apache.hive.jdbc.HiveConnection.<init>(HiveConnection.java:195)
>       at org.apache.hive.jdbc.HiveDriver.connect(HiveDriver.java:105)
>       at java.sql.DriverManager.getConnection(DriverManager.java:664)
>       at java.sql.DriverManager.getConnection(DriverManager.java:270)
>       at 
> com.atscale.engine.connection.pool.main$$anonfun$30$$anonfun$apply$2.apply$mcV$sp(ConnectionManager.scala:896)
>       at 
> com.atscale.engine.connection.pool.main$$anonfun$30$$anonfun$apply$2.apply(ConnectionManager.scala:893)
>       at 
> com.atscale.engine.connection.pool.main$$anonfun$30$$anonfun$apply$2.apply(ConnectionManager.scala:893)
>       at 
> scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24)
>       at 
> scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24)
>       at 
> scala.concurrent.impl.ExecutionContextImpl$AdaptedForkJoinTask.exec(ExecutionContextImpl.scala:121)
>       at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>       at 
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
>       at 
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>       at 
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> Failed: 2
> Finishing: 2
> Finishing: 1
> done!
> {code}
> Here are the errors from Spark Logs:
> {code}
> 14/12/19 21:44:55 INFO thrift.ThriftCLIService: Client protocol version: 
> HIVE_CLI_SERVICE_PROTOCOL_V6
> 14/12/19 21:44:55 INFO session.SessionState: No Tez session required at this 
> point. hive.execution.engine=mr.
> 14/12/19 21:44:55 INFO session.SessionState: No Tez session required at this 
> point. hive.execution.engine=mr.
> 14/12/19 21:44:55 INFO thriftserver.SparkExecuteStatementOperation: Running 
> query 'use as_adventure'
> 14/12/19 21:44:55 INFO parse.ParseDriver: Parsing command: use as_adventure
> 14/12/19 21:44:55 INFO parse.ParseDriver: Parse Completed
> 14/12/19 21:44:55 INFO session.SessionState: No Tez session required at this 
> point. hive.execution.engine=mr.
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=Driver.run 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=TimeToSubmit 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO ql.Driver: Concurrency mode is disabled, not creating 
> a lock manager
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=compile 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=parse 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO parse.ParseDriver: Parsing command: use as_adventure
> 14/12/19 21:44:55 INFO parse.ParseDriver: Parse Completed
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=parse 
> start=1419025495084 end=1419025495084 duration=0 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=semanticAnalyze 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO ql.Driver: Semantic Analysis Completed
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=semanticAnalyze 
> start=1419025495084 end=1419025495084 duration=0 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO ql.Driver: Returning Hive schema: 
> Schema(fieldSchemas:null, properties:null)
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=compile 
> start=1419025495084 end=1419025495085 duration=1 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=Driver.execute 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO ql.Driver: Starting command: use as_adventure
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=TimeToSubmit 
> start=1419025495084 end=1419025495085 duration=1 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=runTasks 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=task.DDL.Stage-0 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=runTasks 
> start=1419025495085 end=1419025495098 duration=13 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=Driver.execute 
> start=1419025495085 end=1419025495098 duration=13 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO ql.Driver: OK
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=releaseLocks 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=releaseLocks 
> start=1419025495098 end=1419025495098 duration=0 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=Driver.run 
> start=1419025495084 end=1419025495098 duration=14 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO thrift.ThriftCLIService: Client protocol version: 
> HIVE_CLI_SERVICE_PROTOCOL_V6
> 14/12/19 21:44:55 INFO thrift.ThriftCLIService: Client protocol version: 
> HIVE_CLI_SERVICE_PROTOCOL_V6
> 14/12/19 21:44:55 INFO session.SessionState: No Tez session required at this 
> point. hive.execution.engine=mr.
> 14/12/19 21:44:55 INFO session.SessionState: No Tez session required at this 
> point. hive.execution.engine=mr.
> 14/12/19 21:44:55 INFO session.SessionState: No Tez session required at this 
> point. hive.execution.engine=mr.
> 14/12/19 21:44:55 INFO session.SessionState: No Tez session required at this 
> point. hive.execution.engine=mr.
> 14/12/19 21:44:55 INFO thriftserver.SparkExecuteStatementOperation: Running 
> query 'use as_adventure'
> 14/12/19 21:44:55 INFO thriftserver.SparkExecuteStatementOperation: Running 
> query 'use as_adventure'
> 14/12/19 21:44:55 INFO thriftserver.SparkExecuteStatementOperation: Result 
> Schema: List(result#274)
> 14/12/19 21:44:55 INFO parse.ParseDriver: Parsing command: use as_adventure
> 14/12/19 21:44:55 INFO parse.ParseDriver: Parse Completed
> 14/12/19 21:44:55 INFO session.SessionState: No Tez session required at this 
> point. hive.execution.engine=mr.
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=Driver.run 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=TimeToSubmit 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO ql.Driver: Concurrency mode is disabled, not creating 
> a lock manager
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=compile 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=parse 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO parse.ParseDriver: Parsing command: use as_adventure
> 14/12/19 21:44:55 INFO parse.ParseDriver: Parse Completed
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=parse 
> start=1419025495165 end=1419025495165 duration=0 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=semanticAnalyze 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO ql.Driver: Semantic Analysis Completed
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=semanticAnalyze 
> start=1419025495165 end=1419025495166 duration=1 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO ql.Driver: Returning Hive schema: 
> Schema(fieldSchemas:null, properties:null)
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=compile 
> start=1419025495165 end=1419025495166 duration=1 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=Driver.execute 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO ql.Driver: Starting command: use as_adventure
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=TimeToSubmit 
> start=1419025495165 end=1419025495166 duration=1 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=runTasks 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=task.DDL.Stage-0 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO parse.ParseDriver: Parsing command: use as_adventure
> 14/12/19 21:44:55 INFO parse.ParseDriver: Parse Completed
> 14/12/19 21:44:55 INFO session.SessionState: No Tez session required at this 
> point. hive.execution.engine=mr.
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=Driver.run 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=TimeToSubmit 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO ql.Driver: Concurrency mode is disabled, not creating 
> a lock manager
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=compile 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 WARN ql.Driver: Shutting down task : Stage-0:DDL
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=parse 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO parse.ParseDriver: Parsing command: use as_adventure
> 14/12/19 21:44:55 INFO parse.ParseDriver: Parse Completed
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=parse 
> start=1419025495173 end=1419025495174 duration=1 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=semanticAnalyze 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO ql.Driver: Semantic Analysis Completed
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=semanticAnalyze 
> start=1419025495174 end=1419025495174 duration=0 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO ql.Driver: Returning Hive schema: 
> Schema(fieldSchemas:null, properties:null)
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=compile 
> start=1419025495172 end=1419025495177 duration=5 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=Driver.execute 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO ql.Driver: Starting command: use as_adventure
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=TimeToSubmit 
> start=1419025495172 end=1419025495177 duration=5 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=runTasks 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=task.DDL.Stage-0 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=runTasks 
> start=1419025495167 end=1419025495188 duration=21 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 ERROR ql.Driver: FAILED: Operation cancelled
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=Driver.execute 
> start=1419025495166 end=1419025495189 duration=23 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=releaseLocks 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=releaseLocks 
> start=1419025495189 end=1419025495189 duration=0 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 WARN ql.Driver: Shutting down task : Stage-0:DDL
> 14/12/19 21:44:55 ERROR hive.HiveContext:
> ======================
> HIVE FAILURE OUTPUT
> ======================
> RDDLike.scala:58)
>       at org.apache.spark.sql.SchemaRDD.<init>(SchemaRDD.scala:108)
>       at org.apache.spark.sql.hive.HiveContext.sql(HiveContext.scala:94)
>       at 
> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.run(Shim13.scala:161)
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:231)
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync(HiveSessionImpl.java:218)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:483)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:79)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:37)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:64)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:422)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>       at 
> org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:493)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:60)
>       at com.sun.proxy.$Proxy18.executeStatementAsync(Unknown Source)
>       at 
> org.apache.hive.service.cli.CLIService.executeStatementAsync(CLIService.java:233)
>       at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:344)
>       at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1313)
>       at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1298)
>       at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>       at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>       at 
> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:55)
>       at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>       at java.lang.Thread.run(Thread.java:745)
> FAILED: Hive Internal Error: 
> org.apache.hadoop.hive.ql.metadata.HiveException(FAILED: Operation cancelled)
> org.apache.hadoop.hive.ql.metadata.HiveException: FAILED: Operation cancelled
>       at 
> org.apache.hadoop.hive.ql.DriverContext.checkShutdown(DriverContext.java:125)
>       at 
> org.apache.hadoop.hive.ql.DriverContext.launching(DriverContext.java:91)
>       at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1497)
>       at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1270)
>       at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1088)
>       at org.apache.hadoop.hive.ql.Driver.run(Driver.java:911)
>       at org.apache.hadoop.hive.ql.Driver.run(Driver.java:901)
>       at org.apache.spark.sql.hive.HiveContext.runHive(HiveContext.scala:305)
>       at 
> org.apache.spark.sql.hive.HiveContext.runSqlHive(HiveContext.scala:276)
>       at 
> org.apache.spark.sql.hive.execution.NativeCommand.sideEffectResult$lzycompute(NativeCommand.scala:35)
>       at 
> org.apache.spark.sql.hive.execution.NativeCommand.sideEffectResult(NativeCommand.scala:35)
>       at 
> org.apache.spark.sql.execution.Command$class.execute(commands.scala:46)
>       at 
> org.apache.spark.sql.hive.execution.NativeCommand.execute(NativeCommand.scala:30)
>       at 
> org.apache.spark.sql.SQLContext$QueryExecution.toRdd$lzycompute(SQLContext.scala:425)
>       at 
> org.apache.spark.sql.SQLContext$QueryExecution.toRdd(SQLContext.scala:425)
>       at 
> org.apache.spark.sql.SchemaRDDLike$class.$init$(SchemaRDDLike.scala:58)
>       at org.apache.spark.sql.SchemaRDD.<init>(SchemaRDD.scala:108)
>       at org.apache.spark.sql.hive.HiveContext.sql(HiveContext.scala:94)
>       at 
> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.run(Shim13.scala:161)
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:231)
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync(HiveSessionImpl.java:218)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:483)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:79)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:37)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:64)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:422)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>       at 
> org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:493)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:60)
>       at com.sun.proxy.$Proxy18.executeStatementAsync(Unknown Source)
>       at 
> org.apache.hive.service.cli.CLIService.executeStatementAsync(CLIService.java:233)
>       at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:344)
>       at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1313)
>       at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1298)
>       at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>       at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>       at 
> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:55)
>       at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>       at java.lang.Thread.run(Thread.java:745)
> OK
> OK
> OK
> OK
> FAILED: Operation cancelled
> FAILED: Operation cancelled
> OK
> OK
> FAILED: Operation cancelled
> FAILED: Hive Internal Error: java.lang.NullPointerException(null)
> java.lang.NullPointerException
>       at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1194)
>       at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1088)
>       at org.apache.hadoop.hive.ql.Driver.run(Driver.java:911)
>       at org.apache.hadoop.hive.ql.Driver.run(Driver.java:901)
>       at org.apache.spark.sql.hive.HiveContext.runHive(HiveContext.scala:305)
>       at 
> org.apache.spark.sql.hive.HiveContext.runSqlHive(HiveContext.scala:276)
>       at 
> org.apache.spark.sql.hive.execution.NativeCommand.sideEffectResult$lzycompute(NativeCommand.scala:35)
>       at 
> org.apache.spark.sql.hive.execution.NativeCommand.sideEffectResult(NativeCommand.scala:35)
>       at 
> org.apache.spark.sql.execution.Command$class.execute(commands.scala:46)
>       at 
> org.apache.spark.sql.hive.execution.NativeCommand.execute(NativeCommand.scala:30)
>       at 
> org.apache.spark.sql.SQLContext$QueryExecution.toRdd$lzycompute(SQLContext.scala:425)
>       at 
> org.apache.spark.sql.SQLContext$QueryExecution.toRdd(SQLContext.scala:425)
>       at 
> org.apache.spark.sql.SchemaRDDLike$class.$init$(SchemaRDDLike.scala:58)
>       at org.apache.spark.sql.SchemaRDD.<init>(SchemaRDD.scala:108)
>       at org.apache.spark.sql.hive.HiveContext.sql(HiveContext.scala:94)
>       at 
> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.run(Shim13.scala:161)
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:231)
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync(HiveSessionImpl.java:218)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:483)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:79)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:37)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:64)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:422)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>       at 
> org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:493)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:60)
>       at com.sun.proxy.$Proxy18.executeStatementAsync(Unknown Source)
>       at 
> org.apache.hive.service.cli.CLIService.executeStatementAsync(CLIService.java:233)
>       at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:344)
>       at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1313)
>       at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1298)
>       at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>       at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>       at 
> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:55)
>       at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>       at java.lang.Thread.run(Thread.java:745)
> FAILED: Operation cancelled
> OK
> FAILED: Operation cancelled
> FAILED: Operation cancelled
> OK
> FAILED: Operation cancelled
> OK
> OK
> FAILED: Operation cancelled
> OK
> OK
> FAILED: Operation cancelled
> OK
> OK
> FAILED: Operation cancelled
> FAILED: Operation cancelled
> OK
> FAILED: Operation cancelled
> FAILED: Operation cancelled
> FAILED: Operation cancelled
> OK
> FAILED: Operation cancelled
> FAILED: Operation cancelled
> OK
> FAILED: Operation cancelled
> ======================
> END HIVE FAILURE OUTPUT
> ======================
> 14/12/19 21:44:55 ERROR thriftserver.SparkExecuteStatementOperation: Error 
> executing query:
> org.apache.spark.sql.execution.QueryExecutionException: FAILED: Operation 
> cancelled
>       at org.apache.spark.sql.hive.HiveContext.runHive(HiveContext.scala:309)
>       at 
> org.apache.spark.sql.hive.HiveContext.runSqlHive(HiveContext.scala:276)
>       at 
> org.apache.spark.sql.hive.execution.NativeCommand.sideEffectResult$lzycompute(NativeCommand.scala:35)
>       at 
> org.apache.spark.sql.hive.execution.NativeCommand.sideEffectResult(NativeCommand.scala:35)
>       at 
> org.apache.spark.sql.execution.Command$class.execute(commands.scala:46)
>       at 
> org.apache.spark.sql.hive.execution.NativeCommand.execute(NativeCommand.scala:30)
>       at 
> org.apache.spark.sql.SQLContext$QueryExecution.toRdd$lzycompute(SQLContext.scala:425)
>       at 
> org.apache.spark.sql.SQLContext$QueryExecution.toRdd(SQLContext.scala:425)
>       at 
> org.apache.spark.sql.SchemaRDDLike$class.$init$(SchemaRDDLike.scala:58)
>       at org.apache.spark.sql.SchemaRDD.<init>(SchemaRDD.scala:108)
>       at org.apache.spark.sql.hive.HiveContext.sql(HiveContext.scala:94)
>       at 
> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.run(Shim13.scala:161)
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:231)
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync(HiveSessionImpl.java:218)
>       at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:483)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:79)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:37)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:64)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:422)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>       at 
> org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:493)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:60)
>       at com.sun.proxy.$Proxy18.executeStatementAsync(Unknown Source)
>       at 
> org.apache.hive.service.cli.CLIService.executeStatementAsync(CLIService.java:233)
>       at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:344)
>       at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1313)
>       at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1298)
>       at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>       at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>       at 
> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:55)
>       at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>       at java.lang.Thread.run(Thread.java:745)
> 14/12/19 21:44:55 WARN thrift.ThriftCLIService: Error executing statement:
> org.apache.hive.service.cli.HiveSQLException: 
> org.apache.spark.sql.execution.QueryExecutionException: FAILED: Operation 
> cancelled
>       at 
> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.run(Shim13.scala:192)
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:231)
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync(HiveSessionImpl.java:218)
>       at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:483)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:79)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:37)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:64)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:422)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>       at 
> org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:493)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:60)
>       at com.sun.proxy.$Proxy18.executeStatementAsync(Unknown Source)
>       at 
> org.apache.hive.service.cli.CLIService.executeStatementAsync(CLIService.java:233)
>       at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:344)
>       at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1313)
>       at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1298)
>       at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>       at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>       at 
> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:55)
>       at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>       at java.lang.Thread.run(Thread.java:745)
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=runTasks 
> start=1419025495177 end=1419025495197 duration=20 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 ERROR ql.Driver: FAILED: Operation cancelled
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=Driver.execute 
> start=1419025495177 end=1419025495200 duration=23 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: <PERFLOG method=releaseLocks 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 INFO log.PerfLogger: </PERFLOG method=releaseLocks 
> start=1419025495200 end=1419025495200 duration=0 
> from=org.apache.hadoop.hive.ql.Driver>
> 14/12/19 21:44:55 ERROR hive.HiveContext:
> ======================
> HIVE FAILURE OUTPUT
> ======================
> ache.spark.sql.SchemaRDD.<init>(SchemaRDD.scala:108)
>       at org.apache.spark.sql.hive.HiveContext.sql(HiveContext.scala:94)
>       at 
> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.run(Shim13.scala:161)
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:231)
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync(HiveSessionImpl.java:218)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:483)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:79)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:37)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:64)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:422)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>       at 
> org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:493)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:60)
>       at com.sun.proxy.$Proxy18.executeStatementAsync(Unknown Source)
>       at 
> org.apache.hive.service.cli.CLIService.executeStatementAsync(CLIService.java:233)
>       at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:344)
>       at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1313)
>       at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1298)
>       at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>       at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>       at 
> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:55)
>       at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>       at java.lang.Thread.run(Thread.java:745)
> FAILED: Hive Internal Error: 
> org.apache.hadoop.hive.ql.metadata.HiveException(FAILED: Operation cancelled)
> org.apache.hadoop.hive.ql.metadata.HiveException: FAILED: Operation cancelled
>       at 
> org.apache.hadoop.hive.ql.DriverContext.checkShutdown(DriverContext.java:125)
>       at 
> org.apache.hadoop.hive.ql.DriverContext.launching(DriverContext.java:91)
>       at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1497)
>       at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1270)
>       at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1088)
>       at org.apache.hadoop.hive.ql.Driver.run(Driver.java:911)
>       at org.apache.hadoop.hive.ql.Driver.run(Driver.java:901)
>       at org.apache.spark.sql.hive.HiveContext.runHive(HiveContext.scala:305)
>       at 
> org.apache.spark.sql.hive.HiveContext.runSqlHive(HiveContext.scala:276)
>       at 
> org.apache.spark.sql.hive.execution.NativeCommand.sideEffectResult$lzycompute(NativeCommand.scala:35)
>       at 
> org.apache.spark.sql.hive.execution.NativeCommand.sideEffectResult(NativeCommand.scala:35)
>       at 
> org.apache.spark.sql.execution.Command$class.execute(commands.scala:46)
>       at 
> org.apache.spark.sql.hive.execution.NativeCommand.execute(NativeCommand.scala:30)
>       at 
> org.apache.spark.sql.SQLContext$QueryExecution.toRdd$lzycompute(SQLContext.scala:425)
>       at 
> org.apache.spark.sql.SQLContext$QueryExecution.toRdd(SQLContext.scala:425)
>       at 
> org.apache.spark.sql.SchemaRDDLike$class.$init$(SchemaRDDLike.scala:58)
>       at org.apache.spark.sql.SchemaRDD.<init>(SchemaRDD.scala:108)
>       at org.apache.spark.sql.hive.HiveContext.sql(HiveContext.scala:94)
>       at 
> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.run(Shim13.scala:161)
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:231)
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync(HiveSessionImpl.java:218)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:483)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:79)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:37)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:64)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:422)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>       at 
> org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:493)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:60)
>       at com.sun.proxy.$Proxy18.executeStatementAsync(Unknown Source)
>       at 
> org.apache.hive.service.cli.CLIService.executeStatementAsync(CLIService.java:233)
>       at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:344)
>       at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1313)
>       at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1298)
>       at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>       at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>       at 
> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:55)
>       at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>       at java.lang.Thread.run(Thread.java:745)
> OK
> OK
> OK
> OK
> FAILED: Operation cancelled
> FAILED: Operation cancelled
> OK
> OK
> FAILED: Operation cancelled
> FAILED: Hive Internal Error: java.lang.NullPointerException(null)
> java.lang.NullPointerException
>       at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1194)
>       at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1088)
>       at org.apache.hadoop.hive.ql.Driver.run(Driver.java:911)
>       at org.apache.hadoop.hive.ql.Driver.run(Driver.java:901)
>       at org.apache.spark.sql.hive.HiveContext.runHive(HiveContext.scala:305)
>       at 
> org.apache.spark.sql.hive.HiveContext.runSqlHive(HiveContext.scala:276)
>       at 
> org.apache.spark.sql.hive.execution.NativeCommand.sideEffectResult$lzycompute(NativeCommand.scala:35)
>       at 
> org.apache.spark.sql.hive.execution.NativeCommand.sideEffectResult(NativeCommand.scala:35)
>       at 
> org.apache.spark.sql.execution.Command$class.execute(commands.scala:46)
>       at 
> org.apache.spark.sql.hive.execution.NativeCommand.execute(NativeCommand.scala:30)
>       at 
> org.apache.spark.sql.SQLContext$QueryExecution.toRdd$lzycompute(SQLContext.scala:425)
>       at 
> org.apache.spark.sql.SQLContext$QueryExecution.toRdd(SQLContext.scala:425)
>       at 
> org.apache.spark.sql.SchemaRDDLike$class.$init$(SchemaRDDLike.scala:58)
>       at org.apache.spark.sql.SchemaRDD.<init>(SchemaRDD.scala:108)
>       at org.apache.spark.sql.hive.HiveContext.sql(HiveContext.scala:94)
>       at 
> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.run(Shim13.scala:161)
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:231)
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync(HiveSessionImpl.java:218)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:483)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:79)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:37)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:64)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:422)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>       at 
> org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:493)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:60)
>       at com.sun.proxy.$Proxy18.executeStatementAsync(Unknown Source)
>       at 
> org.apache.hive.service.cli.CLIService.executeStatementAsync(CLIService.java:233)
>       at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:344)
>       at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1313)
>       at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1298)
>       at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>       at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>       at 
> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:55)
>       at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>       at java.lang.Thread.run(Thread.java:745)
> FAILED: Operation cancelled
> OK
> FAILED: Operation cancelled
> FAILED: Operation cancelled
> OK
> FAILED: Operation cancelled
> OK
> OK
> FAILED: Operation cancelled
> OK
> OK
> FAILED: Operation cancelled
> OK
> OK
> FAILED: Operation cancelled
> FAILED: Operation cancelled
> OK
> FAILED: Operation cancelled
> FAILED: Operation cancelled
> FAILED: Operation cancelled
> OK
> FAILED: Operation cancelled
> FAILED: Operation cancelled
> OK
> FAILED: Operation cancelled
> FAILED: Operation cancelled
> ======================
> END HIVE FAILURE OUTPUT
> ======================
> 14/12/19 21:44:55 ERROR thriftserver.SparkExecuteStatementOperation: Error 
> executing query:
> org.apache.spark.sql.execution.QueryExecutionException: FAILED: Operation 
> cancelled
>       at org.apache.spark.sql.hive.HiveContext.runHive(HiveContext.scala:309)
>       at 
> org.apache.spark.sql.hive.HiveContext.runSqlHive(HiveContext.scala:276)
>       at 
> org.apache.spark.sql.hive.execution.NativeCommand.sideEffectResult$lzycompute(NativeCommand.scala:35)
>       at 
> org.apache.spark.sql.hive.execution.NativeCommand.sideEffectResult(NativeCommand.scala:35)
>       at 
> org.apache.spark.sql.execution.Command$class.execute(commands.scala:46)
>       at 
> org.apache.spark.sql.hive.execution.NativeCommand.execute(NativeCommand.scala:30)
>       at 
> org.apache.spark.sql.SQLContext$QueryExecution.toRdd$lzycompute(SQLContext.scala:425)
>       at 
> org.apache.spark.sql.SQLContext$QueryExecution.toRdd(SQLContext.scala:425)
>       at 
> org.apache.spark.sql.SchemaRDDLike$class.$init$(SchemaRDDLike.scala:58)
>       at org.apache.spark.sql.SchemaRDD.<init>(SchemaRDD.scala:108)
>       at org.apache.spark.sql.hive.HiveContext.sql(HiveContext.scala:94)
>       at 
> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.run(Shim13.scala:161)
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:231)
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync(HiveSessionImpl.java:218)
>       at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:483)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:79)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:37)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:64)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:422)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>       at 
> org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:493)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:60)
>       at com.sun.proxy.$Proxy18.executeStatementAsync(Unknown Source)
>       at 
> org.apache.hive.service.cli.CLIService.executeStatementAsync(CLIService.java:233)
>       at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:344)
>       at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1313)
>       at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1298)
>       at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>       at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>       at 
> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:55)
>       at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>       at java.lang.Thread.run(Thread.java:745)
> 14/12/19 21:44:55 WARN thrift.ThriftCLIService: Error executing statement:
> org.apache.hive.service.cli.HiveSQLException: 
> org.apache.spark.sql.execution.QueryExecutionException: FAILED: Operation 
> cancelled
>       at 
> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.run(Shim13.scala:192)
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:231)
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync(HiveSessionImpl.java:218)
>       at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:483)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:79)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:37)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:64)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:422)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>       at 
> org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:493)
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:60)
>       at com.sun.proxy.$Proxy18.executeStatementAsync(Unknown Source)
>       at 
> org.apache.hive.service.cli.CLIService.executeStatementAsync(CLIService.java:233)
>       at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:344)
>       at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1313)
>       at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1298)
>       at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>       at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>       at 
> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:55)
>       at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>       at java.lang.Thread.run(Thread.java:745)
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to