Hi Josh It was great meeting u in person at the spark-summit SFO yesterday. Thanks for discussing potential solutions to the problem. I verified that 2 hive gateway nodes had not been configured correctly. My bad. I added hive-site.xml to the spark Conf directories for these 2 additional hive gateway nodes.
Plus I increased the driver-memory parameter to 1gb. That solved the memory issue. So good news is I can get spark-SQL running in standalone mode (on a CDH 5.3.3 with spark 1.2 on YARN) Not so good news is that the following params have no effect --master yarn --deployment-mode client So the spark-SQL query runs with only ONE executor :-( I am planning on bugging u for 5-10 minutes at the Spark office hours :-) and hopefully we can solve this. Thanks Best regards Sanjay Sent from my iPhone > On Jun 13, 2015, at 5:38 PM, Josh Rosen <rosenvi...@gmail.com> wrote: > > Try using Spark 1.4.0 with SQL code generation turned on; this should make a > huge difference. > >> On Sat, Jun 13, 2015 at 5:08 PM, Sanjay Subramanian >> <sanjaysubraman...@yahoo.com> wrote: >> hey guys >> >> I tried the following settings as well. No luck >> >> --total-executor-cores 24 --executor-memory 4G >> >> >> BTW on the same cluster , impala absolutely kills it. same query 9 seconds. >> no memory issues. no issues. >> >> In fact I am pretty disappointed with Spark-SQL. >> I have worked with Hive during the 0.9.x stages and taken projects to >> production successfully and Hive actually very rarely craps out. >> >> Whether the spark folks like what I say or not, yes my expectations are >> pretty high of Spark-SQL if I were to change the ways we are doing things at >> my workplace. >> Until that time, we are going to be hugely dependent on Impala and >> Hive(with SSD speeding up the shuffle stage , even MR jobs are not that slow >> now). >> >> I want to clarify for those of u who may be asking - why I am not using >> spark with Scala and insisting on using spark-sql ? >> >> - I have already pipelined data from enterprise tables to Hive >> - I am using CDH 5.3.3 (Cloudera starving developers version) >> - I have close to 300 tables defined in Hive external tables. >> - Data if on HDFS >> - On an average we have 150 columns per table >> - One an everyday basis , we do crazy amounts of ad-hoc joining of new and >> old tables in getting datasets ready for supervised ML >> - I thought that quite simply I can point Spark to the Hive meta and do >> queries as I do - in fact the existing queries would work as is unless I am >> using some esoteric Hive/Impala function >> >> Anyway, if there are some settings I can use and get spark-sql to run even >> on standalone mode that will be huge help. >> >> On the pre-production cluster I have spark on YARN but could never get it to >> run fairly complex queries and I have no answers from this group of the CDH >> groups. >> >> So my assumption is that its possibly not solved , else I have always got >> very quick answers and responses :-) to my questions on all CDH groups, >> Spark, Hive >> >> best regards >> >> sanjay >> >> >> >> From: Josh Rosen <rosenvi...@gmail.com> >> To: Sanjay Subramanian <sanjaysubraman...@yahoo.com> >> Cc: "user@spark.apache.org" <user@spark.apache.org> >> Sent: Friday, June 12, 2015 7:15 AM >> Subject: Re: spark-sql from CLI --->EXCEPTION: java.lang.OutOfMemoryError: >> Java heap space >> >> It sounds like this might be caused by a memory configuration problem. In >> addition to looking at the executor memory, I'd also bump up the driver >> memory, since it appears that your shell is running out of memory when >> collecting a large query result. >> >> Sent from my phone >> >> >> >>> On Jun 11, 2015, at 8:43 AM, Sanjay Subramanian >>> <sanjaysubraman...@yahoo.com.INVALID> wrote: >>> >>> hey guys >>> >>> Using Hive and Impala daily intensively. >>> Want to transition to spark-sql in CLI mode >>> >>> Currently in my sandbox I am using the Spark (standalone mode) in the CDH >>> distribution (starving developer version 5.3.3) >>> 3 datanode hadoop cluster >>> 32GB RAM per node >>> 8 cores per node >>> >>> spark >>> 1.2.0+cdh5.3.3+371 >>> >>> >>> I am testing some stuff on one view and getting memory errors >>> Possibly reason is default memory per executor showing on 18080 is >>> 512M >>> >>> These options when used to start the spark-sql CLI does not seem to have >>> any effect >>> --total-executor-cores 12 --executor-memory 4G >>> >>> >>> >>> /opt/cloudera/parcels/CDH/lib/spark/bin/spark-sql -e "select distinct >>> isr,event_dt,age,age_cod,sex,year,quarter from aers.aers_demo_view" >>> >>> aers.aers_demo_view (7 million+ records) >>> =================== >>> isr bigint case id >>> event_dt bigint Event date >>> age double age of patient >>> age_cod string days,months years >>> sex string M or F >>> year int >>> quarter int >>> >>> >>> VIEW DEFINITION >>> ================ >>> CREATE VIEW `aers.aers_demo_view` AS SELECT `isr` AS `isr`, `event_dt` AS >>> `event_dt`, `age` AS `age`, `age_cod` AS `age_cod`, `gndr_cod` AS `sex`, >>> `year` AS `year`, `quarter` AS `quarter` FROM (SELECT >>> `aers_demo_v1`.`isr`, >>> `aers_demo_v1`.`event_dt`, >>> `aers_demo_v1`.`age`, >>> `aers_demo_v1`.`age_cod`, >>> `aers_demo_v1`.`gndr_cod`, >>> `aers_demo_v1`.`year`, >>> `aers_demo_v1`.`quarter` >>> FROM >>> `aers`.`aers_demo_v1` >>> UNION ALL >>> SELECT >>> `aers_demo_v2`.`isr`, >>> `aers_demo_v2`.`event_dt`, >>> `aers_demo_v2`.`age`, >>> `aers_demo_v2`.`age_cod`, >>> `aers_demo_v2`.`gndr_cod`, >>> `aers_demo_v2`.`year`, >>> `aers_demo_v2`.`quarter` >>> FROM >>> `aers`.`aers_demo_v2` >>> UNION ALL >>> SELECT >>> `aers_demo_v3`.`isr`, >>> `aers_demo_v3`.`event_dt`, >>> `aers_demo_v3`.`age`, >>> `aers_demo_v3`.`age_cod`, >>> `aers_demo_v3`.`gndr_cod`, >>> `aers_demo_v3`.`year`, >>> `aers_demo_v3`.`quarter` >>> FROM >>> `aers`.`aers_demo_v3` >>> UNION ALL >>> SELECT >>> `aers_demo_v4`.`isr`, >>> `aers_demo_v4`.`event_dt`, >>> `aers_demo_v4`.`age`, >>> `aers_demo_v4`.`age_cod`, >>> `aers_demo_v4`.`gndr_cod`, >>> `aers_demo_v4`.`year`, >>> `aers_demo_v4`.`quarter` >>> FROM >>> `aers`.`aers_demo_v4` >>> UNION ALL >>> SELECT >>> `aers_demo_v5`.`primaryid` AS `ISR`, >>> `aers_demo_v5`.`event_dt`, >>> `aers_demo_v5`.`age`, >>> `aers_demo_v5`.`age_cod`, >>> `aers_demo_v5`.`gndr_cod`, >>> `aers_demo_v5`.`year`, >>> `aers_demo_v5`.`quarter` >>> FROM >>> `aers`.`aers_demo_v5` >>> UNION ALL >>> SELECT >>> `aers_demo_v6`.`primaryid` AS `ISR`, >>> `aers_demo_v6`.`event_dt`, >>> `aers_demo_v6`.`age`, >>> `aers_demo_v6`.`age_cod`, >>> `aers_demo_v6`.`sex` AS `GNDR_COD`, >>> `aers_demo_v6`.`year`, >>> `aers_demo_v6`.`quarter` >>> FROM >>> `aers`.`aers_demo_v6`) `aers_demo_view` >>> >>> >>> >>> >>> >>> >>> >>> 15/06/11 08:36:36 WARN DefaultChannelPipeline: An exception was thrown by a >>> user handler while handling an exception event ([id: 0x01b99855, >>> /10.0.0.19:58117 => /10.0.0.19:52016] EXCEPTION: >>> java.lang.OutOfMemoryError: Java heap space) >>> java.lang.OutOfMemoryError: Java heap space >>> at >>> org.jboss.netty.buffer.HeapChannelBuffer.<init>(HeapChannelBuffer.java:42) >>> at >>> org.jboss.netty.buffer.BigEndianHeapChannelBuffer.<init>(BigEndianHeapChannelBuffer.java:34) >>> at >>> org.jboss.netty.buffer.ChannelBuffers.buffer(ChannelBuffers.java:134) >>> at >>> org.jboss.netty.buffer.HeapChannelBufferFactory.getBuffer(HeapChannelBufferFactory.java:68) >>> at >>> org.jboss.netty.buffer.AbstractChannelBufferFactory.getBuffer(AbstractChannelBufferFactory.java:48) >>> at >>> org.jboss.netty.handler.codec.frame.FrameDecoder.newCumulationBuffer(FrameDecoder.java:507) >>> at >>> org.jboss.netty.handler.codec.frame.FrameDecoder.updateCumulation(FrameDecoder.java:345) >>> at >>> org.jboss.netty.handler.codec.frame.FrameDecoder.messageReceived(FrameDecoder.java:312) >>> at >>> org.jboss.netty.channel.Channels.fireMessageReceived(Channels.java:268) >>> at >>> org.jboss.netty.channel.Channels.fireMessageReceived(Channels.java:255) >>> at >>> org.jboss.netty.channel.socket.nio.NioWorker.read(NioWorker.java:88) >>> at >>> org.jboss.netty.channel.socket.nio.AbstractNioWorker.process(AbstractNioWorker.java:109) >>> at >>> org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:312) >>> at >>> org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:90) >>> at >>> org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) >>> at >>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>> at java.lang.Thread.run(Thread.java:745) >>> 15/06/11 08:36:40 ERROR Utils: Uncaught exception in thread >>> task-result-getter-0 >>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>> at java.lang.Long.valueOf(Long.java:577) >>> at >>> com.esotericsoftware.kryo.serializers.DefaultSerializers$LongSerializer.read(DefaultSerializers.java:113) >>> at >>> com.esotericsoftware.kryo.serializers.DefaultSerializers$LongSerializer.read(DefaultSerializers.java:103) >>> at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:732) >>> at >>> com.esotericsoftware.kryo.serializers.DefaultArraySerializers$ObjectArraySerializer.read(DefaultArraySerializers.java:338) >>> at >>> com.esotericsoftware.kryo.serializers.DefaultArraySerializers$ObjectArraySerializer.read(DefaultArraySerializers.java:293) >>> at com.esotericsoftware.kryo.Kryo.readObject(Kryo.java:651) >>> at >>> com.esotericsoftware.kryo.serializers.FieldSerializer$ObjectField.read(FieldSerializer.java:605) >>> at >>> com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:221) >>> at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:732) >>> at >>> com.esotericsoftware.kryo.serializers.DefaultArraySerializers$ObjectArraySerializer.read(DefaultArraySerializers.java:338) >>> at >>> com.esotericsoftware.kryo.serializers.DefaultArraySerializers$ObjectArraySerializer.read(DefaultArraySerializers.java:293) >>> at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:732) >>> at >>> org.apache.spark.serializer.KryoSerializerInstance.deserialize(KryoSerializer.scala:171) >>> at >>> org.apache.spark.scheduler.DirectTaskResult.value(TaskResult.scala:79) >>> at >>> org.apache.spark.scheduler.TaskSetManager.handleSuccessfulTask(TaskSetManager.scala:558) >>> at >>> org.apache.spark.scheduler.TaskSchedulerImpl.handleSuccessfulTask(TaskSchedulerImpl.scala:352) >>> at >>> org.apache.spark.scheduler.TaskResultGetter$$anon$2$$anonfun$run$1.apply$mcV$sp(TaskResultGetter.scala:80) >>> at >>> org.apache.spark.scheduler.TaskResultGetter$$anon$2$$anonfun$run$1.apply(TaskResultGetter.scala:49) >>> at >>> org.apache.spark.scheduler.TaskResultGetter$$anon$2$$anonfun$run$1.apply(TaskResultGetter.scala:49) >>> at >>> org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1468) >>> at >>> org.apache.spark.scheduler.TaskResultGetter$$anon$2.run(TaskResultGetter.scala:48) >>> at >>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>> at java.lang.Thread.run(Thread.java:745) >>> 15/06/11 08:36:38 ERROR ActorSystemImpl: exception on LARS’ timer thread >>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>> at akka.dispatch.AbstractNodeQueue.<init>(AbstractNodeQueue.java:19) >>> at >>> akka.actor.LightArrayRevolverScheduler$TaskQueue.<init>(Scheduler.scala:431) >>> at >>> akka.actor.LightArrayRevolverScheduler$$anon$12.nextTick(Scheduler.scala:397) >>> at >>> akka.actor.LightArrayRevolverScheduler$$anon$12.run(Scheduler.scala:363) >>> at java.lang.Thread.run(Thread.java:745) >>> Exception in thread "task-result-getter-0" java.lang.OutOfMemoryError: GC >>> overhead limit exceeded >>> at java.lang.Long.valueOf(Long.java:577) >>> at >>> com.esotericsoftware.kryo.serializers.DefaultSerializers$LongSerializer.read(DefaultSerializers.java:113) >>> at >>> com.esotericsoftware.kryo.serializers.DefaultSerializers$LongSerializer.read(DefaultSerializers.java:103) >>> at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:732) >>> at >>> com.esotericsoftware.kryo.serializers.DefaultArraySerializers$ObjectArraySerializer.read(DefaultArraySerializers.java:338) >>> at >>> com.esotericsoftware.kryo.serializers.DefaultArraySerializers$ObjectArraySerializer.read(DefaultArraySerializers.java:293) >>> at com.esotericsoftware.kryo.Kryo.readObject(Kryo.java:651) >>> at >>> com.esotericsoftware.kryo.serializers.FieldSerializer$ObjectField.read(FieldSerializer.java:605) >>> at >>> com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:221) >>> at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:732) >>> at >>> com.esotericsoftware.kryo.serializers.DefaultArraySerializers$ObjectArraySerializer.read(DefaultArraySerializers.java:338) >>> at >>> com.esotericsoftware.kryo.serializers.DefaultArraySerializers$ObjectArraySerializer.read(DefaultArraySerializers.java:293) >>> at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:732) >>> at >>> org.apache.spark.serializer.KryoSerializerInstance.deserialize(KryoSerializer.scala:171) >>> at >>> org.apache.spark.scheduler.DirectTaskResult.value(TaskResult.scala:79) >>> at >>> org.apache.spark.scheduler.TaskSetManager.handleSuccessfulTask(TaskSetManager.scala:558) >>> at >>> org.apache.spark.scheduler.TaskSchedulerImpl.handleSuccessfulTask(TaskSchedulerImpl.scala:352) >>> at >>> org.apache.spark.scheduler.TaskResultGetter$$anon$2$$anonfun$run$1.apply$mcV$sp(TaskResultGetter.scala:80) >>> at >>> org.apache.spark.scheduler.TaskResultGetter$$anon$2$$anonfun$run$1.apply(TaskResultGetter.scala:49) >>> at >>> org.apache.spark.scheduler.TaskResultGetter$$anon$2$$anonfun$run$1.apply(TaskResultGetter.scala:49) >>> at >>> org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1468) >>> at >>> org.apache.spark.scheduler.TaskResultGetter$$anon$2.run(TaskResultGetter.scala:48) >>> at >>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>> at java.lang.Thread.run(Thread.java:745) >>> 15/06/11 08:36:41 ERROR ActorSystemImpl: Uncaught fatal error from thread >>> [sparkDriver-scheduler-1] shutting down ActorSystem [sparkDriver] >>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>> at akka.dispatch.AbstractNodeQueue.<init>(AbstractNodeQueue.java:19) >>> at >>> akka.actor.LightArrayRevolverScheduler$TaskQueue.<init>(Scheduler.scala:431) >>> at >>> akka.actor.LightArrayRevolverScheduler$$anon$12.nextTick(Scheduler.scala:397) >>> at >>> akka.actor.LightArrayRevolverScheduler$$anon$12.run(Scheduler.scala:363) >>> at java.lang.Thread.run(Thread.java:745) >>> 15/06/11 08:36:46 ERROR ActorSystemImpl: Uncaught fatal error from thread >>> [sparkDriver-akka.actor.default-dispatcher-4] shutting down ActorSystem >>> [sparkDriver] >>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>> 15/06/11 08:36:46 ERROR SparkSQLDriver: Failed in [select distinct >>> isr,event_dt,age,age_cod,sex,year,quarter from aers.aers_demo_view] >>> org.apache.spark.SparkException: Job cancelled because SparkContext was >>> shut down >>> at >>> org.apache.spark.scheduler.DAGScheduler$$anonfun$cleanUpAfterSchedulerStop$1.apply(DAGScheduler.scala:702) >>> at >>> org.apache.spark.scheduler.DAGScheduler$$anonfun$cleanUpAfterSchedulerStop$1.apply(DAGScheduler.scala:701) >>> at scala.collection.mutable.HashSet.foreach(HashSet.scala:79) >>> at >>> org.apache.spark.scheduler.DAGScheduler.cleanUpAfterSchedulerStop(DAGScheduler.scala:701) >>> at >>> org.apache.spark.scheduler.DAGSchedulerEventProcessActor.postStop(DAGScheduler.scala:1428) >>> at >>> akka.actor.dungeon.FaultHandling$class.akka$actor$dungeon$FaultHandling$$finishTerminate(FaultHandling.scala:201) >>> at >>> akka.actor.dungeon.FaultHandling$class.terminate(FaultHandling.scala:163) >>> at akka.actor.ActorCell.terminate(ActorCell.scala:338) >>> at akka.actor.ActorCell.invokeAll$1(ActorCell.scala:431) >>> at akka.actor.ActorCell.systemInvoke(ActorCell.scala:447) >>> at akka.dispatch.Mailbox.processAllSystemMessages(Mailbox.scala:262) >>> at akka.dispatch.Mailbox.run(Mailbox.scala:218) >>> at >>> akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386) >>> at >>> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) >>> at >>> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) >>> at >>> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) >>> at >>> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) >>> 15/06/11 08:36:51 WARN DefaultChannelPipeline: An exception was thrown by a >>> user handler while handling an exception event ([id: 0x79935a9b, >>> /10.0.0.35:54028 => /10.0.0.19:52016] EXCEPTION: >>> java.lang.OutOfMemoryError: Java heap space) >>> java.lang.OutOfMemoryError: Java heap space >>> 15/06/11 08:36:52 ERROR ActorSystemImpl: Uncaught fatal error from thread >>> [sparkDriver-akka.actor.default-dispatcher-5] shutting down ActorSystem >>> [sparkDriver] >>> java.lang.OutOfMemoryError: Java heap space >>> 15/06/11 08:36:53 WARN DefaultChannelPipeline: An exception was thrown by a >>> user handler while handling an exception event ([id: 0xcb8c4b5d, >>> /10.0.0.18:46744 => /10.0.0.19:52016] EXCEPTION: >>> java.lang.OutOfMemoryError: Java heap space) >>> java.lang.OutOfMemoryError: Java heap space >>> 15/06/11 08:36:56 WARN NioEventLoop: Unexpected exception in the selector >>> loop. >>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>> 15/06/11 08:36:57 ERROR ActorSystemImpl: Uncaught fatal error from thread >>> [sparkDriver-akka.actor.default-dispatcher-18] shutting down ActorSystem >>> [sparkDriver] >>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>> 15/06/11 08:36:58 ERROR Utils: Uncaught exception in thread >>> task-result-getter-3 >>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>> Exception in thread "task-result-getter-3" java.lang.OutOfMemoryError: GC >>> overhead limit exceeded >>> 15/06/11 08:37:01 ERROR ActorSystemImpl: Uncaught fatal error from thread >>> [sparkDriver-akka.actor.default-dispatcher-4] shutting down ActorSystem >>> [sparkDriver] >>> java.lang.OutOfMemoryError: Java heap space >>> Time taken: 70.982 seconds >>> 15/06/11 08:37:06 WARN QueuedThreadPool: 4 threads could not be stopped >>> 15/06/11 08:37:11 ERROR MapOutputTrackerMaster: Error communicating with >>> MapOutputTracker >>> akka.pattern.AskTimeoutException: >>> Recipient[Actor[akka://sparkDriver/user/MapOutputTracker#-2109395547]] had >>> already been terminated. >>> at akka.pattern.AskableActorRef$.ask$extension(AskSupport.scala:134) >>> at >>> org.apache.spark.MapOutputTracker.askTracker(MapOutputTracker.scala:111) >>> at >>> org.apache.spark.MapOutputTracker.sendTracker(MapOutputTracker.scala:122) >>> at >>> org.apache.spark.MapOutputTrackerMaster.stop(MapOutputTracker.scala:330) >>> at org.apache.spark.SparkEnv.stop(SparkEnv.scala:83) >>> at org.apache.spark.SparkContext.stop(SparkContext.scala:1210) >>> at >>> org.apache.spark.sql.hive.thriftserver.SparkSQLEnv$.stop(SparkSQLEnv.scala:66) >>> at >>> org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$$anon$1.run(SparkSQLCLIDriver.scala:107) >>> Exception in thread "Thread-3" org.apache.spark.SparkException: Error >>> communicating with MapOutputTracker >>> at >>> org.apache.spark.MapOutputTracker.askTracker(MapOutputTracker.scala:116) >>> at >>> org.apache.spark.MapOutputTracker.sendTracker(MapOutputTracker.scala:122) >>> at >>> org.apache.spark.MapOutputTrackerMaster.stop(MapOutputTracker.scala:330) >>> at org.apache.spark.SparkEnv.stop(SparkEnv.scala:83) >>> at org.apache.spark.SparkContext.stop(SparkContext.scala:1210) >>> at >>> org.apache.spark.sql.hive.thriftserver.SparkSQLEnv$.stop(SparkSQLEnv.scala:66) >>> at >>> org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$$anon$1.run(SparkSQLCLIDriver.scala:107) >>> Caused by: akka.pattern.AskTimeoutException: >>> Recipient[Actor[akka://sparkDriver/user/MapOutputTracker#-2109395547]] had >>> already been terminated. >>> at akka.pattern.AskableActorRef$.ask$extension(AskSupport.scala:134) >>> at >>> org.apache.spark.MapOutputTracker.askTracker(MapOutputTracker.scala:111) >