Hello and happy holiday to those who are already enjoying it!

I am still having trouble running Hive with Spark. I downloaded Spark 1.5.2 and 
built it like this (my Hadoop is version 2.7.1):

./make-distribution.sh --name "hadoop2-without-hive" --tgz 
"-Pyarn,hadoop-provided,hadoop-2.4,parquet-provided”

When trying to run it with Hive 1.2.1 (a simple command that creates a Spark 
job like ‘Select count(*) from userstweetsdailystatistics;') get the following 
error

15/12/24 17:12:54 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:54 INFO log.PerfLogger: <PERFLOG method=SparkBuildPlan 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator>
15/12/24 17:12:54 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:54 INFO log.PerfLogger: <PERFLOG method=SparkCreateTran.Map 1 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator>
15/12/24 17:12:54 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:54 INFO Configuration.deprecation: mapred.task.is.map is deprecated. 
Instead, use mapreduce.task.ismap
15/12/24 17:12:54 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:54 INFO exec.Utilities: Processing alias userstweetsdailystatistics
15/12/24 17:12:54 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:54 INFO exec.Utilities: Adding input file 
hdfs://hadoop-master:8020/user/ubuntu/hive/warehouse/userstweetsdailystatistics
15/12/24 17:12:55 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:55 INFO log.PerfLogger: <PERFLOG method=serializePlan 
from=org.apache.hadoop.hive.ql.exec.Utilities>
15/12/24 17:12:55 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:55 INFO exec.Utilities: Serializing MapWork via kryo
15/12/24 17:12:56 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:56 INFO log.PerfLogger: </PERFLOG method=serializePlan 
start=1450973575887 end=1450973576279 duration=392 
from=org.apache.hadoop.hive.ql.exec.Utilities>
15/12/24 17:12:57 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:57 INFO storage.MemoryStore: ensureFreeSpace(572800) called with 
curMem=0, maxMem=556038881
15/12/24 17:12:57 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:57 INFO storage.MemoryStore: Block broadcast_0 stored as values in memory 
(estimated size 559.4 KB, free 529.7 MB)
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO storage.MemoryStore: ensureFreeSpace(43075) called with 
curMem=572800, maxMem=556038881
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO storage.MemoryStore: Block broadcast_0_piece0 stored as bytes in 
memory (estimated size 42.1 KB, free 529.7 MB)
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO storage.BlockManagerInfo: Added broadcast_0_piece0 in memory on 
192.168.1.64:49690 (size: 42.1 KB, free: 530.2 MB)
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 ERROR util.Utils: uncaught error in thread SparkListenerBus, stopping 
SparkContext
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 
java.lang.AbstractMethodError
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.spark.scheduler.SparkListenerBus$class.onPostEvent(SparkListenerBus.scala:62)
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.spark.scheduler.LiveListenerBus.onPostEvent(LiveListenerBus.scala:31)
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.spark.scheduler.LiveListenerBus.onPostEvent(LiveListenerBus.scala:31)
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.spark.util.ListenerBus$class.postToAll(ListenerBus.scala:56)
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.spark.util.AsynchronousListenerBus.postToAll(AsynchronousListenerBus.scala:37)
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.spark.util.AsynchronousListenerBus$$anon$1$$anonfun$run$1.apply$mcV$sp(AsynchronousListenerBus.scala:79)
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.spark.util.Utils$.tryOrStopSparkContext(Utils.scala:1136)
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.spark.util.AsynchronousListenerBus$$anon$1.run(AsynchronousListenerBus.scala:63)
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/metrics/json,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages/stage/kill,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/api,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/static,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/executors/threadDump/json,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/executors/threadDump,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/executors/json,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/executors,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/environment/json,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/environment,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/storage/rdd/json,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/storage/rdd,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/storage/json,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/storage,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages/pool/json,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages/pool,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages/stage/json,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages/stage,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages/json,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/jobs/job/json,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/jobs/job,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/jobs/json,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/jobs,null}
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO spark.SparkContext: Created broadcast 0 from hadoopRDD at 
SparkPlanGenerator.java:188
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO ui.SparkUI: Stopped Spark web UI at http://192.168.1.64:4040
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO scheduler.DAGScheduler: Stopping DAGScheduler
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO cluster.SparkDeploySchedulerBackend: Shutting down all executors
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO cluster.SparkDeploySchedulerBackend: Asking each executor to shut 
down
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO log.PerfLogger: </PERFLOG method=SparkCreateTran.Map 1 
start=1450973574712 end=1450973578874 duration=4162 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator>
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO log.PerfLogger: <PERFLOG method=SparkCreateTran.Reducer 2 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator>
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO log.PerfLogger: <PERFLOG method=serializePlan 
from=org.apache.hadoop.hive.ql.exec.Utilities>
15/12/24 17:12:58 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:58 INFO exec.Utilities: Serializing ReduceWork via kryo
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:59 INFO log.PerfLogger: </PERFLOG method=serializePlan 
start=1450973578926 end=1450973579000 duration=74 
from=org.apache.hadoop.hive.ql.exec.Utilities>
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:59 INFO log.PerfLogger: </PERFLOG method=SparkCreateTran.Reducer 2 
start=1450973578874 end=1450973579073 duration=199 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator>
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:59 INFO log.PerfLogger: </PERFLOG method=SparkBuildPlan 
start=1450973574707 end=1450973579074 duration=4367 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator>
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:59 INFO log.PerfLogger: <PERFLOG method=SparkBuildRDDGraph 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlan>
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:59 WARN remote.ReliableDeliverySupervisor: Association with remote system 
[akka.tcp://sparkExecutor@192.168.1.64:35089] has failed, address is now gated 
for [5000] ms. Reason: [Disassociated] 
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:59 INFO log.PerfLogger: </PERFLOG method=SparkBuildRDDGraph 
start=1450973579074 end=1450973579273 duration=199 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlan>
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl: 15/12/24 
17:12:59 INFO client.RemoteDriver: Failed to run job 
d3746d11-eac8-4bf9-9897-bef27fd0423e
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl: 
java.lang.IllegalStateException: Cannot call methods on a stopped SparkContext
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.spark.SparkContext.org$apache$spark$SparkContext$$assertNotStopped(SparkContext.scala:104)
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.spark.SparkContext.submitJob(SparkContext.scala:1981)
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$1.apply(AsyncRDDActions.scala:118)
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$1.apply(AsyncRDDActions.scala:116)
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:147)
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:108)
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.spark.rdd.RDD.withScope(RDD.scala:310)
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.spark.rdd.AsyncRDDActions.foreachAsync(AsyncRDDActions.scala:116)
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.spark.api.java.JavaRDDLike$class.foreachAsync(JavaRDDLike.scala:690)
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.spark.api.java.AbstractJavaRDDLike.foreachAsync(JavaRDDLike.scala:47)
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.hadoop.hive.ql.exec.spark.RemoteHiveSparkClient$JobStatusJob.call(RemoteHiveSparkClient.java:257)
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.hive.spark.client.RemoteDriver$JobWrapper.call(RemoteDriver.java:366)
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
org.apache.hive.spark.client.RemoteDriver$JobWrapper.call(RemoteDriver.java:335)
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
java.util.concurrent.FutureTask.run(FutureTask.java:262)
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
15/12/24 17:12:59 [stderr-redir-1]: INFO client.SparkClientImpl:        at 
java.lang.Thread.run(Thread.java:745)
15/12/24 17:12:59 [RPC-Handler-3]: INFO client.SparkClientImpl: Received result 
for d3746d11-eac8-4bf9-9897-bef27fd0423e
Status: Failed
15/12/24 17:12:59 [Thread-8]: ERROR status.SparkJobMonitor: Status: Failed
15/12/24 17:12:59 [Thread-8]: INFO log.PerfLogger: </PERFLOG method=SparkRunJob 
start=1450973569576 end=1450973579584 duration=10008 
from=org.apache.hadoop.hive.ql.exec.spark.status.SparkJobMonitor>
FAILED: Execution Error, return code 3 from 
org.apache.hadoop.hive.ql.exec.spark.SparkTask
15/12/24 17:13:01 [main]: ERROR ql.Driver: FAILED: Execution Error, return code 
3 from org.apache.hadoop.hive.ql.exec.spark.SparkTask
15/12/24 17:13:01 [main]: INFO log.PerfLogger: </PERFLOG method=Driver.execute 
start=1450973565261 end=1450973581307 duration=16046 
from=org.apache.hadoop.hive.ql.Driver>
15/12/24 17:13:01 [main]: INFO log.PerfLogger: <PERFLOG method=releaseLocks 
from=org.apache.hadoop.hive.ql.Driver>
15/12/24 17:13:01 [main]: INFO log.PerfLogger: </PERFLOG method=releaseLocks 
start=1450973581308 end=1450973581308 duration=0 
from=org.apache.hadoop.hive.ql.Driver>
15/12/24 17:13:01 [main]: INFO exec.ListSinkOperator: 7 finished. closing... 
15/12/24 17:13:01 [main]: INFO exec.ListSinkOperator: 7 Close done
15/12/24 17:13:01 [main]: INFO log.PerfLogger: <PERFLOG method=releaseLocks 
from=org.apache.hadoop.hive.ql.Driver>
15/12/24 17:13:01 [main]: INFO log.PerfLogger: </PERFLOG method=releaseLocks 
start=1450973581362 end=1450973581362 duration=0 
from=org.apache.hadoop.hive.ql.Driver>


The only useful thing I can find at the Spark side is in the worker log:

15/12/24 17:12:53 INFO worker.Worker: Asked to launch executor 
app-20151224171253-0000/0 for Hive on Spark
15/12/24 17:12:53 INFO spark.SecurityManager: Changing view acls to: ubuntu
15/12/24 17:12:53 INFO spark.SecurityManager: Changing modify acls to: ubuntu
15/12/24 17:12:53 INFO spark.SecurityManager: SecurityManager: authentication 
disabled; ui acls disabled; users with view permissions: Set(ubuntu); users 
with modify permissions: Set(ubuntu)
15/12/24 17:12:53 INFO worker.ExecutorRunner: Launch command: 
"/usr/lib/jvm/java-7-openjdk-amd64/bin/java" "-cp" 
"/usr/local/hadoop/etc/hadoop/:/usr/local/hadoop/share/hadoop/common/lib/*:/usr/local/hadoop/share/hadoop/common/*:/usr/local/hadoop/share/hadoop/hdfs/:/usr/local/hadoop/share/hadoop/hdfs/lib/*:/usr/local/hadoop/share/hadoop/hdfs/*:/usr/local/hadoop/share/hadoop/yarn/lib/*:/usr/local/hadoop/share/hadoop/yarn/*:/usr/local/hadoop/share/hadoop/mapreduce/lib/*:/usr/local/hadoop/share/hadoop/mapreduce/*:/usr/local/hadoop/contrib/capacity-scheduler/*.jar:/home/ubuntu/Downloads/spark-1.5.2-bin-hadoop2-without-hive/lib/spark-assembly-1.5.2-hadoop2.4.0.jar:/home/ubuntu/Downloads/spark-1.5.2-bin-hadoop2-without-hive/lib/spark-assembly-1.5.2-hadoop2.4.0.jar:/usr/local/hadoop/etc/hadoop/:/usr/local/hadoop/share/hadoop/common/lib/*:/usr/local/hadoop/share/hadoop/common/*:/usr/local/hadoop/share/hadoop/hdfs/:/usr/local/hadoop/share/hadoop/hdfs/lib/*:/usr/local/hadoop/share/hadoop/hdfs/*:/usr/local/hadoop/share/hadoop/yarn/lib/*:/usr/local/hadoop/share/hadoop/yarn/*:/usr/local/hadoop/share/hadoop/mapreduce/lib/*:/usr/local/hadoop/share/hadoop/mapreduce/*:/usr/local/hadoop/contrib/capacity-scheduler/*.jar:/home/ubuntu/Downloads/spark-1.5.2-bin-hadoop2-without-hive/lib/spark-assembly-1.5.2-hadoop2.4.0.jar:/home/ubuntu/Downloads/spark-1.5.2-bin-hadoop2-without-hive/lib/spark-assembly-1.5.2-hadoop2.4.0.jar:/home/ubuntu/Downloads/spark-1.5.2-bin-hadoop2-without-hive/sbin/../conf/:/home/ubuntu/Downloads/spark-1.5.2-bin-hadoop2-without-hive/lib/spark-assembly-1.5.2-hadoop2.4.0.jar:/usr/local/hadoop/etc/hadoop/:/usr/local/hadoop/etc/hadoop/:/usr/local/hadoop/share/hadoop/common/lib/*:/usr/local/hadoop/share/hadoop/common/*:/usr/local/hadoop/share/hadoop/hdfs/:/usr/local/hadoop/share/hadoop/hdfs/lib/*:/usr/local/hadoop/share/hadoop/hdfs/*:/usr/local/hadoop/share/hadoop/yarn/lib/*:/usr/local/hadoop/share/hadoop/yarn/*:/usr/local/hadoop/share/hadoop/mapreduce/lib/*:/usr/local/hadoop/share/hadoop/mapreduce/*:/usr/local/hadoop/contrib/capacity-scheduler/*.jar"
 "-Xms1024M" "-Xmx1024M" "-Dspark.driver.port=44858" 
"-Dhive.spark.log.dir=/home/ubuntu/Downloads/spark-1.5.2-bin-hadoop2-without-hive/logs/"
 "-XX:MaxPermSize=256m" 
"org.apache.spark.executor.CoarseGrainedExecutorBackend" "--driver-url" 
"akka.tcp://sparkDriver@192.168.1.64:44858/user/CoarseGrainedScheduler" 
"--executor-id" "0" "--hostname" "192.168.1.64" "--cores" "3" "--app-id" 
"app-20151224171253-0000" "--worker-url" 
"akka.tcp://sparkWorker@192.168.1.64:54209/user/Worker"
15/12/24 17:12:58 INFO worker.Worker: Asked to kill executor 
app-20151224171253-0000/0
15/12/24 17:12:58 INFO worker.ExecutorRunner: Runner thread for executor 
app-20151224171253-0000/0 interrupted
15/12/24 17:12:58 INFO worker.ExecutorRunner: Killing process!
15/12/24 17:12:58 ERROR logging.FileAppender: Error writing stream to file 
/home/ubuntu/Downloads/spark-1.5.2-bin-hadoop2-without-hive/work/app-20151224171253-0000/0/stderr
java.io.IOException: Stream closed
        at 
java.io.BufferedInputStream.getBufIfOpen(BufferedInputStream.java:162)
        at java.io.BufferedInputStream.read1(BufferedInputStream.java:272)
        at java.io.BufferedInputStream.read(BufferedInputStream.java:334)
        at java.io.FilterInputStream.read(FilterInputStream.java:107)
        at 
org.apache.spark.util.logging.FileAppender.appendStreamToFile(FileAppender.scala:70)
        at 
org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply$mcV$sp(FileAppender.scala:39)
        at 
org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)
        at 
org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)
        at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1699)
        at 
org.apache.spark.util.logging.FileAppender$$anon$1.run(FileAppender.scala:38)
15/12/24 17:12:59 INFO worker.Worker: Executor app-20151224171253-0000/0 
finished with state KILLED exitStatus 143
15/12/24 17:12:59 INFO worker.Worker: Cleaning up local directories for 
application app-20151224171253-0000
15/12/24 17:12:59 INFO shuffle.ExternalShuffleBlockResolver: Application 
app-20151224171253-0000 removed, cleanupLocalDirs = true

Here is my Spark configuration

export HADOOP_HOME=/usr/local/hadoop
export PATH=$PATH:$HADOOP_HOME/bin
export SPARK_DIST_CLASSPATH=`hadoop class path`


Any hints as to what could be going wrong? Why is the executor getting killed? 
Have I built Spark wrongly? I have tried building it in several different ways 
and I keep failing.
I must admit I am confused with the information I find online on how to 
use/build Spark on Hive and which version goes with what.
Can I download a pre-built version of Spark that would be suitable with my 
existing Hadoop 2.7.1 and my Hive 1.2.1?
This error has been baffling me for weeks..


More than grateful for any help!
Sofia


Reply via email to