[ 
https://issues.apache.org/jira/browse/HIVE-7980?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14744676#comment-14744676
 ] 

lgh commented on HIVE-7980:
---------------------------

Thanks for your replay. I am sure that I am using hive 1.2.1.

The blow is the hive script:
hive> set spark.home=/usr/local/spark;
hive> set hive.execution.engine=spark;
hive> set spark.master=yarn;
hive> set spark.eventLog.enabled=true;
hive> set spark.eventLog.dir=hdfs://server1:9000/directory;
hive> set spark.serializer=org.apache.spark.serializer.KryoSerializer;
hive> select count(*) from test;
Query ID = hadoop_20150915094049_9a83ffe9-63b4-4847-b7f4-0e566f9f71d9
Total jobs = 1
Launching Job 1 out of 1
In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
  set mapreduce.job.reduces=<number>
Starting Spark Job = 0368cc5b-dd5d-4c7c-b48e-636d53ed350b

Query Hive on Spark job[0] stages:
0
1

Status: Running (Hive on Spark job[0])
Job Progress Format
CurrentTime StageId_StageAttemptId: 
SucceededTasksCount(+RunningTasksCount-FailedTasksCount)/TotalTasksCount 
[StageCost]
2015-09-15 09:41:20,764 Stage-0_0: 0(+1)/1      Stage-1_0: 0/1  
2015-09-15 09:41:22,784 Stage-0_0: 0(+1,-1)/1   Stage-1_0: 0/1  
Status: Failed
FAILED: Execution Error, return code 3 from 
org.apache.hadoop.hive.ql.exec.spark.SparkTask



Then,the blow is  the worker errlog:
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in 
[jar:file:/data/slot0/yarn/data/usercache/hadoop/filecache/17/spark-assembly-1.3.1-hadoop2.6.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in 
[jar:file:/letv/usr/local/hadoop-2.6.0/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
15/09/15 09:41:14 INFO executor.CoarseGrainedExecutorBackend: Registered signal 
handlers for [TERM, HUP, INT]
15/09/15 09:41:15 INFO spark.SecurityManager: Changing view acls to: hadoop
15/09/15 09:41:15 INFO spark.SecurityManager: Changing modify acls to: hadoop
15/09/15 09:41:15 INFO spark.SecurityManager: SecurityManager: authentication 
disabled; ui acls disabled; users with view permissions: Set(hadoop); users 
with modify permissions: Set(hadoop)
15/09/15 09:41:16 INFO slf4j.Slf4jLogger: Slf4jLogger started
15/09/15 09:41:16 INFO Remoting: Starting remoting
15/09/15 09:41:16 INFO Remoting: Remoting started; listening on addresses 
:[akka.tcp://driverPropsFetcher@10-140-110-157:21269]
15/09/15 09:41:16 INFO util.Utils: Successfully started service 
'driverPropsFetcher' on port 21269.
15/09/15 09:41:16 INFO remote.RemoteActorRefProvider$RemotingTerminator: 
Shutting down remote daemon.
15/09/15 09:41:16 INFO spark.SecurityManager: Changing view acls to: hadoop
15/09/15 09:41:16 INFO spark.SecurityManager: Changing modify acls to: hadoop
15/09/15 09:41:16 INFO spark.SecurityManager: SecurityManager: authentication 
disabled; ui acls disabled; users with view permissions: Set(hadoop); users 
with modify permissions: Set(hadoop)
15/09/15 09:41:16 INFO remote.RemoteActorRefProvider$RemotingTerminator: Remote 
daemon shut down; proceeding with flushing remote transports.
15/09/15 09:41:16 INFO slf4j.Slf4jLogger: Slf4jLogger started
15/09/15 09:41:16 INFO Remoting: Starting remoting
15/09/15 09:41:16 INFO remote.RemoteActorRefProvider$RemotingTerminator: 
Remoting shut down.
15/09/15 09:41:16 INFO Remoting: Remoting started; listening on addresses 
:[akka.tcp://sparkExecutor@10-140-110-157:15895]
15/09/15 09:41:16 INFO util.Utils: Successfully started service 'sparkExecutor' 
on port 15895.
15/09/15 09:41:16 INFO util.AkkaUtils: Connecting to MapOutputTracker: 
akka.tcp://sparkDriver@server1:12682/user/MapOutputTracker
15/09/15 09:41:16 INFO util.AkkaUtils: Connecting to BlockManagerMaster: 
akka.tcp://sparkDriver@server1:12682/user/BlockManagerMaster
15/09/15 09:41:16 INFO storage.DiskBlockManager: Created local directory at 
/data/hadoop/data12/yarn/data/usercache/hadoop/appcache/application_1441701008982_0038/blockmgr-8e6d671f-63c6-4269-a504-03ef36ae2b0f
15/09/15 09:41:16 INFO storage.DiskBlockManager: Created local directory at 
/data/hadoop/data11/yarn/data/usercache/hadoop/appcache/application_1441701008982_0038/blockmgr-c6159085-466f-4411-bf6e-224517f92b49
15/09/15 09:41:16 INFO storage.DiskBlockManager: Created local directory at 
/data/hadoop/data10/yarn/data/usercache/hadoop/appcache/application_1441701008982_0038/blockmgr-30dc8099-6827-4a53-8fd3-c150236e61da
15/09/15 09:41:16 INFO storage.DiskBlockManager: Created local directory at 
/data/hadoop/data1/yarn/data/usercache/hadoop/appcache/application_1441701008982_0038/blockmgr-3c4a6ca3-f84f-46f7-b365-7f51a605c044
15/09/15 09:41:16 INFO storage.DiskBlockManager: Created local directory at 
/data/hadoop/data4/yarn/data/usercache/hadoop/appcache/application_1441701008982_0038/blockmgr-c3c56c5b-e368-45bb-9a50-d660301d5939
15/09/15 09:41:16 INFO storage.DiskBlockManager: Created local directory at 
/data/hadoop/data5/yarn/data/usercache/hadoop/appcache/application_1441701008982_0038/blockmgr-f19e07ea-6614-4035-92bb-e89674506c68
15/09/15 09:41:16 INFO storage.DiskBlockManager: Created local directory at 
/data/hadoop/data2/yarn/data/usercache/hadoop/appcache/application_1441701008982_0038/blockmgr-c245e542-6bd1-470c-ae8d-2a3ec389b879
15/09/15 09:41:16 INFO storage.DiskBlockManager: Created local directory at 
/data/hadoop/data3/yarn/data/usercache/hadoop/appcache/application_1441701008982_0038/blockmgr-752fd55f-e15f-43cf-af75-a726bfb498d0
15/09/15 09:41:16 INFO storage.DiskBlockManager: Created local directory at 
/data/hadoop/data8/yarn/data/usercache/hadoop/appcache/application_1441701008982_0038/blockmgr-bbeae731-7a63-455d-b453-4c1915e878dc
15/09/15 09:41:16 INFO storage.DiskBlockManager: Created local directory at 
/data/hadoop/data9/yarn/data/usercache/hadoop/appcache/application_1441701008982_0038/blockmgr-7cf05ed3-aff3-455b-8db7-c7242eda8b97
15/09/15 09:41:16 INFO storage.DiskBlockManager: Created local directory at 
/data/hadoop/data6/yarn/data/usercache/hadoop/appcache/application_1441701008982_0038/blockmgr-8cb7c93b-c15a-465d-ba0a-3be5abe5c899
15/09/15 09:41:16 INFO storage.DiskBlockManager: Created local directory at 
/data/hadoop/data7/yarn/data/usercache/hadoop/appcache/application_1441701008982_0038/blockmgr-1d0087cd-0c35-4f67-8bdd-331c14235290
15/09/15 09:41:17 INFO storage.MemoryStore: MemoryStore started with capacity 
1060.3 MB
15/09/15 09:41:17 INFO util.AkkaUtils: Connecting to OutputCommitCoordinator: 
akka.tcp://sparkDriver@server1:12682/user/OutputCommitCoordinator
15/09/15 09:41:17 INFO executor.CoarseGrainedExecutorBackend: Connecting to 
driver: akka.tcp://sparkDriver@server1:12682/user/CoarseGrainedScheduler
15/09/15 09:41:17 INFO executor.CoarseGrainedExecutorBackend: Successfully 
registered with driver
15/09/15 09:41:17 INFO executor.Executor: Starting executor ID 2 on host 
10-140-110-157
15/09/15 09:41:17 INFO netty.NettyBlockTransferService: Server created on 40604
15/09/15 09:41:17 INFO storage.BlockManagerMaster: Trying to register 
BlockManager
15/09/15 09:41:17 INFO storage.BlockManagerMaster: Registered BlockManager
15/09/15 09:41:17 INFO util.AkkaUtils: Connecting to HeartbeatReceiver: 
akka.tcp://sparkDriver@server1:12682/user/HeartbeatReceiver
15/09/15 09:41:20 INFO executor.CoarseGrainedExecutorBackend: Got assigned task 0
15/09/15 09:41:20 INFO executor.Executor: Running task 0.0 in stage 0.0 (TID 0)
15/09/15 09:41:20 INFO executor.Executor: Fetching 
http://10.140.130.38:54874/jars/hive-exec-1.2.1.jar with timestamp 1442281278254
15/09/15 09:41:20 INFO util.Utils: Fetching 
http://10.140.130.38:54874/jars/hive-exec-1.2.1.jar to 
/data/hadoop/data12/yarn/data/usercache/hadoop/appcache/application_1441701008982_0038/fetchFileTemp6606399870470341414.tmp
15/09/15 09:41:20 INFO util.Utils: Copying 
/data/hadoop/data12/yarn/data/usercache/hadoop/appcache/application_1441701008982_0038/12837195511442281278254_cache
 to 
/data/slot3/yarn/data/usercache/hadoop/appcache/application_1441701008982_0038/container_1441701008982_0038_01_000003/./hive-exec-1.2.1.jar
15/09/15 09:41:21 INFO executor.Executor: Adding 
file:/data/slot3/yarn/data/usercache/hadoop/appcache/application_1441701008982_0038/container_1441701008982_0038_01_000003/./hive-exec-1.2.1.jar
 to class loader
15/09/15 09:41:21 INFO broadcast.TorrentBroadcast: Started reading broadcast 
variable 1
15/09/15 09:41:21 INFO storage.MemoryStore: ensureFreeSpace(50134) called with 
curMem=0, maxMem=1111794647
15/09/15 09:41:21 INFO storage.MemoryStore: Block broadcast_1_piece0 stored as 
bytes in memory (estimated size 49.0 KB, free 1060.2 MB)
15/09/15 09:41:21 INFO storage.BlockManagerMaster: Updated info of block 
broadcast_1_piece0
15/09/15 09:41:21 INFO broadcast.TorrentBroadcast: Reading broadcast variable 1 
took 299 ms
15/09/15 09:41:21 INFO storage.MemoryStore: ensureFreeSpace(163768) called with 
curMem=50134, maxMem=1111794647
15/09/15 09:41:21 INFO storage.MemoryStore: Block broadcast_1 stored as values 
in memory (estimated size 159.9 KB, free 1060.1 MB)
15/09/15 09:41:21 INFO rdd.HadoopRDD: Input split: 
Paths:/user/hive/warehouse/temp.db/test/test:0+12InputFormatClass: 
org.apache.hadoop.mapred.TextInputFormat

15/09/15 09:41:21 INFO broadcast.TorrentBroadcast: Started reading broadcast 
variable 0
15/09/15 09:41:22 INFO storage.MemoryStore: ensureFreeSpace(46802) called with 
curMem=213902, maxMem=1111794647
15/09/15 09:41:22 INFO storage.MemoryStore: Block broadcast_0_piece0 stored as 
bytes in memory (estimated size 45.7 KB, free 1060.0 MB)
15/09/15 09:41:22 INFO storage.BlockManagerMaster: Updated info of block 
broadcast_0_piece0
15/09/15 09:41:22 INFO broadcast.TorrentBroadcast: Reading broadcast variable 0 
took 187 ms
15/09/15 09:41:22 INFO Configuration.deprecation: mapred.task.is.map is 
deprecated. Instead, use mapreduce.task.ismap
15/09/15 09:41:22 INFO storage.MemoryStore: ensureFreeSpace(540492) called with 
curMem=260704, maxMem=1111794647
15/09/15 09:41:22 INFO storage.MemoryStore: Block broadcast_0 stored as values 
in memory (estimated size 527.8 KB, free 1059.5 MB)
15/09/15 09:41:22 INFO Configuration.deprecation: mapred.tip.id is deprecated. 
Instead, use mapreduce.task.id
15/09/15 09:41:22 INFO Configuration.deprecation: mapred.task.id is deprecated. 
Instead, use mapreduce.task.attempt.id
15/09/15 09:41:22 INFO Configuration.deprecation: mapred.task.partition is 
deprecated. Instead, use mapreduce.task.partition
15/09/15 09:41:22 INFO Configuration.deprecation: mapred.job.id is deprecated. 
Instead, use mapreduce.job.id
15/09/15 09:41:22 INFO exec.Utilities: No plan file found: 
hdfs://bird-cluster/tmp/hive/hadoop/23ba4f2d-1f4c-42d7-af83-1cfc9411344c/hive_2015-09-15_09-40-49_573_5616367127345554376-1/-mr-10003/f96f663b-b0ad-408c-8cda-534f9a49ea2e/map.xml
15/09/15 09:41:22 ERROR executor.Executor: Exception in task 0.0 in stage 0.0 
(TID 0)
java.lang.NullPointerException
        at 
org.apache.hadoop.hive.ql.io.HiveInputFormat.init(HiveInputFormat.java:255)
        at 
org.apache.hadoop.hive.ql.io.HiveInputFormat.pushProjectionsAndFilters(HiveInputFormat.java:437)
        at 
org.apache.hadoop.hive.ql.io.HiveInputFormat.pushProjectionsAndFilters(HiveInputFormat.java:430)
        at 
org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getRecordReader(CombineHiveInputFormat.java:587)
        at org.apache.spark.rdd.HadoopRDD$$anon$1.<init>(HadoopRDD.scala:236)
        at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:212)
        at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:101)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
        at org.apache.spark.scheduler.Task.run(Task.scala:64)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:203)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:745)

And blow is hive error log:
2015-09-15 09:41:20,110 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO log.PerfLogger: 
</PERFLOG method=SparkBuildRDDGraph start=1442281280024 end=1442281280109 
duration=85 from=org.apache.hadoop.hive.ql.exec.spark.SparkPlan>
2015-09-15 09:41:20,127 INFO  [RPC-Handler-3]: client.SparkClientImpl 
(SparkClientImpl.java:handle(547)) - Received spark job ID: 0 for 
0368cc5b-dd5d-4c7c-b48e-636d53ed350b
2015-09-15 09:41:20,134 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO log.PerfLogger: 
<PERFLOG method=getSplits 
from=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat>
2015-09-15 09:41:20,134 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO exec.Utilities: PLAN 
PATH = 
hdfs://bird-cluster/tmp/hive/hadoop/23ba4f2d-1f4c-42d7-af83-1cfc9411344c/hive_2015-09-15_09-40-49_573_5616367127345554376-1/-mr-10003/f96f663b-b0ad-408c-8cda-534f9a49ea2e/map.xml
2015-09-15 09:41:20,135 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO exec.Utilities: 
***************non-local mode***************
2015-09-15 09:41:20,135 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO exec.Utilities: local 
path = 
hdfs://bird-cluster/tmp/hive/hadoop/23ba4f2d-1f4c-42d7-af83-1cfc9411344c/hive_2015-09-15_09-40-49_573_5616367127345554376-1/-mr-10003/f96f663b-b0ad-408c-8cda-534f9a49ea2e/map.xml
2015-09-15 09:41:20,135 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO exec.Utilities: Open 
file to read in plan: 
hdfs://bird-cluster/tmp/hive/hadoop/23ba4f2d-1f4c-42d7-af83-1cfc9411344c/hive_2015-09-15_09-40-49_573_5616367127345554376-1/-mr-10003/f96f663b-b0ad-408c-8cda-534f9a49ea2e/map.xml
2015-09-15 09:41:20,143 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO log.PerfLogger: 
<PERFLOG method=deserializePlan from=org.apache.hadoop.hive.ql.exec.Utilities>
2015-09-15 09:41:20,144 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO exec.Utilities: 
Deserializing MapWork via kryo
2015-09-15 09:41:20,182 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO log.PerfLogger: 
</PERFLOG method=deserializePlan start=1442281280143 end=1442281280181 
duration=38 from=org.apache.hadoop.hive.ql.exec.Utilities>
2015-09-15 09:41:20,182 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO 
io.CombineHiveInputFormat: Total number of paths: 1, launching 1 threads to 
check non-combinable ones.
2015-09-15 09:41:20,198 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO 
lzo.GPLNativeCodeLoader: Loaded native gpl library from the embedded binaries
2015-09-15 09:41:20,201 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO lzo.LzoCodec: 
Successfully loaded & initialized native-lzo library [hadoop-lzo rev 
826e7d8d3e839964dd9ed2d5f83296254b2c71d3]
2015-09-15 09:41:20,355 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO 
io.CombineHiveInputFormat: CombineHiveInputSplit creating pool for 
hdfs://bird-cluster/user/hive/warehouse/temp.db/test; using filter path 
hdfs://bird-cluster/user/hive/warehouse/temp.db/test
2015-09-15 09:41:20,379 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO input.FileInputFormat: 
Total input paths to process : 1
2015-09-15 09:41:20,399 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO 
input.CombineFileInputFormat: DEBUG: Terminated node allocation with : 
CompletedNodes: 1, size left: 0
2015-09-15 09:41:20,400 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO 
io.CombineHiveInputFormat: number of splits 1
2015-09-15 09:41:20,401 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO 
io.CombineHiveInputFormat: Number of all splits 1
2015-09-15 09:41:20,401 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO log.PerfLogger: 
</PERFLOG method=getSplits start=1442281280133 end=1442281280401 duration=268 
from=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat>
2015-09-15 09:41:20,417 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO 
scheduler.DAGScheduler: Registering RDD 1 (mapPartitionsToPair at 
MapTran.java:31)
2015-09-15 09:41:20,420 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO 
scheduler.DAGScheduler: Got job 0 (foreachAsync at 
RemoteHiveSparkClient.java:257) with 1 output partitions (allowLocal=false)
2015-09-15 09:41:20,421 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO 
scheduler.DAGScheduler: Final stage: Stage 1(foreachAsync at 
RemoteHiveSparkClient.java:257)
2015-09-15 09:41:20,422 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO 
scheduler.DAGScheduler: Parents of final stage: List(Stage 0)
2015-09-15 09:41:20,429 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO 
scheduler.DAGScheduler: Missing parents: List(Stage 0)
2015-09-15 09:41:20,440 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO 
scheduler.DAGScheduler: Submitting Stage 0 (MapPartitionsRDD[1] at 
mapPartitionsToPair at MapTran.java:31), which has no missing parents
2015-09-15 09:41:20,467 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO storage.MemoryStore: 
ensureFreeSpace(163768) called with curMem=587318, maxMem=278302556
2015-09-15 09:41:20,468 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO storage.MemoryStore: 
Block broadcast_1 stored as values in memory (estimated size 159.9 KB, free 
264.7 MB)
2015-09-15 09:41:20,490 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO storage.MemoryStore: 
ensureFreeSpace(50134) called with curMem=751086, maxMem=278302556
2015-09-15 09:41:20,490 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO storage.MemoryStore: 
Block broadcast_1_piece0 stored as bytes in memory (estimated size 49.0 KB, 
free 264.6 MB)
2015-09-15 09:41:20,491 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO 
storage.BlockManagerInfo: Added broadcast_1_piece0 in memory on server1:37766 
(size: 49.0 KB, free: 265.3 MB)
2015-09-15 09:41:20,492 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO 
storage.BlockManagerMaster: Updated info of block broadcast_1_piece0
2015-09-15 09:41:20,493 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO spark.SparkContext: 
Created broadcast 1 from broadcast at DAGScheduler.scala:839
2015-09-15 09:41:20,504 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO 
scheduler.DAGScheduler: Submitting 1 missing tasks from Stage 0 
(MapPartitionsRDD[1] at mapPartitionsToPair at MapTran.java:31)
2015-09-15 09:41:20,506 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO cluster.YarnScheduler: 
Adding task set 0.0 with 1 tasks
2015-09-15 09:41:20,552 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:20 INFO 
scheduler.TaskSetManager: Starting task 0.0 in stage 0.0 (TID 0, 
10-140-110-157, RACK_LOCAL, 1515 bytes)
2015-09-15 09:41:20,760 INFO  [main]: log.PerfLogger 
(PerfLogger.java:PerfLogEnd(148)) - </PERFLOG method=SparkSubmitToRunning 
start=1442281254723 end=1442281280760 duration=26037 
from=org.apache.hadoop.hive.ql.exec.spark.status.SparkJobMonitor>
2015-09-15 09:41:20,760 INFO  [main]: status.SparkJobMonitor 
(SessionState.java:printInfo(951)) - 
Query Hive on Spark job[0] stages:
2015-09-15 09:41:20,763 INFO  [main]: status.SparkJobMonitor 
(SessionState.java:printInfo(951)) - 0
2015-09-15 09:41:20,763 INFO  [main]: status.SparkJobMonitor 
(SessionState.java:printInfo(951)) - 1
2015-09-15 09:41:20,763 INFO  [main]: status.SparkJobMonitor 
(SessionState.java:printInfo(951)) - 
Status: Running (Hive on Spark job[0])
2015-09-15 09:41:20,763 INFO  [main]: status.SparkJobMonitor 
(SessionState.java:printInfo(951)) - Job Progress Format
CurrentTime StageId_StageAttemptId: 
SucceededTasksCount(+RunningTasksCount-FailedTasksCount)/TotalTasksCount 
[StageCost]
2015-09-15 09:41:20,764 INFO  [main]: log.PerfLogger 
(PerfLogger.java:PerfLogBegin(121)) - <PERFLOG method=SparkRunStage.0_0 
from=org.apache.hadoop.hive.ql.exec.spark.status.SparkJobMonitor>
2015-09-15 09:41:20,765 INFO  [main]: status.SparkJobMonitor 
(SessionState.java:printInfo(951)) - 2015-09-15 09:41:20,764       Stage-0_0: 
0(+1)/1      Stage-1_0: 0/1  
2015-09-15 09:41:21,432 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:21 INFO 
storage.BlockManagerInfo: Added broadcast_1_piece0 in memory on 
10-140-110-157:40604 (size: 49.0 KB, free: 1060.2 MB)
2015-09-15 09:41:22,179 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:22 INFO 
storage.BlockManagerInfo: Added broadcast_0_piece0 in memory on 
10-140-110-157:40604 (size: 45.7 KB, free: 1060.2 MB)
2015-09-15 09:41:22,612 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:22 WARN 
scheduler.TaskSetManager: Lost task 0.0 in stage 0.0 (TID 0, 10-140-110-157): 
java.lang.NullPointerException
2015-09-15 09:41:22,612 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.hadoop.hive.ql.io.HiveInputFormat.init(HiveInputFormat.java:255)
2015-09-15 09:41:22,613 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.hadoop.hive.ql.io.HiveInputFormat.pushProjectionsAndFilters(HiveInputFormat.java:437)
2015-09-15 09:41:22,613 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.hadoop.hive.ql.io.HiveInputFormat.pushProjectionsAndFilters(HiveInputFormat.java:430)
2015-09-15 09:41:22,613 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getRecordReader(CombineHiveInputFormat.java:587)
2015-09-15 09:41:22,613 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.rdd.HadoopRDD$$anon$1.<init>(HadoopRDD.scala:236)
2015-09-15 09:41:22,613 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:212)
2015-09-15 09:41:22,613 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:101)
2015-09-15 09:41:22,613 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
2015-09-15 09:41:22,614 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
2015-09-15 09:41:22,614 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
2015-09-15 09:41:22,614 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
2015-09-15 09:41:22,614 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
2015-09-15 09:41:22,614 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
2015-09-15 09:41:22,614 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
2015-09-15 09:41:22,615 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.scheduler.Task.run(Task.scala:64)
2015-09-15 09:41:22,615 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:203)
2015-09-15 09:41:22,615 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
2015-09-15 09:41:22,615 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
2015-09-15 09:41:22,615 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
java.lang.Thread.run(Thread.java:745)
2015-09-15 09:41:22,615 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 
2015-09-15 09:41:22,618 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:22 INFO 
scheduler.TaskSetManager: Starting task 0.1 in stage 0.0 (TID 1, 
10-140-110-155, RACK_LOCAL, 1515 bytes)
2015-09-15 09:41:22,784 INFO  [main]: status.SparkJobMonitor 
(SessionState.java:printInfo(951)) - 2015-09-15 09:41:22,784       Stage-0_0: 
0(+1,-1)/1   Stage-1_0: 0/1  
2015-09-15 09:41:23,436 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:23 INFO 
storage.BlockManagerInfo: Added broadcast_1_piece0 in memory on 
10-140-110-155:42408 (size: 49.0 KB, free: 1060.2 MB)
2015-09-15 09:41:23,954 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:23 INFO 
storage.BlockManagerInfo: Added broadcast_0_piece0 in memory on 
10-140-110-155:42408 (size: 45.7 KB, free: 1060.2 MB)
2015-09-15 09:41:24,377 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:24 INFO 
scheduler.TaskSetManager: Lost task 0.1 in stage 0.0 (TID 1) on executor 
10-140-110-155: java.lang.NullPointerException (null) [duplicate 1]
2015-09-15 09:41:24,380 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:24 INFO 
scheduler.TaskSetManager: Starting task 0.2 in stage 0.0 (TID 2, 
10-140-110-155, RACK_LOCAL, 1515 bytes)
2015-09-15 09:41:24,426 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:24 INFO 
scheduler.TaskSetManager: Lost task 0.2 in stage 0.0 (TID 2) on executor 
10-140-110-155: java.lang.NullPointerException (null) [duplicate 2]
2015-09-15 09:41:24,428 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:24 INFO 
scheduler.TaskSetManager: Starting task 0.3 in stage 0.0 (TID 3, 
10-140-110-155, RACK_LOCAL, 1515 bytes)
2015-09-15 09:41:24,467 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:24 INFO 
scheduler.TaskSetManager: Lost task 0.3 in stage 0.0 (TID 3) on executor 
10-140-110-155: java.lang.NullPointerException (null) [duplicate 3]
2015-09-15 09:41:24,468 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:24 ERROR 
scheduler.TaskSetManager: Task 0 in stage 0.0 failed 4 times; aborting job
2015-09-15 09:41:24,471 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:24 INFO cluster.YarnScheduler: 
Removed TaskSet 0.0, whose tasks have all completed, from pool 
2015-09-15 09:41:24,475 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:24 INFO cluster.YarnScheduler: 
Cancelling stage 0
2015-09-15 09:41:24,477 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:24 INFO 
scheduler.DAGScheduler: Stage 0 (mapPartitionsToPair at MapTran.java:31) failed 
in 3.945 s
2015-09-15 09:41:24,506 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 15/09/15 09:41:24 INFO client.RemoteDriver: 
Failed to run job 0368cc5b-dd5d-4c7c-b48e-636d53ed350b
2015-09-15 09:41:24,506 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - java.util.concurrent.ExecutionException: 
Exception thrown by job
2015-09-15 09:41:24,506 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.JavaFutureActionWrapper.getImpl(FutureAction.scala:311)
2015-09-15 09:41:24,506 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.JavaFutureActionWrapper.get(FutureAction.scala:316)
2015-09-15 09:41:24,506 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.hive.spark.client.RemoteDriver$JobWrapper.call(RemoteDriver.java:382)
2015-09-15 09:41:24,506 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.hive.spark.client.RemoteDriver$JobWrapper.call(RemoteDriver.java:335)
2015-09-15 09:41:24,506 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
java.util.concurrent.FutureTask.run(FutureTask.java:262)
2015-09-15 09:41:24,506 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
2015-09-15 09:41:24,507 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
2015-09-15 09:41:24,507 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
java.lang.Thread.run(Thread.java:745)
2015-09-15 09:41:24,507 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - Caused by: org.apache.spark.SparkException: 
Job aborted due to stage failure: Task 0 in stage 0.0 failed 4 times, most 
recent failure: Lost task 0.3 in stage 0.0 (TID 3, 10-140-110-155): 
java.lang.NullPointerException
2015-09-15 09:41:24,507 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.hadoop.hive.ql.io.HiveInputFormat.init(HiveInputFormat.java:255)
2015-09-15 09:41:24,507 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.hadoop.hive.ql.io.HiveInputFormat.pushProjectionsAndFilters(HiveInputFormat.java:437)
2015-09-15 09:41:24,507 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.hadoop.hive.ql.io.HiveInputFormat.pushProjectionsAndFilters(HiveInputFormat.java:430)
2015-09-15 09:41:24,507 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getRecordReader(CombineHiveInputFormat.java:587)
2015-09-15 09:41:24,508 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.rdd.HadoopRDD$$anon$1.<init>(HadoopRDD.scala:236)
2015-09-15 09:41:24,508 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:212)
2015-09-15 09:41:24,508 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:101)
2015-09-15 09:41:24,508 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
2015-09-15 09:41:24,508 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
2015-09-15 09:41:24,508 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
2015-09-15 09:41:24,509 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
2015-09-15 09:41:24,509 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
2015-09-15 09:41:24,509 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
2015-09-15 09:41:24,509 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
2015-09-15 09:41:24,509 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.scheduler.Task.run(Task.scala:64)
2015-09-15 09:41:24,509 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:203)
2015-09-15 09:41:24,510 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
2015-09-15 09:41:24,510 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
2015-09-15 09:41:24,510 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
java.lang.Thread.run(Thread.java:745)
2015-09-15 09:41:24,510 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - 
2015-09-15 09:41:24,510 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) - Driver stacktrace:
2015-09-15 09:41:24,510 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1204)
2015-09-15 09:41:24,511 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1193)
2015-09-15 09:41:24,511 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1192)
2015-09-15 09:41:24,511 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
2015-09-15 09:41:24,514 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
2015-09-15 09:41:24,514 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1192)
2015-09-15 09:41:24,514 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:693)
2015-09-15 09:41:24,514 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:693)
2015-09-15 09:41:24,515 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
scala.Option.foreach(Option.scala:236)
2015-09-15 09:41:24,515 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:693)
2015-09-15 09:41:24,515 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1393)
2015-09-15 09:41:24,515 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1354)
2015-09-15 09:41:24,515 INFO  [stderr-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(569)) -        at 
org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
2015-09-15 09:41:24,540 INFO  [RPC-Handler-3]: client.SparkClientImpl 
(SparkClientImpl.java:handle(522)) - Received result for 
0368cc5b-dd5d-4c7c-b48e-636d53ed350b
2015-09-15 09:41:24,795 ERROR [main]: status.SparkJobMonitor 
(SessionState.java:printError(960)) - Status: Failed
2015-09-15 09:41:24,795 INFO  [main]: log.PerfLogger 
(PerfLogger.java:PerfLogEnd(148)) - </PERFLOG method=SparkRunJob 
start=1442281254723 end=1442281284795 duration=30072 
from=org.apache.hadoop.hive.ql.exec.spark.status.SparkJobMonitor>
2015-09-15 09:41:24,817 ERROR [main]: ql.Driver 
(SessionState.java:printError(960)) - FAILED: Execution Error, return code 3 
from org.apache.hadoop.hive.ql.exec.spark.SparkTask
2015-09-15 09:41:24,817 INFO  [main]: log.PerfLogger 
(PerfLogger.java:PerfLogEnd(148)) - </PERFLOG method=Driver.execute 
start=1442281249766 end=1442281284817 duration=35051 
from=org.apache.hadoop.hive.ql.Driver>
2015-09-15 09:41:24,817 INFO  [main]: log.PerfLogger 
(PerfLogger.java:PerfLogBegin(121)) - <PERFLOG method=releaseLocks 
from=org.apache.hadoop.hive.ql.Driver>
2015-09-15 09:41:24,817 INFO  [main]: log.PerfLogger 
(PerfLogger.java:PerfLogEnd(148)) - </PERFLOG method=releaseLocks 
start=1442281284817 end=1442281284817 duration=0 
from=org.apache.hadoop.hive.ql.Driver>
2015-09-15 09:41:24,819 INFO  [main]: exec.ListSinkOperator 
(Operator.java:close(612)) - 7 finished. closing... 
2015-09-15 09:41:24,819 INFO  [main]: exec.ListSinkOperator 
(Operator.java:close(634)) - 7 Close done
2015-09-15 09:41:24,826 INFO  [main]: log.PerfLogger 
(PerfLogger.java:PerfLogBegin(121)) - <PERFLOG method=releaseLocks 
from=org.apache.hadoop.hive.ql.Driver>
2015-09-15 09:41:24,826 INFO  [main]: log.PerfLogger 
(PerfLogger.java:PerfLogEnd(148)) - </PERFLOG method=releaseLocks 
start=1442281284826 end=1442281284826 duration=0 
from=org.apache.hadoop.hive.ql.Driver>






Thanks again.It seems that the spark job is already successfully running on the 
yarn from hive,but the explain job map.xml is lost .




> Hive on spark issue..
> ---------------------
>
>                 Key: HIVE-7980
>                 URL: https://issues.apache.org/jira/browse/HIVE-7980
>             Project: Hive
>          Issue Type: Bug
>          Components: HiveServer2, Spark
>    Affects Versions: spark-branch
>         Environment: Test Environment is..
> . hive 0.14.0(spark branch version)
> . spark 
> (http://ec2-50-18-79-139.us-west-1.compute.amazonaws.com/data/spark-assembly-1.1.0-SNAPSHOT-hadoop2.3.0.jar)
> . hadoop 2.4.0 (yarn)
>            Reporter: alton.jung
>            Assignee: Chao Sun
>             Fix For: spark-branch
>
>
> .I followed this 
> guide(https://cwiki.apache.org/confluence/display/Hive/Hive+on+Spark%3A+Getting+Started).
>  and i compiled hive from spark branch. in the next step i met the below 
> error..
> (*i typed the hive query on beeline, i used the  simple query using "order 
> by" to invoke the palleral works 
>                    ex) select * from test where id = 1 order by id;
> )
> [Error list is]
> 2014-09-04 02:58:08,796 ERROR spark.SparkClient 
> (SparkClient.java:execute(158)) - Error generating Spark Plan
> java.lang.NullPointerException
>       at 
> org.apache.spark.SparkContext.defaultParallelism(SparkContext.scala:1262)
>       at 
> org.apache.spark.SparkContext.defaultMinPartitions(SparkContext.scala:1269)
>       at 
> org.apache.spark.SparkContext.hadoopRDD$default$5(SparkContext.scala:537)
>       at 
> org.apache.spark.api.java.JavaSparkContext.hadoopRDD(JavaSparkContext.scala:318)
>       at 
> org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generateRDD(SparkPlanGenerator.java:160)
>       at 
> org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generate(SparkPlanGenerator.java:88)
>       at 
> org.apache.hadoop.hive.ql.exec.spark.SparkClient.execute(SparkClient.java:156)
>       at 
> org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.submit(SparkSessionImpl.java:52)
>       at 
> org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:77)
>       at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:161)
>       at 
> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
>       at org.apache.hadoop.hive.ql.exec.TaskRunner.run(TaskRunner.java:72)
> 2014-09-04 02:58:11,108 ERROR ql.Driver (SessionState.java:printError(696)) - 
> FAILED: Execution Error, return code 2 from 
> org.apache.hadoop.hive.ql.exec.spark.SparkTask
> 2014-09-04 02:58:11,182 INFO  log.PerfLogger 
> (PerfLogger.java:PerfLogEnd(135)) - </PERFLOG method=Driver.execute 
> start=1409824527954 end=1409824691182 duration=163228 
> from=org.apache.hadoop.hive.ql.Driver>
> 2014-09-04 02:58:11,223 INFO  log.PerfLogger 
> (PerfLogger.java:PerfLogBegin(108)) - <PERFLOG method=releaseLocks 
> from=org.apache.hadoop.hive.ql.Driver>
> 2014-09-04 02:58:11,224 INFO  log.PerfLogger 
> (PerfLogger.java:PerfLogEnd(135)) - </PERFLOG method=releaseLocks 
> start=1409824691223 end=1409824691224 duration=1 
> from=org.apache.hadoop.hive.ql.Driver>
> 2014-09-04 02:58:11,306 ERROR operation.Operation 
> (SQLOperation.java:run(199)) - Error running hive query: 
> org.apache.hive.service.cli.HiveSQLException: Error while processing 
> statement: FAILED: Execution Error, return code 2 from 
> org.apache.hadoop.hive.ql.exec.spark.SparkTask
>       at 
> org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:284)
>       at 
> org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:146)
>       at 
> org.apache.hive.service.cli.operation.SQLOperation.access$100(SQLOperation.java:69)
>       at 
> org.apache.hive.service.cli.operation.SQLOperation$1$1.run(SQLOperation.java:196)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:415)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1556)
>       at 
> org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:508)
>       at 
> org.apache.hive.service.cli.operation.SQLOperation$1.run(SQLOperation.java:208)
>       at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
>       at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:166)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:722)
> 2014-09-04 02:58:11,634 INFO  exec.ListSinkOperator 
> (Operator.java:close(580)) - 47 finished. closing... 
> 2014-09-04 02:58:11,683 INFO  exec.ListSinkOperator 
> (Operator.java:close(598)) - 47 Close done
> 2014-09-04 02:58:12,190 INFO  log.PerfLogger 
> (PerfLogger.java:PerfLogBegin(108)) - <PERFLOG method=releaseLocks 
> from=org.apache.hadoop.hive.ql.Driver>
> 2014-09-04 02:58:12,234 INFO  log.PerfLogger 
> (PerfLogger.java:PerfLogEnd(135)) - </PERFLOG method=releaseLocks 
> start=1409824692190 end=1409824692191 duration=1 
> from=org.apache.hadoop.hive.ql.Driver>



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to