HIVE-10476 - Hive query should fail when it fails to initialize a session in 
SetSparkReducerParallelism [Spark Branch] (Chao, reviewed by Rui)


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/cee1183d
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/cee1183d
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/cee1183d

Branch: refs/heads/hbase-metastore
Commit: cee1183d3044c4fa0a7882c3499d59721f1b5492
Parents: 88dfd27
Author: Chao Sun <sunc...@apache.org>
Authored: Tue Apr 28 09:39:18 2015 -0700
Committer: Xuefu Zhang <xzh...@cloudera.com>
Committed: Mon Jun 1 14:03:09 2015 -0700

----------------------------------------------------------------------
 .../hadoop/hive/ql/optimizer/spark/SetSparkReducerParallelism.java | 2 ++
 1 file changed, 2 insertions(+)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hive/blob/cee1183d/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SetSparkReducerParallelism.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SetSparkReducerParallelism.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SetSparkReducerParallelism.java
index dcfbc6b..f9ef474 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SetSparkReducerParallelism.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SetSparkReducerParallelism.java
@@ -115,6 +115,8 @@ public class SetSparkReducerParallelism implements 
NodeProcessor {
             sparkSession = SparkUtilities.getSparkSession(
               context.getConf(), sparkSessionManager);
             sparkMemoryAndCores = sparkSession.getMemoryAndCores();
+          } catch (HiveException e) {
+            throw new SemanticException("Failed to get a spark session: " + e);
           } catch (Exception e) {
             LOG.warn("Failed to get spark memory/core info", e);
           } finally {

Reply via email to