[
https://issues.apache.org/jira/browse/HIVE-9258?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Jimmy Xiang resolved HIVE-9258.
-------------------------------
Resolution: Not a Problem
Closed it as Not a Problem. Thanks. As to sparkMemoryAndCores in
SetSparkReducerParallelism, it is used only during one query. In one query,
there could be several reducers so it's better cached. Since we create a
SetSparkReducerParallelism instance per query, so it should not be cached in
the entire user session, I think.
> Explain query should share the same Spark application with regular queries
> [Spark Branch]
> -----------------------------------------------------------------------------------------
>
> Key: HIVE-9258
> URL: https://issues.apache.org/jira/browse/HIVE-9258
> Project: Hive
> Issue Type: Sub-task
> Components: Spark
> Reporter: Xuefu Zhang
> Assignee: Jimmy Xiang
>
> Currently for Hive on Spark, query plan includes the number of reducers,
> which is determined partly by the Spark cluster. Thus, explain query will
> need to launch a Spark application (Spark remote context), which should be
> shared with regular queries so that we don't launch additional Spark remote
> context.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)