[jira] [Assigned] (SPARK-19658) Set NumPartitions of RepartitionByExpression In Analyzer

2017-02-18 Thread Apache Spark (JIRA)

 [ 
https://issues.apache.org/jira/browse/SPARK-19658?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-19658:


Assignee: Apache Spark  (was: Xiao Li)

> Set NumPartitions of RepartitionByExpression In Analyzer
> 
>
> Key: SPARK-19658
> URL: https://issues.apache.org/jira/browse/SPARK-19658
> Project: Spark
>  Issue Type: Improvement
>  Components: SQL
>Affects Versions: 2.1.0
>Reporter: Xiao Li
>Assignee: Apache Spark
>
> Currently, if {{NumPartitions}} is not set, we will set it using 
> `spark.sql.shuffle.partitions` in Planner. However, this is not following 
> general resolution process. We should do it in Analyzer and then Optimizer 
> can use the value for optimization. 



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-19658) Set NumPartitions of RepartitionByExpression In Analyzer

2017-02-18 Thread Apache Spark (JIRA)

 [ 
https://issues.apache.org/jira/browse/SPARK-19658?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-19658:


Assignee: Xiao Li  (was: Apache Spark)

> Set NumPartitions of RepartitionByExpression In Analyzer
> 
>
> Key: SPARK-19658
> URL: https://issues.apache.org/jira/browse/SPARK-19658
> Project: Spark
>  Issue Type: Improvement
>  Components: SQL
>Affects Versions: 2.1.0
>Reporter: Xiao Li
>Assignee: Xiao Li
>
> Currently, if {{NumPartitions}} is not set, we will set it using 
> `spark.sql.shuffle.partitions` in Planner. However, this is not following 
> general resolution process. We should do it in Analyzer and then Optimizer 
> can use the value for optimization. 



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org