[jira] [Assigned] (SPARK-20562) Support Maintenance by having a threshold for unavailability

2017-05-03 Thread Apache Spark (JIRA)

 [ 
https://issues.apache.org/jira/browse/SPARK-20562?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-20562:


Assignee: (was: Apache Spark)

> Support Maintenance by having a threshold for unavailability
> 
>
> Key: SPARK-20562
> URL: https://issues.apache.org/jira/browse/SPARK-20562
> Project: Spark
>  Issue Type: Improvement
>  Components: Mesos
>Affects Versions: 2.1.0
>Reporter: Kamal Gurala
>
> Make Spark be aware of offers that have an unavailability period set because 
> of a scheduled Maintenance on the node.
> Have a configurable option that's a threshold which ensures that tasks are 
> not scheduled on offers that are within a threshold for maintenance



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-20562) Support Maintenance by having a threshold for unavailability

2017-05-03 Thread Apache Spark (JIRA)

 [ 
https://issues.apache.org/jira/browse/SPARK-20562?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-20562:


Assignee: Apache Spark

> Support Maintenance by having a threshold for unavailability
> 
>
> Key: SPARK-20562
> URL: https://issues.apache.org/jira/browse/SPARK-20562
> Project: Spark
>  Issue Type: Improvement
>  Components: Mesos
>Affects Versions: 2.1.0
>Reporter: Kamal Gurala
>Assignee: Apache Spark
>
> Make Spark be aware of offers that have an unavailability period set because 
> of a scheduled Maintenance on the node.
> Have a configurable option that's a threshold which ensures that tasks are 
> not scheduled on offers that are within a threshold for maintenance



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org