[jira] [Assigned] (SPARK-23307) Spark UI should sort jobs/stages with the completed timestamp before cleaning up them

2018-02-01 Thread Apache Spark (JIRA)

 [ 
https://issues.apache.org/jira/browse/SPARK-23307?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-23307:


Assignee: Shixiong Zhu  (was: Apache Spark)

> Spark UI should sort jobs/stages with the completed timestamp before cleaning 
> up them
> -
>
> Key: SPARK-23307
> URL: https://issues.apache.org/jira/browse/SPARK-23307
> Project: Spark
>  Issue Type: Bug
>  Components: Web UI
>Affects Versions: 2.3.0
>Reporter: Shixiong Zhu
>Assignee: Shixiong Zhu
>Priority: Blocker
>
> When you have a long running job, it may be deleted from UI quickly when it 
> completes, if you happen to run a small job after it. It's pretty annoying 
> when you run lots of jobs in the same driver concurrently (e.g., running 
> multiple Structured Streaming queries). We should sort jobs/stages with the 
> completed timestamp before cleaning up them.
> In 2.2, Spark has a separated buffer for completed jobs/stages, so it doesn't 
> need to sort the jobs/stages.
> What's the behavior I expect:
> Set "spark.ui.retainedJobs" to 10 and run the following codes, job 0 should 
> be kept in the Spark UI.
>  
> {code:java}
> new Thread() {
>   override def run() {
>     // job 0
>     sc.makeRDD(1 to 1, 1).foreach { i =>
>     Thread.sleep(1)
>    }
>   }
> }.start()
> Thread.sleep(1000)
> for (_ <- 1 to 20) {
>   new Thread() {
>     override def run() {
>       sc.makeRDD(1 to 1, 1).foreach { i =>
>       }
>     }
>   }.start()
> }
> Thread.sleep(15000)
>   sc.makeRDD(1 to 1, 1).foreach { i =>
> }
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-23307) Spark UI should sort jobs/stages with the completed timestamp before cleaning up them

2018-02-01 Thread Apache Spark (JIRA)

 [ 
https://issues.apache.org/jira/browse/SPARK-23307?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-23307:


Assignee: Apache Spark  (was: Shixiong Zhu)

> Spark UI should sort jobs/stages with the completed timestamp before cleaning 
> up them
> -
>
> Key: SPARK-23307
> URL: https://issues.apache.org/jira/browse/SPARK-23307
> Project: Spark
>  Issue Type: Bug
>  Components: Web UI
>Affects Versions: 2.3.0
>Reporter: Shixiong Zhu
>Assignee: Apache Spark
>Priority: Blocker
>
> When you have a long running job, it may be deleted from UI quickly when it 
> completes, if you happen to run a small job after it. It's pretty annoying 
> when you run lots of jobs in the same driver concurrently (e.g., running 
> multiple Structured Streaming queries). We should sort jobs/stages with the 
> completed timestamp before cleaning up them.
> In 2.2, Spark has a separated buffer for completed jobs/stages, so it doesn't 
> need to sort the jobs/stages.
> What's the behavior I expect:
> Set "spark.ui.retainedJobs" to 10 and run the following codes, job 0 should 
> be kept in the Spark UI.
>  
> {code:java}
> new Thread() {
>   override def run() {
>     // job 0
>     sc.makeRDD(1 to 1, 1).foreach { i =>
>     Thread.sleep(1)
>    }
>   }
> }.start()
> Thread.sleep(1000)
> for (_ <- 1 to 20) {
>   new Thread() {
>     override def run() {
>       sc.makeRDD(1 to 1, 1).foreach { i =>
>       }
>     }
>   }.start()
> }
> Thread.sleep(15000)
>   sc.makeRDD(1 to 1, 1).foreach { i =>
> }
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-23307) Spark UI should sort jobs/stages with the completed timestamp before cleaning up them

2018-02-01 Thread Shixiong Zhu (JIRA)

 [ 
https://issues.apache.org/jira/browse/SPARK-23307?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shixiong Zhu reassigned SPARK-23307:


Assignee: Shixiong Zhu

> Spark UI should sort jobs/stages with the completed timestamp before cleaning 
> up them
> -
>
> Key: SPARK-23307
> URL: https://issues.apache.org/jira/browse/SPARK-23307
> Project: Spark
>  Issue Type: Bug
>  Components: Web UI
>Affects Versions: 2.3.0
>Reporter: Shixiong Zhu
>Assignee: Shixiong Zhu
>Priority: Major
>
> When you have a long running job, it may be deleted from UI quickly when it 
> completes, if you happen to run a small job after it. It's pretty annoying 
> when you run lots of jobs in the same driver concurrently (e.g., running 
> multiple Structured Streaming queries). We should sort jobs/stages with the 
> completed timestamp before cleaning up them.
> In 2.2, Spark has a separated buffer for completed jobs/stages, so it doesn't 
> need to sort the jobs/stages.
> What's the behavior I expect:
> Set "spark.ui.retainedJobs" to 10 and run the following codes, job 0 should 
> be kept in the Spark UI.
>  
> {code:java}
> new Thread() {
>   override def run() {
>     // job 0
>     sc.makeRDD(1 to 1, 1).foreach { i =>
>     Thread.sleep(1)
>    }
>   }
> }.start()
> Thread.sleep(1000)
> for (_ <- 1 to 20) {
>   new Thread() {
>     override def run() {
>       sc.makeRDD(1 to 1, 1).foreach { i =>
>       }
>     }
>   }.start()
> }
> Thread.sleep(15000)
>   sc.makeRDD(1 to 1, 1).foreach { i =>
> }
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org