[jira] [Assigned] (SPARK-23307) Spark UI should sort jobs/stages with the completed timestamp before cleaning up them
[ https://issues.apache.org/jira/browse/SPARK-23307?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Apache Spark reassigned SPARK-23307: Assignee: Shixiong Zhu (was: Apache Spark) > Spark UI should sort jobs/stages with the completed timestamp before cleaning > up them > - > > Key: SPARK-23307 > URL: https://issues.apache.org/jira/browse/SPARK-23307 > Project: Spark > Issue Type: Bug > Components: Web UI >Affects Versions: 2.3.0 >Reporter: Shixiong Zhu >Assignee: Shixiong Zhu >Priority: Blocker > > When you have a long running job, it may be deleted from UI quickly when it > completes, if you happen to run a small job after it. It's pretty annoying > when you run lots of jobs in the same driver concurrently (e.g., running > multiple Structured Streaming queries). We should sort jobs/stages with the > completed timestamp before cleaning up them. > In 2.2, Spark has a separated buffer for completed jobs/stages, so it doesn't > need to sort the jobs/stages. > What's the behavior I expect: > Set "spark.ui.retainedJobs" to 10 and run the following codes, job 0 should > be kept in the Spark UI. > > {code:java} > new Thread() { > override def run() { > // job 0 > sc.makeRDD(1 to 1, 1).foreach { i => > Thread.sleep(1) > } > } > }.start() > Thread.sleep(1000) > for (_ <- 1 to 20) { > new Thread() { > override def run() { > sc.makeRDD(1 to 1, 1).foreach { i => > } > } > }.start() > } > Thread.sleep(15000) > sc.makeRDD(1 to 1, 1).foreach { i => > } > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-23307) Spark UI should sort jobs/stages with the completed timestamp before cleaning up them
[ https://issues.apache.org/jira/browse/SPARK-23307?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Apache Spark reassigned SPARK-23307: Assignee: Apache Spark (was: Shixiong Zhu) > Spark UI should sort jobs/stages with the completed timestamp before cleaning > up them > - > > Key: SPARK-23307 > URL: https://issues.apache.org/jira/browse/SPARK-23307 > Project: Spark > Issue Type: Bug > Components: Web UI >Affects Versions: 2.3.0 >Reporter: Shixiong Zhu >Assignee: Apache Spark >Priority: Blocker > > When you have a long running job, it may be deleted from UI quickly when it > completes, if you happen to run a small job after it. It's pretty annoying > when you run lots of jobs in the same driver concurrently (e.g., running > multiple Structured Streaming queries). We should sort jobs/stages with the > completed timestamp before cleaning up them. > In 2.2, Spark has a separated buffer for completed jobs/stages, so it doesn't > need to sort the jobs/stages. > What's the behavior I expect: > Set "spark.ui.retainedJobs" to 10 and run the following codes, job 0 should > be kept in the Spark UI. > > {code:java} > new Thread() { > override def run() { > // job 0 > sc.makeRDD(1 to 1, 1).foreach { i => > Thread.sleep(1) > } > } > }.start() > Thread.sleep(1000) > for (_ <- 1 to 20) { > new Thread() { > override def run() { > sc.makeRDD(1 to 1, 1).foreach { i => > } > } > }.start() > } > Thread.sleep(15000) > sc.makeRDD(1 to 1, 1).foreach { i => > } > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-23307) Spark UI should sort jobs/stages with the completed timestamp before cleaning up them
[ https://issues.apache.org/jira/browse/SPARK-23307?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Shixiong Zhu reassigned SPARK-23307: Assignee: Shixiong Zhu > Spark UI should sort jobs/stages with the completed timestamp before cleaning > up them > - > > Key: SPARK-23307 > URL: https://issues.apache.org/jira/browse/SPARK-23307 > Project: Spark > Issue Type: Bug > Components: Web UI >Affects Versions: 2.3.0 >Reporter: Shixiong Zhu >Assignee: Shixiong Zhu >Priority: Major > > When you have a long running job, it may be deleted from UI quickly when it > completes, if you happen to run a small job after it. It's pretty annoying > when you run lots of jobs in the same driver concurrently (e.g., running > multiple Structured Streaming queries). We should sort jobs/stages with the > completed timestamp before cleaning up them. > In 2.2, Spark has a separated buffer for completed jobs/stages, so it doesn't > need to sort the jobs/stages. > What's the behavior I expect: > Set "spark.ui.retainedJobs" to 10 and run the following codes, job 0 should > be kept in the Spark UI. > > {code:java} > new Thread() { > override def run() { > // job 0 > sc.makeRDD(1 to 1, 1).foreach { i => > Thread.sleep(1) > } > } > }.start() > Thread.sleep(1000) > for (_ <- 1 to 20) { > new Thread() { > override def run() { > sc.makeRDD(1 to 1, 1).foreach { i => > } > } > }.start() > } > Thread.sleep(15000) > sc.makeRDD(1 to 1, 1).foreach { i => > } > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org