[ https://issues.apache.org/jira/browse/SPARK-25903?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16933335#comment-16933335 ]
Jungtaek Lim commented on SPARK-25903: -------------------------------------- Hit this again: [https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/110979/testReport/] > Flaky test: BarrierTaskContextSuite.throw exception on barrier() call timeout > ----------------------------------------------------------------------------- > > Key: SPARK-25903 > URL: https://issues.apache.org/jira/browse/SPARK-25903 > Project: Spark > Issue Type: Bug > Components: Tests > Affects Versions: 2.4.0, 3.0.0 > Reporter: Marcelo Vanzin > Priority: Major > > We hit this in our internal builds. > {noformat} > Expected exception org.apache.spark.SparkException to be thrown, but no > exception was thrown > Stacktrace > org.scalatest.exceptions.TestFailedException: Expected exception > org.apache.spark.SparkException to be thrown, but no exception was thrown > at > org.scalatest.Assertions$class.newAssertionFailedException(Assertions.scala:528) > at > org.scalatest.FunSuite.newAssertionFailedException(FunSuite.scala:1560) > at org.scalatest.Assertions$class.intercept(Assertions.scala:822) > at org.scalatest.FunSuite.intercept(FunSuite.scala:1560) > at > org.apache.spark.scheduler.BarrierTaskContextSuite$$anonfun$7.apply(BarrierTaskContextSuite.scala:94) > at > org.apache.spark.scheduler.BarrierTaskContextSuite$$anonfun$7.apply(BarrierTaskContextSuite.scala:76) > at org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85) > {noformat} > The problem from the logs is that the first task to call {{barrier()}} took a > while, and at that point the "slow" task was already running, so the sleep > finishes before the 2s timeout runs out. > - > https://amplab.cs.berkeley.edu/jenkins/view/Spark%20QA%20Test%20(Dashboard)/job/spark-master-test-maven-hadoop-2.7-ubuntu-scala-2.11/169/ -- This message was sent by Atlassian Jira (v8.3.4#803005) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org