Agree that frequent topic deletion is not a very Kafka-esque thing to do On Fri, Dec 9, 2016 at 12:09 PM, Shixiong(Ryan) Zhu <shixi...@databricks.com> wrote: > Sean, "stress test for failOnDataLoss=false" is because Kafka consumer may > be thrown NPE when a topic is deleted. I added some logic to retry on such > failure, however, it may still fail when topic deletion is too frequent (the > stress test). Just reopened > https://issues.apache.org/jira/browse/SPARK-18588. > > Anyway, this is just a best effort to deal with Kafka issue, and in > practice, people won't delete topic frequently, so this is not a release > blocker. > > On Fri, Dec 9, 2016 at 2:55 AM, Sean Owen <so...@cloudera.com> wrote: >> >> As usual, the sigs / hashes are fine and licenses look fine. >> >> I am still seeing some test failures. A few I've seen over time and aren't >> repeatable, but a few seem persistent. ANyone else observed these? I'm on >> Ubuntu 16 / Java 8 building for -Pyarn -Phadoop-2.7 -Phive >> >> If anyone can confirm I'll investigate the cause if I can. I'd hesitate to >> support the release yet unless the build is definitely passing for others. >> >> >> udf3Test(test.org.apache.spark.sql.JavaUDFSuite) Time elapsed: 0.281 sec >> <<< ERROR! >> java.lang.NoSuchMethodError: >> org.apache.spark.sql.catalyst.JavaTypeInference$.inferDataType(Lcom/google/common/reflect/TypeToken;)Lscala/Tuple2; >> at test.org.apache.spark.sql.JavaUDFSuite.udf3Test(JavaUDFSuite.java:107) >> >> >> >> - caching on disk *** FAILED *** >> java.util.concurrent.TimeoutException: Can't find 2 executors before >> 30000 milliseconds elapsed >> at >> org.apache.spark.ui.jobs.JobProgressListener.waitUntilExecutorsUp(JobProgressListener.scala:584) >> at >> org.apache.spark.DistributedSuite.org$apache$spark$DistributedSuite$$testCaching(DistributedSuite.scala:154) >> at >> org.apache.spark.DistributedSuite$$anonfun$32$$anonfun$apply$1.apply$mcV$sp(DistributedSuite.scala:191) >> at >> org.apache.spark.DistributedSuite$$anonfun$32$$anonfun$apply$1.apply(DistributedSuite.scala:191) >> at >> org.apache.spark.DistributedSuite$$anonfun$32$$anonfun$apply$1.apply(DistributedSuite.scala:191) >> at >> org.scalatest.Transformer$$anonfun$apply$1.apply$mcV$sp(Transformer.scala:22) >> at org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85) >> at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104) >> at org.scalatest.Transformer.apply(Transformer.scala:22) >> at org.scalatest.Transformer.apply(Transformer.scala:20) >> ... >> >> >> - stress test for failOnDataLoss=false *** FAILED *** >> org.apache.spark.sql.streaming.StreamingQueryException: Query [id = >> 3b191b78-7f30-46d3-93f8-5fbeecce94a2, runId = >> 0cab93b6-19d8-47a7-88ad-d296bea72405] terminated with exception: null >> at >> org.apache.spark.sql.execution.streaming.StreamExecution.org$apache$spark$sql$execution$streaming$StreamExecution$$runBatches(StreamExecution.scala:262) >> at >> org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.run(StreamExecution.scala:160) >> ... >> Cause: java.lang.NullPointerException: >> ... >> >> >> >> On Thu, Dec 8, 2016 at 4:40 PM Reynold Xin <r...@databricks.com> wrote: >>> >>> Please vote on releasing the following candidate as Apache Spark version >>> 2.1.0. The vote is open until Sun, December 11, 2016 at 1:00 PT and passes >>> if a majority of at least 3 +1 PMC votes are cast. >>> >>> [ ] +1 Release this package as Apache Spark 2.1.0 >>> [ ] -1 Do not release this package because ... >>> >>> >>> To learn more about Apache Spark, please see http://spark.apache.org/ >>> >>> The tag to be voted on is v2.1.0-rc2 >>> (080717497365b83bc202ab16812ced93eb1ea7bd) >>> >>> List of JIRA tickets resolved are: >>> https://issues.apache.org/jira/issues/?jql=project%20%3D%20SPARK%20AND%20fixVersion%20%3D%202.1.0 >>> >>> The release files, including signatures, digests, etc. can be found at: >>> http://people.apache.org/~pwendell/spark-releases/spark-2.1.0-rc2-bin/ >>> >>> Release artifacts are signed with the following key: >>> https://people.apache.org/keys/committer/pwendell.asc >>> >>> The staging repository for this release can be found at: >>> https://repository.apache.org/content/repositories/orgapachespark-1217 >>> >>> The documentation corresponding to this release can be found at: >>> http://people.apache.org/~pwendell/spark-releases/spark-2.1.0-rc2-docs/ >>> >>> >>> (Note that the docs and staging repo are still being uploaded and will be >>> available soon) >>> >>> >>> ======================================= >>> How can I help test this release? >>> ======================================= >>> If you are a Spark user, you can help us test this release by taking an >>> existing Spark workload and running on this release candidate, then >>> reporting any regressions. >>> >>> =============================================================== >>> What should happen to JIRA tickets still targeting 2.1.0? >>> =============================================================== >>> Committers should look at those and triage. Extremely important bug >>> fixes, documentation, and API tweaks that impact compatibility should be >>> worked on immediately. Everything else please retarget to 2.1.1 or 2.2.0. > >
--------------------------------------------------------------------- To unsubscribe e-mail: dev-unsubscr...@spark.apache.org