This vote passes with nine +1s (five binding) and one binding +0! Thanks to everyone who tested/voted. I'll start work on publishing the release today.
+1: Mark Hamstra* Moshe Eshel Egor Pahomov Reynold Xin* Yin Huai* Andrew Or* Burak Yavuz Kousuke Saruta Michael Armbrust* 0: Sean Owen* -1: (none) *Binding On Wed, Mar 9, 2016 at 3:29 PM, Michael Armbrust <mich...@databricks.com> wrote: > +1 - Ported all our internal jobs to run on 1.6.1 with no regressions. > > On Wed, Mar 9, 2016 at 7:04 AM, Kousuke Saruta <saru...@oss.nttdata.co.jp> > wrote: > >> +1 (non-binding) >> >> >> On 2016/03/09 4:28, Burak Yavuz wrote: >> >> +1 >> >> On Tue, Mar 8, 2016 at 10:59 AM, Andrew Or <and...@databricks.com> wrote: >> >>> +1 >>> >>> 2016-03-08 10:59 GMT-08:00 Yin Huai < <yh...@databricks.com> >>> yh...@databricks.com>: >>> >>>> +1 >>>> >>>> On Mon, Mar 7, 2016 at 12:39 PM, Reynold Xin < <r...@databricks.com> >>>> r...@databricks.com> wrote: >>>> >>>>> +1 (binding) >>>>> >>>>> >>>>> On Sun, Mar 6, 2016 at 12:08 PM, Egor Pahomov < >>>>> <pahomov.e...@gmail.com>pahomov.e...@gmail.com> wrote: >>>>> >>>>>> +1 >>>>>> >>>>>> Spark ODBC server is fine, SQL is fine. >>>>>> >>>>>> 2016-03-03 12:09 GMT-08:00 Yin Yang < <yy201...@gmail.com> >>>>>> yy201...@gmail.com>: >>>>>> >>>>>>> Skipping docker tests, the rest are green: >>>>>>> >>>>>>> [INFO] Spark Project External Kafka ....................... SUCCESS >>>>>>> [01:28 min] >>>>>>> [INFO] Spark Project Examples ............................. SUCCESS >>>>>>> [02:59 min] >>>>>>> [INFO] Spark Project External Kafka Assembly .............. SUCCESS >>>>>>> [ 11.680 s] >>>>>>> [INFO] >>>>>>> ------------------------------------------------------------------------ >>>>>>> [INFO] BUILD SUCCESS >>>>>>> [INFO] >>>>>>> ------------------------------------------------------------------------ >>>>>>> [INFO] Total time: 02:16 h >>>>>>> [INFO] Finished at: 2016-03-03T11:17:07-08:00 >>>>>>> [INFO] Final Memory: 152M/4062M >>>>>>> >>>>>>> On Thu, Mar 3, 2016 at 8:55 AM, Yin Yang < <yy201...@gmail.com> >>>>>>> yy201...@gmail.com> wrote: >>>>>>> >>>>>>>> When I ran test suite using the following command: >>>>>>>> >>>>>>>> build/mvn clean -Phive -Phive-thriftserver -Pyarn -Phadoop-2.6 >>>>>>>> -Dhadoop.version=2.7.0 package >>>>>>>> >>>>>>>> I got failure in Spark Project Docker Integration Tests : >>>>>>>> >>>>>>>> 16/03/02 17:36:46 INFO RemoteActorRefProvider$RemotingTerminator: >>>>>>>> Remote daemon shut down; proceeding with flushing remote transports. >>>>>>>> ^[[31m*** RUN ABORTED ***^[[0m >>>>>>>> ^[[31m com.spotify.docker.client.DockerException: >>>>>>>> java.util.concurrent.ExecutionException: >>>>>>>> com.spotify.docker.client.shaded.javax.ws.rs.ProcessingException: >>>>>>>> java.io. IOException: No such file or directory^[[0m >>>>>>>> ^[[31m at >>>>>>>> com.spotify.docker.client.DefaultDockerClient.propagate(DefaultDockerClient.java:1141)^[[0m >>>>>>>> ^[[31m at >>>>>>>> com.spotify.docker.client.DefaultDockerClient.request(DefaultDockerClient.java:1082)^[[0m >>>>>>>> ^[[31m at >>>>>>>> com.spotify.docker.client.DefaultDockerClient.ping(DefaultDockerClient.java:281)^[[0m >>>>>>>> ^[[31m at >>>>>>>> org.apache.spark.sql.jdbc.DockerJDBCIntegrationSuite.beforeAll(DockerJDBCIntegrationSuite.scala:76)^[[0m >>>>>>>> ^[[31m at >>>>>>>> org.scalatest.BeforeAndAfterAll$class.beforeAll(BeforeAndAfterAll.scala:187)^[[0m >>>>>>>> ^[[31m at >>>>>>>> org.apache.spark.sql.jdbc.DockerJDBCIntegrationSuite.beforeAll(DockerJDBCIntegrationSuite.scala:58)^[[0m >>>>>>>> ^[[31m at >>>>>>>> org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:253)^[[0m >>>>>>>> ^[[31m at >>>>>>>> org.apache.spark.sql.jdbc.DockerJDBCIntegrationSuite.run(DockerJDBCIntegrationSuite.scala:58)^[[0m >>>>>>>> ^[[31m at >>>>>>>> org.scalatest.Suite$class.callExecuteOnSuite$1(Suite.scala:1492)^[[0m >>>>>>>> ^[[31m at >>>>>>>> org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1528)^[[0m >>>>>>>> ^[[31m ...^[[0m >>>>>>>> ^[[31m Cause: java.util.concurrent.ExecutionException: >>>>>>>> com.spotify.docker.client.shaded.javax.ws.rs.ProcessingException: >>>>>>>> java.io.IOException: No such file or directory^[[0m >>>>>>>> ^[[31m at >>>>>>>> jersey.repackaged.com.google.common.util.concurrent.AbstractFuture$Sync.getValue(AbstractFuture.java:299)^[[0m >>>>>>>> ^[[31m at >>>>>>>> jersey.repackaged.com.google.common.util.concurrent.AbstractFuture$Sync.get(AbstractFuture.java:286)^[[0m >>>>>>>> ^[[31m at >>>>>>>> jersey.repackaged.com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:116)^[[0m >>>>>>>> ^[[31m at >>>>>>>> com.spotify.docker.client.DefaultDockerClient.request(DefaultDockerClient.java:1080)^[[0m >>>>>>>> ^[[31m at >>>>>>>> com.spotify.docker.client.DefaultDockerClient.ping(DefaultDockerClient.java:281)^[[0m >>>>>>>> ^[[31m at >>>>>>>> org.apache.spark.sql.jdbc.DockerJDBCIntegrationSuite.beforeAll(DockerJDBCIntegrationSuite.scala:76)^[[0m >>>>>>>> ^[[31m at >>>>>>>> org.scalatest.BeforeAndAfterAll$class.beforeAll(BeforeAndAfterAll.scala:187)^[[0m >>>>>>>> ^[[31m at >>>>>>>> org.apache.spark.sql.jdbc.DockerJDBCIntegrationSuite.beforeAll(DockerJDBCIntegrationSuite.scala:58)^[[0m >>>>>>>> ^[[31m at >>>>>>>> org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:253)^[[0m >>>>>>>> ^[[31m at >>>>>>>> org.apache.spark.sql.jdbc.DockerJDBCIntegrationSuite.run(DockerJDBCIntegrationSuite.scala:58)^[[0m >>>>>>>> ^[[31m ...^[[0m >>>>>>>> ^[[31m Cause: >>>>>>>> com.spotify.docker.client.shaded.javax.ws.rs.ProcessingException: >>>>>>>> java.io.IOException: No such file or directory^[[0m >>>>>>>> ^[[31m at >>>>>>>> org.glassfish.jersey.apache.connector.ApacheConnector.apply(ApacheConnector.java:481)^[[0m >>>>>>>> ^[[31m at >>>>>>>> org.glassfish.jersey.apache.connector.ApacheConnector$1.run(ApacheConnector.java:491)^[[0m >>>>>>>> ^[[31m at >>>>>>>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)^[[0m >>>>>>>> ^[[31m at >>>>>>>> java.util.concurrent.FutureTask.run(FutureTask.java:262)^[[0m >>>>>>>> ^[[31m at >>>>>>>> jersey.repackaged.com.google.common.util.concurrent.MoreExecutors$DirectExecutorService.execute(MoreExecutors.java:299)^[[0m >>>>>>>> ^[[31m at >>>>>>>> java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:110)^[[0m >>>>>>>> ^[[31m at >>>>>>>> jersey.repackaged.com.google.common.util.concurrent.AbstractListeningExecutorService.submit(AbstractListeningExecutorService.java:50)^[[0m >>>>>>>> ^[[31m at >>>>>>>> jersey.repackaged.com.google.common.util.concurrent.AbstractListeningExecutorService.submit(AbstractListeningExecutorService.java:37)^[[0m >>>>>>>> ^[[31m at >>>>>>>> org.glassfish.jersey.apache.connector.ApacheConnector.apply(ApacheConnector.java:487)^[[0m >>>>>>>> ^[[31m at >>>>>>>> org.glassfish.jersey.client.ClientRuntime$2.run(ClientRuntime.java:177)^[[0m >>>>>>>> ^[[31m ...^[[0m >>>>>>>> ^[[31m Cause: java.io.IOException: No such file or directory^[[0m >>>>>>>> ^[[31m at >>>>>>>> jnr.unixsocket.UnixSocketChannel.doConnect(UnixSocketChannel.java:94)^[[0m >>>>>>>> >>>>>>>> Has anyone seen the above ? >>>>>>>> >>>>>>>> On Wed, Mar 2, 2016 at 2:45 PM, Michael Armbrust < >>>>>>>> <mich...@databricks.com>mich...@databricks.com> wrote: >>>>>>>> >>>>>>>>> Please vote on releasing the following candidate as Apache Spark >>>>>>>>> version 1.6.1! >>>>>>>>> >>>>>>>>> The vote is open until Saturday, March 5, 2016 at 20:00 UTC and >>>>>>>>> passes if a majority of at least 3+1 PMC votes are cast. >>>>>>>>> >>>>>>>>> [ ] +1 Release this package as Apache Spark 1.6.1 >>>>>>>>> [ ] -1 Do not release this package because ... >>>>>>>>> >>>>>>>>> To learn more about Apache Spark, please see >>>>>>>>> <http://spark.apache.org/>http://spark.apache.org/ >>>>>>>>> >>>>>>>>> The tag to be voted on is *v1.6.1-rc1 >>>>>>>>> (15de51c238a7340fa81cb0b80d029a05d97bfc5c) >>>>>>>>> <https://github.com/apache/spark/tree/v1.6.1-rc1>* >>>>>>>>> >>>>>>>>> The release files, including signatures, digests, etc. can be >>>>>>>>> found at: >>>>>>>>> >>>>>>>>> <https://home.apache.org/%7Epwendell/spark-releases/spark-1.6.1-rc1-bin/> >>>>>>>>> https://home.apache.org/~pwendell/spark-releases/spark-1.6.1-rc1-bin/ >>>>>>>>> >>>>>>>>> Release artifacts are signed with the following key: >>>>>>>>> <https://people.apache.org/keys/committer/pwendell.asc> >>>>>>>>> https://people.apache.org/keys/committer/pwendell.asc >>>>>>>>> >>>>>>>>> The staging repository for this release can be found at: >>>>>>>>> >>>>>>>>> <https://repository.apache.org/content/repositories/orgapachespark-1180/> >>>>>>>>> https://repository.apache.org/content/repositories/orgapachespark-1180/ >>>>>>>>> >>>>>>>>> The test repository (versioned as v1.6.1-rc1) for this release can >>>>>>>>> be found at: >>>>>>>>> >>>>>>>>> <https://repository.apache.org/content/repositories/orgapachespark-1179/> >>>>>>>>> https://repository.apache.org/content/repositories/orgapachespark-1179/ >>>>>>>>> >>>>>>>>> The documentation corresponding to this release can be found at: >>>>>>>>> >>>>>>>>> <https://home.apache.org/%7Epwendell/spark-releases/spark-1.6.1-rc1-docs/> >>>>>>>>> https://home.apache.org/~pwendell/spark-releases/spark-1.6.1-rc1-docs/ >>>>>>>>> >>>>>>>>> >>>>>>>>> ======================================= >>>>>>>>> == How can I help test this release? == >>>>>>>>> ======================================= >>>>>>>>> If you are a Spark user, you can help us test this release by >>>>>>>>> taking an existing Spark workload and running on this release >>>>>>>>> candidate, >>>>>>>>> then reporting any regressions from 1.6.0. >>>>>>>>> >>>>>>>>> ================================================ >>>>>>>>> == What justifies a -1 vote for this release? == >>>>>>>>> ================================================ >>>>>>>>> This is a maintenance release in the 1.6.x series. Bugs already >>>>>>>>> present in 1.6.0, missing features, or bugs related to new features >>>>>>>>> will >>>>>>>>> not necessarily block this release. >>>>>>>>> >>>>>>>>> =============================================================== >>>>>>>>> == What should happen to JIRA tickets still targeting 1.6.0? == >>>>>>>>> =============================================================== >>>>>>>>> 1. It is OK for documentation patches to target 1.6.1 and still go >>>>>>>>> into branch-1.6, since documentations will be published separately >>>>>>>>> from the >>>>>>>>> release. >>>>>>>>> 2. New features for non-alpha-modules should target 1.7+. >>>>>>>>> 3. Non-blocker bug fixes should target 1.6.2 or 2.0.0, or drop the >>>>>>>>> target version. >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> >>>>>> >>>>>> *Sincerely yours Egor Pakhomov * >>>>>> >>>>> >>>>> >>>> >>> >> >> >