This vote fails.  Following shortly with RC3

On Thu, Jun 1, 2017 at 8:28 PM, Reynold Xin <r...@databricks.com> wrote:

> Again (I've probably said this more than 10 times already in different
> threads), SPARK-18350 has no impact on whether the timestamp type is with
> timezone or without timezone. It simply allows a session specific timezone
> setting rather than having Spark always rely on the machine timezone.
>
> On Wed, May 31, 2017 at 11:58 AM, Kostas Sakellis <kos...@cloudera.com>
> wrote:
>
>> Hey Michael,
>>
>> There is a discussion on TIMESTAMP semantics going on the thread "SQL
>> TIMESTAMP semantics vs. SPARK-18350" which might impact Spark 2.2. Should
>> we make a decision there before voting on the next RC for Spark 2.2?
>>
>> Thanks,
>> Kostas
>>
>> On Tue, May 30, 2017 at 12:09 PM, Michael Armbrust <
>> mich...@databricks.com> wrote:
>>
>>> Last call, anything else important in-flight for 2.2?
>>>
>>> On Thu, May 25, 2017 at 10:56 AM, Michael Allman <mich...@videoamp.com>
>>> wrote:
>>>
>>>> PR is here: https://github.com/apache/spark/pull/18112
>>>>
>>>>
>>>> On May 25, 2017, at 10:28 AM, Michael Allman <mich...@videoamp.com>
>>>> wrote:
>>>>
>>>> Michael,
>>>>
>>>> If you haven't started cutting the new RC, I'm working on a
>>>> documentation PR right now I'm hoping we can get into Spark 2.2 as a
>>>> migration note, even if it's just a mention: https://issues.apache
>>>> .org/jira/browse/SPARK-20888.
>>>>
>>>> Michael
>>>>
>>>>
>>>> On May 22, 2017, at 11:39 AM, Michael Armbrust <mich...@databricks.com>
>>>> wrote:
>>>>
>>>> I'm waiting for SPARK-20814
>>>> <https://issues.apache.org/jira/browse/SPARK-20814> at Marcelo's
>>>> request and I'd also like to include SPARK-20844
>>>> <https://issues.apache.org/jira/browse/SPARK-20844>.  I think we
>>>> should be able to cut another RC midweek.
>>>>
>>>> On Fri, May 19, 2017 at 11:53 AM, Nick Pentreath <
>>>> nick.pentre...@gmail.com> wrote:
>>>>
>>>>> All the outstanding ML QA doc and user guide items are done for 2.2 so
>>>>> from that side we should be good to cut another RC :)
>>>>>
>>>>>
>>>>> On Thu, 18 May 2017 at 00:18 Russell Spitzer <
>>>>> russell.spit...@gmail.com> wrote:
>>>>>
>>>>>> Seeing an issue with the DataScanExec and some of our integration
>>>>>> tests for the SCC. Running dataframe read and writes from the shell seems
>>>>>> fine but the Redaction code seems to get a "None" when doing
>>>>>> SparkSession.getActiveSession.get in our integration tests. I'm not
>>>>>> sure why but i'll dig into this later if I get a chance.
>>>>>>
>>>>>> Example Failed Test
>>>>>> https://github.com/datastax/spark-cassandra-connector/blob/v
>>>>>> 2.0.1/spark-cassandra-connector/src/it/scala/com/datastax/sp
>>>>>> ark/connector/sql/CassandraSQLSpec.scala#L311
>>>>>>
>>>>>> ```[info]   org.apache.spark.SparkException: Job aborted due to
>>>>>> stage failure: Task serialization failed: 
>>>>>> java.util.NoSuchElementException:
>>>>>> None.get
>>>>>> [info] java.util.NoSuchElementException: None.get
>>>>>> [info] at scala.None$.get(Option.scala:347)
>>>>>> [info] at scala.None$.get(Option.scala:345)
>>>>>> [info] at org.apache.spark.sql.execution.DataSourceScanExec$class.org
>>>>>> $apache$spark$sql$execution$DataSourceScanExec$$redact(DataSo
>>>>>> urceScanExec.scala:70)
>>>>>> [info] at org.apache.spark.sql.execution
>>>>>> .DataSourceScanExec$$anonfun$4.apply(DataSourceScanExec.scala:54)
>>>>>> [info] at org.apache.spark.sql.execution
>>>>>> .DataSourceScanExec$$anonfun$4.apply(DataSourceScanExec.scala:52)
>>>>>> ```
>>>>>>
>>>>>> Again this only seems to repo in our IT suite so i'm not sure if this
>>>>>> is a real issue.
>>>>>>
>>>>>>
>>>>>> On Tue, May 16, 2017 at 1:40 PM Joseph Bradley <jos...@databricks.com>
>>>>>> wrote:
>>>>>>
>>>>>>> All of the ML/Graph/SparkR QA blocker JIRAs have been resolved.
>>>>>>> Thanks everyone who helped out on those!
>>>>>>>
>>>>>>> We still have open ML/Graph/SparkR JIRAs targeted at 2.2, but they
>>>>>>> are essentially all for documentation.
>>>>>>>
>>>>>>> Joseph
>>>>>>>
>>>>>>> On Thu, May 11, 2017 at 3:08 PM, Marcelo Vanzin <van...@cloudera.com
>>>>>>> > wrote:
>>>>>>>
>>>>>>>> Since you'll be creating a new RC, I'd wait until SPARK-20666 is
>>>>>>>> fixed, since the change that caused it is in branch-2.2. Probably a
>>>>>>>> good idea to raise it to blocker at this point.
>>>>>>>>
>>>>>>>> On Thu, May 11, 2017 at 2:59 PM, Michael Armbrust
>>>>>>>> <mich...@databricks.com> wrote:
>>>>>>>> > I'm going to -1 given the outstanding issues and lack of +1s.
>>>>>>>> I'll create
>>>>>>>> > another RC once ML has had time to take care of the more critical
>>>>>>>> problems.
>>>>>>>> > In the meantime please keep testing this release!
>>>>>>>> >
>>>>>>>> > On Tue, May 9, 2017 at 2:00 AM, Kazuaki Ishizaki <
>>>>>>>> ishiz...@jp.ibm.com>
>>>>>>>> > wrote:
>>>>>>>> >>
>>>>>>>> >> +1 (non-binding)
>>>>>>>> >>
>>>>>>>> >> I tested it on Ubuntu 16.04 and OpenJDK8 on ppc64le. All of the
>>>>>>>> tests for
>>>>>>>> >> core have passed.
>>>>>>>> >>
>>>>>>>> >> $ java -version
>>>>>>>> >> openjdk version "1.8.0_111"
>>>>>>>> >> OpenJDK Runtime Environment (build
>>>>>>>> >> 1.8.0_111-8u111-b14-2ubuntu0.16.04.2-b14)
>>>>>>>> >> OpenJDK 64-Bit Server VM (build 25.111-b14, mixed mode)
>>>>>>>> >> $ build/mvn -DskipTests -Phive -Phive-thriftserver -Pyarn
>>>>>>>> -Phadoop-2.7
>>>>>>>> >> package install
>>>>>>>> >> $ build/mvn -Phive -Phive-thriftserver -Pyarn -Phadoop-2.7 test
>>>>>>>> -pl core
>>>>>>>> >> ...
>>>>>>>> >> Run completed in 15 minutes, 12 seconds.
>>>>>>>> >> Total number of tests run: 1940
>>>>>>>> >> Suites: completed 206, aborted 0
>>>>>>>> >> Tests: succeeded 1940, failed 0, canceled 4, ignored 8, pending 0
>>>>>>>> >> All tests passed.
>>>>>>>> >> [INFO]
>>>>>>>> >> ------------------------------------------------------------
>>>>>>>> ------------
>>>>>>>> >> [INFO] BUILD SUCCESS
>>>>>>>> >> [INFO]
>>>>>>>> >> ------------------------------------------------------------
>>>>>>>> ------------
>>>>>>>> >> [INFO] Total time: 16:51 min
>>>>>>>> >> [INFO] Finished at: 2017-05-09T17:51:04+09:00
>>>>>>>> >> [INFO] Final Memory: 53M/514M
>>>>>>>> >> [INFO]
>>>>>>>> >> ------------------------------------------------------------
>>>>>>>> ------------
>>>>>>>> >> [WARNING] The requested profile "hive" could not be activated
>>>>>>>> because it
>>>>>>>> >> does not exist.
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >> Kazuaki Ishizaki,
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >> From:        Michael Armbrust <mich...@databricks.com>
>>>>>>>> >> To:        "dev@spark.apache.org" <dev@spark.apache.org>
>>>>>>>> >> Date:        2017/05/05 02:08
>>>>>>>> >> Subject:        [VOTE] Apache Spark 2.2.0 (RC2)
>>>>>>>> >> ________________________________
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >> Please vote on releasing the following candidate as Apache Spark
>>>>>>>> version
>>>>>>>> >> 2.2.0. The vote is open until Tues, May 9th, 2017 at 12:00 PST
>>>>>>>> and passes if
>>>>>>>> >> a majority of at least 3 +1 PMC votes are cast.
>>>>>>>> >>
>>>>>>>> >> [ ] +1 Release this package as Apache Spark 2.2.0
>>>>>>>> >> [ ] -1 Do not release this package because ...
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >> To learn more about Apache Spark, please see
>>>>>>>> http://spark.apache.org/
>>>>>>>> >>
>>>>>>>> >> The tag to be voted on is v2.2.0-rc2
>>>>>>>> >> (1d4017b44d5e6ad156abeaae6371747f111dd1f9)
>>>>>>>> >>
>>>>>>>> >> List of JIRA tickets resolved can be found with this filter.
>>>>>>>> >>
>>>>>>>> >> The release files, including signatures, digests, etc. can be
>>>>>>>> found at:
>>>>>>>> >> http://home.apache.org/~pwendell/spark-releases/spark-2.2.0-
>>>>>>>> rc2-bin/
>>>>>>>> >>
>>>>>>>> >> Release artifacts are signed with the following key:
>>>>>>>> >> https://people.apache.org/keys/committer/pwendell.asc
>>>>>>>> >>
>>>>>>>> >> The staging repository for this release can be found at:
>>>>>>>> >> https://repository.apache.org/content/repositories/orgapache
>>>>>>>> spark-1236/
>>>>>>>> >>
>>>>>>>> >> The documentation corresponding to this release can be found at:
>>>>>>>> >> http://people.apache.org/~pwendell/spark-releases/spark-2.2.
>>>>>>>> 0-rc2-docs/
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >> FAQ
>>>>>>>> >>
>>>>>>>> >> How can I help test this release?
>>>>>>>> >>
>>>>>>>> >> If you are a Spark user, you can help us test this release by
>>>>>>>> taking an
>>>>>>>> >> existing Spark workload and running on this release candidate,
>>>>>>>> then
>>>>>>>> >> reporting any regressions.
>>>>>>>> >>
>>>>>>>> >> What should happen to JIRA tickets still targeting 2.2.0?
>>>>>>>> >>
>>>>>>>> >> Committers should look at those and triage. Extremely important
>>>>>>>> bug fixes,
>>>>>>>> >> documentation, and API tweaks that impact compatibility should
>>>>>>>> be worked on
>>>>>>>> >> immediately. Everything else please retarget to 2.3.0 or 2.2.1.
>>>>>>>> >>
>>>>>>>> >> But my bug isn't fixed!??!
>>>>>>>> >>
>>>>>>>> >> In order to make timely releases, we will typically not hold the
>>>>>>>> release
>>>>>>>> >> unless the bug in question is a regression from 2.1.1.
>>>>>>>> >>
>>>>>>>> >
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Marcelo
>>>>>>>>
>>>>>>>> ------------------------------------------------------------
>>>>>>>> ---------
>>>>>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Joseph Bradley
>>>>>>> Software Engineer - Machine Learning
>>>>>>> Databricks, Inc.
>>>>>>> [image: http://databricks.com] <http://databricks.com/>
>>>>>>>
>>>>>>
>>>>
>>>>
>>>>
>>>
>>
>

Reply via email to