This vote passes! I'll follow up with a formal release announcement soon.
+1:
Sean Owen (binding)
Wenchen Fan (binding)
Hyukjin Kwon (binding)
Dongjoon Hyun (binding)
Takeshi Yamamuro
+0: None
-1: None
Thanks, everyone!
Xingbo
On Mon, Nov 4, 2019 at 9:35 AM Dongjoon Hyun
wrote:
> Hi,
Hi, Xingbo.
Could you sent a vote result email to finalize this vote, please?
Bests,
Dongjoon.
On Fri, Nov 1, 2019 at 2:55 PM Takeshi Yamamuro
wrote:
> +1, too.
>
> On Sat, Nov 2, 2019 at 3:36 AM Hyukjin Kwon wrote:
>
>> +1
>>
>> On Fri, 1 Nov 2019, 15:36 Wenchen Fan, wrote:
>>
>>> The PR
+1, too.
On Sat, Nov 2, 2019 at 3:36 AM Hyukjin Kwon wrote:
> +1
>
> On Fri, 1 Nov 2019, 15:36 Wenchen Fan, wrote:
>
>> The PR builder uses Hadoop 2.7 profile, which makes me think that 2.7 is
>> more stable and we should make releases using 2.7 by default.
>>
>> +1
>>
>> On Fri, Nov 1, 2019
+1
On Fri, 1 Nov 2019, 15:36 Wenchen Fan, wrote:
> The PR builder uses Hadoop 2.7 profile, which makes me think that 2.7 is
> more stable and we should make releases using 2.7 by default.
>
> +1
>
> On Fri, Nov 1, 2019 at 7:16 AM Xiao Li wrote:
>
>> Spark 3.0 will still use the Hadoop 2.7
+1 for Apache Spark 3.0.0-preview (RC2).
Bests,
Dongjoon.
On Thu, Oct 31, 2019 at 11:36 PM Wenchen Fan wrote:
> The PR builder uses Hadoop 2.7 profile, which makes me think that 2.7 is
> more stable and we should make releases using 2.7 by default.
>
> +1
>
> On Fri, Nov 1, 2019 at 7:16 AM
The PR builder uses Hadoop 2.7 profile, which makes me think that 2.7 is
more stable and we should make releases using 2.7 by default.
+1
On Fri, Nov 1, 2019 at 7:16 AM Xiao Li wrote:
> Spark 3.0 will still use the Hadoop 2.7 profile by default, I think.
> Hadoop 2.7 profile is much more
Spark 3.0 will still use the Hadoop 2.7 profile by default, I think. Hadoop
2.7 profile is much more stable than Hadoop 3.2 profile.
On Thu, Oct 31, 2019 at 3:54 PM Sean Owen wrote:
> This isn't a big thing, but I see that the pyspark build includes
> Hadoop 2.7 rather than 3.2. Maybe later we
This isn't a big thing, but I see that the pyspark build includes
Hadoop 2.7 rather than 3.2. Maybe later we change the build to put in
3.2 by default.
Otherwise, the tests all seems to pass with JDK 8 / 11 with all
profiles enabled, so I'm +1 on it.
On Thu, Oct 31, 2019 at 1:00 AM Xingbo Jiang
Please vote on releasing the following candidate as Apache Spark version
3.0.0-preview.
The vote is open until November 3 PST and passes if a majority +1 PMC votes
are cast, with
a minimum of 3 +1 votes.
[ ] +1 Release this package as Apache Spark 3.0.0-preview
[ ] -1 Do not release this package