Thank you, Yi.
+1

On Mon, Jun 21, 2021 at 3:00 PM Jungtaek Lim <kabhwan.opensou...@gmail.com>
wrote:

> +1 (non-binding) Thanks for your efforts!
>
> On Mon, Jun 21, 2021 at 2:40 PM Kent Yao <yaooq...@gmail.com> wrote:
>
>> +1 (non-binding)
>>
>> *Kent Yao *
>> @ Data Science Center, Hangzhou Research Institute, NetEase Corp.
>> *a spark enthusiast*
>> *kyuubi <https://github.com/yaooqinn/kyuubi>is a
>> unified multi-tenant JDBC interface for large-scale data processing and
>> analytics, built on top of Apache Spark <http://spark.apache.org/>.*
>> *spark-authorizer <https://github.com/yaooqinn/spark-authorizer>A Spark
>> SQL extension which provides SQL Standard Authorization for **Apache
>> Spark <http://spark.apache.org/>.*
>> *spark-postgres <https://github.com/yaooqinn/spark-postgres> A library
>> for reading data from and transferring data to Postgres / Greenplum with
>> Spark SQL and DataFrames, 10~100x faster.*
>> *itatchi <https://github.com/yaooqinn/spark-func-extras>A** library t**hat
>> brings useful functions from various modern database management systems to 
>> **Apache
>> Spark <http://spark.apache.org/>.*
>>
>>
>>
>> On 06/21/2021 13:40,Hyukjin Kwon<gurwls...@gmail.com>
>> <gurwls...@gmail.com> wrote:
>>
>> +1
>>
>> 2021년 6월 21일 (월) 오후 2:19, Dongjoon Hyun <dongjoon.h...@gmail.com>님이 작성:
>>
>>> +1
>>>
>>> Thank you, Yi.
>>>
>>> Bests,
>>> Dongjoon.
>>>
>>>
>>> On Sat, Jun 19, 2021 at 6:57 PM Yuming Wang <wgy...@gmail.com> wrote:
>>>
>>>> +1
>>>>
>>>> Tested a batch of production query with Thrift Server.
>>>>
>>>> On Sat, Jun 19, 2021 at 3:04 PM Mridul Muralidharan <mri...@gmail.com>
>>>> wrote:
>>>>
>>>>>
>>>>> +1
>>>>>
>>>>> Signatures, digests, etc check out fine.
>>>>> Checked out tag and build/tested with -Pyarn -Phadoop-2.7 -Pmesos
>>>>> -Pkubernetes
>>>>>
>>>>> Regards,
>>>>> Mridul
>>>>>
>>>>> PS: Might be related to some quirk of my local env - the first test
>>>>> run (after clean + package) usually fails for me (typically for hive 
>>>>> tests)
>>>>> - with a second run succeeding : this is not specific to this RC though.
>>>>>
>>>>> On Fri, Jun 18, 2021 at 6:14 PM Liang-Chi Hsieh <vii...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> +1. Docs looks good. Binary looks good.
>>>>>>
>>>>>> Ran simple test and some tpcds queries.
>>>>>>
>>>>>> Thanks for working on this!
>>>>>>
>>>>>>
>>>>>> wuyi wrote
>>>>>> > Please vote on releasing the following candidate as Apache Spark
>>>>>> version
>>>>>> > 3.0.3.
>>>>>> >
>>>>>> > The vote is open until Jun 21th 3AM (PST) and passes if a majority
>>>>>> +1 PMC
>>>>>> > votes are cast, with
>>>>>> > a minimum of 3 +1 votes.
>>>>>> >
>>>>>> > [ ] +1 Release this package as Apache Spark 3.0.3
>>>>>> > [ ] -1 Do not release this package because ...
>>>>>> >
>>>>>> > To learn more about Apache Spark, please see
>>>>>> https://spark.apache.org/
>>>>>> >
>>>>>> > The tag to be voted on is v3.0.3-rc1 (commit
>>>>>> > 65ac1e75dc468f53fc778cd2ce1ba3f21067aab8):
>>>>>> > https://github.com/apache/spark/tree/v3.0.3-rc1
>>>>>> >
>>>>>> > The release files, including signatures, digests, etc. can be found
>>>>>> at:
>>>>>> > https://dist.apache.org/repos/dist/dev/spark/v3.0.3-rc1-bin/
>>>>>> >
>>>>>> > Signatures used for Spark RCs can be found in this file:
>>>>>> > https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>>>> >
>>>>>> > The staging repository for this release can be found at:
>>>>>> >
>>>>>> https://repository.apache.org/content/repositories/orgapachespark-1386/
>>>>>> >
>>>>>> > The documentation corresponding to this release can be found at:
>>>>>> > https://dist.apache.org/repos/dist/dev/spark/v3.0.3-rc1-docs/
>>>>>> >
>>>>>> > The list of bug fixes going into 3.0.3 can be found at the
>>>>>> following URL:
>>>>>> > https://issues.apache.org/jira/projects/SPARK/versions/12349723
>>>>>> >
>>>>>> > This release is using the release script of the tag v3.0.3-rc1.
>>>>>> >
>>>>>> > FAQ
>>>>>> >
>>>>>> > =========================
>>>>>> > How can I help test this release?
>>>>>> > =========================
>>>>>> >
>>>>>> > If you are a Spark user, you can help us test this release by taking
>>>>>> > an existing Spark workload and running on this release candidate,
>>>>>> then
>>>>>> > reporting any regressions.
>>>>>> >
>>>>>> > If you're working in PySpark you can set up a virtual env and
>>>>>> install
>>>>>> > the current RC and see if anything important breaks, in the
>>>>>> Java/Scala
>>>>>> > you can add the staging repository to your projects resolvers and
>>>>>> test
>>>>>> > with the RC (make sure to clean up the artifact cache before/after
>>>>>> so
>>>>>> > you don't end up building with a out of date RC going forward).
>>>>>> >
>>>>>> > ===========================================
>>>>>> > What should happen to JIRA tickets still targeting 3.0.3?
>>>>>> > ===========================================
>>>>>> >
>>>>>> > The current list of open tickets targeted at 3.0.3 can be found at:
>>>>>> > https://issues.apache.org/jira/projects/SPARK and search for
>>>>>> "Target
>>>>>> > Version/s" = 3.0.3
>>>>>> >
>>>>>> > Committers should look at those and triage. Extremely important bug
>>>>>> > fixes, documentation, and API tweaks that impact compatibility
>>>>>> should
>>>>>> > be worked on immediately. Everything else please retarget to an
>>>>>> > appropriate release.
>>>>>> >
>>>>>> > ==================
>>>>>> > But my bug isn't fixed?
>>>>>> > ==================
>>>>>> >
>>>>>> > In order to make timely releases, we will typically not hold the
>>>>>> > release unless the bug in question is a regression from the previous
>>>>>> > release. That being said, if there is something which is a
>>>>>> regression
>>>>>> > that has not been correctly targeted please ping me or a committer
>>>>>> to
>>>>>> > help target the issue.
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> Sent from: http://apache-spark-developers-list.1001551.n3.nabble.com/
>>>>>>
>>>>>> ---------------------------------------------------------------------
>>>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>>>
>>>>>> ---------------------------------------------------------------------
>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>
>

-- 
---
Takeshi Yamamuro

Reply via email to