Hi All,

Thanks for testing the 3.4.0 RC! I apologize for the maven testing failures
for the Spark Connect Scala Client. We will try to get those sorted as soon
as possible.

This is an artifact of having multiple build systems, and only running CI
for one (SBT). That, however, is a debate for another day :)...

Cheers,
Herman

On Wed, Feb 22, 2023 at 5:32 PM Bjørn Jørgensen <bjornjorgen...@gmail.com>
wrote:

> ./build/mvn clean package
>
> I'm using ubuntu rolling, python 3.11 openjdk 17
>
> CompatibilitySuite:
> - compatibility MiMa tests *** FAILED ***
>   java.lang.AssertionError: assertion failed: Failed to find the jar
> inside folder: /home/bjorn/spark-3.4.0/connector/connect/client/jvm/target
>   at scala.Predef$.assert(Predef.scala:223)
>   at
> org.apache.spark.sql.connect.client.util.IntegrationTestUtils$.findJar(IntegrationTestUtils.scala:67)
>   at
> org.apache.spark.sql.connect.client.CompatibilitySuite.clientJar$lzycompute(CompatibilitySuite.scala:57)
>   at
> org.apache.spark.sql.connect.client.CompatibilitySuite.clientJar(CompatibilitySuite.scala:53)
>   at
> org.apache.spark.sql.connect.client.CompatibilitySuite.$anonfun$new$1(CompatibilitySuite.scala:69)
>   at org.scalatest.OutcomeOf.outcomeOf(OutcomeOf.scala:85)
>   at org.scalatest.OutcomeOf.outcomeOf$(OutcomeOf.scala:83)
>   at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
>   at org.scalatest.Transformer.apply(Transformer.scala:22)
>   at org.scalatest.Transformer.apply(Transformer.scala:20)
>   ...
> - compatibility API tests: Dataset *** FAILED ***
>   java.lang.AssertionError: assertion failed: Failed to find the jar
> inside folder: /home/bjorn/spark-3.4.0/connector/connect/client/jvm/target
>   at scala.Predef$.assert(Predef.scala:223)
>   at
> org.apache.spark.sql.connect.client.util.IntegrationTestUtils$.findJar(IntegrationTestUtils.scala:67)
>   at
> org.apache.spark.sql.connect.client.CompatibilitySuite.clientJar$lzycompute(CompatibilitySuite.scala:57)
>   at
> org.apache.spark.sql.connect.client.CompatibilitySuite.clientJar(CompatibilitySuite.scala:53)
>   at
> org.apache.spark.sql.connect.client.CompatibilitySuite.$anonfun$new$7(CompatibilitySuite.scala:110)
>   at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
>   at org.scalatest.OutcomeOf.outcomeOf(OutcomeOf.scala:85)
>   at org.scalatest.OutcomeOf.outcomeOf$(OutcomeOf.scala:83)
>   at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
>   at org.scalatest.Transformer.apply(Transformer.scala:22)
>   ...
> SparkConnectClientSuite:
> - Placeholder test: Create SparkConnectClient
> - Test connection
> - Test connection string
> - Check URI: sc://host, isCorrect: true
> - Check URI: sc://localhost/, isCorrect: true
> - Check URI: sc://localhost:1234/, isCorrect: true
> - Check URI: sc://localhost/;, isCorrect: true
> - Check URI: sc://host:123, isCorrect: true
> - Check URI: sc://host:123/;user_id=a94, isCorrect: true
> - Check URI: scc://host:12, isCorrect: false
> - Check URI: http://host, isCorrect: false
> - Check URI: sc:/host:1234/path, isCorrect: false
> - Check URI: sc://host/path, isCorrect: false
> - Check URI: sc://host/;parm1;param2, isCorrect: false
> - Check URI: sc://host:123;user_id=a94, isCorrect: false
> - Check URI: sc:///user_id=123, isCorrect: false
> - Check URI: sc://host:-4, isCorrect: false
> - Check URI: sc://:123/, isCorrect: false
> - Non user-id parameters throw unsupported errors
> DatasetSuite:
> - limit
> - select
> - filter
> - write
> UserDefinedFunctionSuite:
> - udf and encoder serialization
> Run completed in 21 seconds, 944 milliseconds.
> Total number of tests run: 389
> Suites: completed 10, aborted 0
> Tests: succeeded 386, failed 3, canceled 0, ignored 0, pending 0
> *** 3 TESTS FAILED ***
> [INFO]
> ------------------------------------------------------------------------
> [INFO] Reactor Summary for Spark Project Parent POM 3.4.0:
> [INFO]
> [INFO] Spark Project Parent POM ........................... SUCCESS [
> 47.096 s]
> [INFO] Spark Project Tags ................................. SUCCESS [
> 14.759 s]
> [INFO] Spark Project Sketch ............................... SUCCESS [
> 21.628 s]
> [INFO] Spark Project Local DB ............................. SUCCESS [
> 20.311 s]
> [INFO] Spark Project Networking ........................... SUCCESS [01:07
> min]
> [INFO] Spark Project Shuffle Streaming Service ............ SUCCESS [
> 15.921 s]
> [INFO] Spark Project Unsafe ............................... SUCCESS [
> 16.020 s]
> [INFO] Spark Project Launcher ............................. SUCCESS [
> 10.873 s]
> [INFO] Spark Project Core ................................. SUCCESS [37:10
> min]
> [INFO] Spark Project ML Local Library ..................... SUCCESS [
> 40.841 s]
> [INFO] Spark Project GraphX ............................... SUCCESS [02:39
> min]
> [INFO] Spark Project Streaming ............................ SUCCESS [05:53
> min]
> [INFO] Spark Project Catalyst ............................. SUCCESS [11:22
> min]
> [INFO] Spark Project SQL .................................. SUCCESS [
>  02:27 h]
> [INFO] Spark Project ML Library ........................... SUCCESS [22:45
> min]
> [INFO] Spark Project Tools ................................ SUCCESS [
>  7.263 s]
> [INFO] Spark Project Hive ................................. SUCCESS [
>  01:21 h]
> [INFO] Spark Project REPL ................................. SUCCESS [02:07
> min]
> [INFO] Spark Project Assembly ............................. SUCCESS [
> 11.704 s]
> [INFO] Kafka 0.10+ Token Provider for Streaming ........... SUCCESS [
> 26.748 s]
> [INFO] Spark Integration for Kafka 0.10 ................... SUCCESS [01:44
> min]
> [INFO] Kafka 0.10+ Source for Structured Streaming ........ SUCCESS [33:27
> min]
> [INFO] Spark Project Examples ............................. SUCCESS [01:17
> min]
> [INFO] Spark Integration for Kafka 0.10 Assembly .......... SUCCESS [
> 27.292 s]
> [INFO] Spark Avro ......................................... SUCCESS [02:18
> min]
> [INFO] Spark Project Connect Common ....................... SUCCESS [
> 43.728 s]
> [INFO] Spark Project Connect Server ....................... SUCCESS [01:19
> min]
> [INFO] Spark Project Connect Client ....................... FAILURE [
> 53.524 s]
> [INFO] Spark Protobuf ..................................... SKIPPED
> [INFO]
> ------------------------------------------------------------------------
> [INFO] BUILD FAILURE
> [INFO]
> ------------------------------------------------------------------------
> [INFO] Total time:  05:58 h
> [INFO] Finished at: 2023-02-22T22:28:38+01:00
> [INFO]
> ------------------------------------------------------------------------
> [ERROR] Failed to execute goal
> org.scalatest:scalatest-maven-plugin:2.2.0:test (test) on project
> spark-connect-client-jvm_2.12: There are test failures -> [Help 1]
>
> ons. 22. feb. 2023 kl. 21:41 skrev Mridul Muralidharan <mri...@gmail.com>:
>
>>
>> Signatures, digests, etc check out fine - thanks for updating them !
>> Checked out tag and build/tested with -Phive -Pyarn -Pmesos -Pkubernetes
>>
>>
>> The test ClientE2ETestSuite.simple udf failed [1] in "Connect Client "
>> module ... yet to test "Spark Protobuf" module due to the failure.
>>
>>
>> Regards,
>> Mridul
>>
>> [1]
>>
>> - simple udf *** FAILED ***
>>
>>   io.grpc.StatusRuntimeException: INTERNAL:
>> org.apache.spark.sql.ClientE2ETestSuite
>>
>>   at io.grpc.Status.asRuntimeException(Status.java:535)
>>
>>   at
>> io.grpc.stub.ClientCalls$BlockingResponseStream.hasNext(ClientCalls.java:660)
>>
>>   at org.apache.spark.sql.connect.client.SparkResult.org
>> $apache$spark$sql$connect$client$SparkResult$$processResponses(SparkResult.scala:50)
>>
>>   at
>> org.apache.spark.sql.connect.client.SparkResult.length(SparkResult.scala:95)
>>
>>   at
>> org.apache.spark.sql.connect.client.SparkResult.toArray(SparkResult.scala:112)
>>
>>   at org.apache.spark.sql.Dataset.$anonfun$collect$1(Dataset.scala:2037)
>>
>>   at org.apache.spark.sql.Dataset.withResult(Dataset.scala:2267)
>>
>>   at org.apache.spark.sql.Dataset.collect(Dataset.scala:2036)
>>
>>   at
>> org.apache.spark.sql.ClientE2ETestSuite.$anonfun$new$5(ClientE2ETestSuite.scala:65)
>>
>>   at
>> scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
>>
>>   ...
>>
>>
>>
>>
>>
>> On Wed, Feb 22, 2023 at 2:07 AM Mridul Muralidharan <mri...@gmail.com>
>> wrote:
>>
>>>
>>> Thanks Xinrong !
>>> The signature verifications are fine now ... will continue with testing
>>> the release.
>>>
>>>
>>> Regards,
>>> Mridul
>>>
>>>
>>> On Wed, Feb 22, 2023 at 1:27 AM Xinrong Meng <xinrong.apa...@gmail.com>
>>> wrote:
>>>
>>>> Hi Mridul,
>>>>
>>>> Would you please try that again? It should work now.
>>>>
>>>> On Wed, Feb 22, 2023 at 2:04 PM Mridul Muralidharan <mri...@gmail.com>
>>>> wrote:
>>>>
>>>>>
>>>>> Hi Xinrong,
>>>>>
>>>>>   Was it signed with the same key as present in KEYS [1] ?
>>>>> I am seeing errors with gpg when validating. For example:
>>>>>
>>>>>
>>>>> $ gpg --verify pyspark-3.4.0.tar.gz.asc
>>>>>
>>>>> gpg: assuming signed data in 'pyspark-3.4.0.tar.gz'
>>>>>
>>>>> gpg: Signature made Tue 21 Feb 2023 05:56:05 AM CST
>>>>>
>>>>> gpg:                using RSA key
>>>>> CC68B3D16FE33A766705160BA7E57908C7A4E1B1
>>>>>
>>>>> gpg:                issuer "xinr...@apache.org"
>>>>>
>>>>> gpg: Can't check signature: No public key
>>>>>
>>>>>
>>>>>
>>>>> Regards,
>>>>> Mridul
>>>>>
>>>>> [1] https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>>>
>>>>>
>>>>> On Tue, Feb 21, 2023 at 10:36 PM Xinrong Meng <
>>>>> xinrong.apa...@gmail.com> wrote:
>>>>>
>>>>>> Please vote on releasing the following candidate as Apache Spark
>>>>>> version 3.4.0.
>>>>>>
>>>>>> The vote is open until 11:59pm Pacific time *February 27th* and
>>>>>> passes if a majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
>>>>>>
>>>>>> [ ] +1 Release this package as Apache Spark 3.4.0
>>>>>> [ ] -1 Do not release this package because ...
>>>>>>
>>>>>> To learn more about Apache Spark, please see http://spark.apache.org/
>>>>>>
>>>>>> The tag to be voted on is *v3.4.0-rc1* (commit
>>>>>> e2484f626bb338274665a49078b528365ea18c3b):
>>>>>> https://github.com/apache/spark/tree/v3.4.0-rc1
>>>>>>
>>>>>> The release files, including signatures, digests, etc. can be found
>>>>>> at:
>>>>>> https://dist.apache.org/repos/dist/dev/spark/v3.4.0-rc1-bin/
>>>>>>
>>>>>> Signatures used for Spark RCs can be found in this file:
>>>>>> https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>>>>
>>>>>> The staging repository for this release can be found at:
>>>>>> https://repository.apache.org/content/repositories/orgapachespark-1435
>>>>>>
>>>>>> The documentation corresponding to this release can be found at:
>>>>>> https://dist.apache.org/repos/dist/dev/spark/v3.4.0-rc1-docs/
>>>>>>
>>>>>> The list of bug fixes going into 3.4.0 can be found at the following
>>>>>> URL:
>>>>>> https://issues.apache.org/jira/projects/SPARK/versions/12351465
>>>>>>
>>>>>> This release is using the release script of the tag v3.4.0-rc1.
>>>>>>
>>>>>>
>>>>>> FAQ
>>>>>>
>>>>>> =========================
>>>>>> How can I help test this release?
>>>>>> =========================
>>>>>> If you are a Spark user, you can help us test this release by taking
>>>>>> an existing Spark workload and running on this release candidate, then
>>>>>> reporting any regressions.
>>>>>>
>>>>>> If you're working in PySpark you can set up a virtual env and install
>>>>>> the current RC and see if anything important breaks, in the Java/Scala
>>>>>> you can add the staging repository to your projects resolvers and test
>>>>>> with the RC (make sure to clean up the artifact cache before/after so
>>>>>> you don't end up building with a out of date RC going forward).
>>>>>>
>>>>>> ===========================================
>>>>>> What should happen to JIRA tickets still targeting 3.4.0?
>>>>>> ===========================================
>>>>>> The current list of open tickets targeted at 3.4.0 can be found at:
>>>>>> https://issues.apache.org/jira/projects/SPARK and search for "Target
>>>>>> Version/s" = 3.4.0
>>>>>>
>>>>>> Committers should look at those and triage. Extremely important bug
>>>>>> fixes, documentation, and API tweaks that impact compatibility should
>>>>>> be worked on immediately. Everything else please retarget to an
>>>>>> appropriate release.
>>>>>>
>>>>>> ==================
>>>>>> But my bug isn't fixed?
>>>>>> ==================
>>>>>> In order to make timely releases, we will typically not hold the
>>>>>> release unless the bug in question is a regression from the previous
>>>>>> release. That being said, if there is something which is a regression
>>>>>> that has not been correctly targeted please ping me or a committer to
>>>>>> help target the issue.
>>>>>>
>>>>>> Thanks,
>>>>>> Xinrong Meng
>>>>>>
>>>>>
>
> --
> Bjørn Jørgensen
> Vestre Aspehaug 4, 6010 Ålesund
> Norge
>
> +47 480 94 297
>

Reply via email to