BTW, for that integration suite, I saw the related artifacts in the RC4
staging directory.

Does Spark 2.4.0 need to start to release these
`spark-kubernetes-integration-tests`
artifacts?

   -
   
https://repository.apache.org/content/repositories/orgapachespark-1290/org/apache/spark/spark-kubernetes-integration-tests_2.11/
   -
   
https://repository.apache.org/content/repositories/orgapachespark-1290/org/apache/spark/spark-kubernetes-integration-tests_2.12/

Historically, Spark released `spark-docker-integration-tests` at Spark
1.6.x era and stopped since Spark 2.0.0.

   -
   
http://central.maven.org/maven2/org/apache/spark/spark-docker-integration-tests_2.10/
   -
   
http://central.maven.org/maven2/org/apache/spark/spark-docker-integration-tests_2.11/


Bests,
Dongjoon.

On Tue, Oct 23, 2018 at 11:43 AM Stavros Kontopoulos <
stavros.kontopou...@lightbend.com> wrote:

> Sean,
>
> Ok makes sense, im using a cloned repo. I built with Scala 2.12 profile
> using the related tag v2.4.0-rc4:
>
> ./dev/change-scala-version.sh 2.12
> ./dev/make-distribution.sh  --name test --r --tgz -Pscala-2.12 -Psparkr
> -Phadoop-2.7 -Pkubernetes -Phive
> Pushed images to dockerhub (previous email) since I didnt use the minikube
> daemon (default behavior).
>
> Then run tests successfully against minikube:
>
> TGZ_PATH=$(pwd)/spark-2.4.0-bin-test.gz
> cd resource-managers/kubernetes/integration-tests
>
> ./dev/dev-run-integration-tests.sh --spark-tgz $TGZ_PATH --service-account
> default --namespace default --image-tag k8s-scala-12 --image-repo skonto
>
>
> [INFO]
> [INFO] --- scalatest-maven-plugin:1.0:test (integration-test) @
> spark-kubernetes-integration-tests_2.12 ---
> Discovery starting.
> Discovery completed in 229 milliseconds.
> Run starting. Expected test count is: 14
> KubernetesSuite:
> - Run SparkPi with no resources
> - Run SparkPi with a very long application name.
> - Use SparkLauncher.NO_RESOURCE
> - Run SparkPi with a master URL without a scheme.
> - Run SparkPi with an argument.
> - Run SparkPi with custom labels, annotations, and environment variables.
> - Run extraJVMOptions check on driver
> - Run SparkRemoteFileTest using a remote data file
> - Run SparkPi with env and mount secrets.
> - Run PySpark on simple pi.py example
> - Run PySpark with Python2 to test a pyfiles example
> - Run PySpark with Python3 to test a pyfiles example
> - Run PySpark with memory customization
> - Run in client mode.
> Run completed in 5 minutes, 24 seconds.
> Total number of tests run: 14
> Suites: completed 2, aborted 0
> Tests: succeeded 14, failed 0, canceled 0, ignored 0, pending 0
> All tests passed.
> [INFO]
> ------------------------------------------------------------------------
> [INFO] Reactor Summary:
> [INFO]
> [INFO] Spark Project Parent POM 2.4.0 ..................... SUCCESS [
> 4.491 s]
> [INFO] Spark Project Tags ................................. SUCCESS [
> 3.833 s]
> [INFO] Spark Project Local DB ............................. SUCCESS [
> 2.680 s]
> [INFO] Spark Project Networking ........................... SUCCESS [
> 4.817 s]
> [INFO] Spark Project Shuffle Streaming Service ............ SUCCESS [
> 2.541 s]
> [INFO] Spark Project Unsafe ............................... SUCCESS [
> 2.795 s]
> [INFO] Spark Project Launcher ............................. SUCCESS [
> 5.593 s]
> [INFO] Spark Project Core ................................. SUCCESS [
> 25.160 s]
> [INFO] Spark Project Kubernetes Integration Tests 2.4.0 ... SUCCESS [05:30
> min]
> [INFO]
> ------------------------------------------------------------------------
> [INFO] BUILD SUCCESS
> [INFO]
> ------------------------------------------------------------------------
> [INFO] Total time: 06:23 min
> [INFO] Finished at: 2018-10-23T18:39:11Z
> [INFO]
> ------------------------------------------------------------------------
>
>
> but had to modify this line
> <https://github.com/apache/spark/blob/master/resource-managers/kubernetes/integration-tests/dev/dev-run-integration-tests.sh#L106>
>  and
> added -Pscala-2.12 , otherwise it fails (these tests inherit from the
> parent pom but the profile is not propagated to the mvn command that
> launches the tests, I can create a PR to fix that).
>
>
> On Tue, Oct 23, 2018 at 7:44 PM, Hyukjin Kwon <gurwls...@gmail.com> wrote:
>
>> https://github.com/apache/spark/pull/22514 sounds like a regression that
>> affects Hive CTAS in write path (by not replacing them into Spark internal
>> datasources; therefore performance regression).
>> but yea I suspect if we should block the release by this.
>>
>> https://github.com/apache/spark/pull/22144 is just being discussed if I
>> am not mistaken.
>>
>> Thanks.
>>
>> 2018년 10월 24일 (수) 오전 12:27, Xiao Li <gatorsm...@gmail.com>님이 작성:
>>
>>> https://github.com/apache/spark/pull/22144 is also not a blocker of
>>> Spark 2.4 release, as discussed in the PR.
>>>
>>> Thanks,
>>>
>>> Xiao
>>>
>>> Xiao Li <gatorsm...@gmail.com> 于2018年10月23日周二 上午9:20写道:
>>>
>>>> Thanks for reporting this. https://github.com/apache/spark/pull/22514
>>>> is not a blocker. We can fix it in the next minor release, if we are unable
>>>> to make it in this release.
>>>>
>>>> Thanks,
>>>>
>>>> Xiao
>>>>
>>>> Sean Owen <sro...@gmail.com> 于2018年10月23日周二 上午9:14写道:
>>>>
>>>>> (I should add, I only observed this with the Scala 2.12 build. It all
>>>>> seemed to work with 2.11. Therefore I'm not too worried about it. I
>>>>> don't think it's a Scala version issue, but perhaps something looking
>>>>> for a spark 2.11 tarball and not finding it. See
>>>>> https://github.com/apache/spark/pull/22805#issuecomment-432304622 for
>>>>> a change that might address this kind of thing.)
>>>>>
>>>>> On Tue, Oct 23, 2018 at 11:05 AM Sean Owen <sro...@gmail.com> wrote:
>>>>> >
>>>>> > Yeah, that's maybe the issue here. This is a source release, not a
>>>>> git checkout, and it still needs to work in this context.
>>>>> >
>>>>> > I just added -Pkubernetes to my build and didn't do anything else. I
>>>>> think the ideal is that a "mvn -P... -P... install" to work from a source
>>>>> release; that's a good expectation and consistent with docs.
>>>>> >
>>>>> > Maybe these tests simply don't need to run with the normal suite of
>>>>> tests, and can be considered tests run manually by developers running 
>>>>> these
>>>>> scripts? Basically, KubernetesSuite shouldn't run in a normal mvn install?
>>>>> >
>>>>> > I don't think this has to block the release even if so, just trying
>>>>> to get to the bottom of it.
>>>>>
>>>>> ---------------------------------------------------------------------
>>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>>
>>>>>
>
>
> --
> Stavros Kontopoulos
>
> *Senior Software Engineer*
> *Lightbend, Inc.*
>
> *p:  +30 6977967274 <%2B1%20650%20678%200020>*
> *e: stavros.kontopou...@lightbend.com* <dave.mar...@lightbend.com>
>
>
>

Reply via email to