Those still appear to be env problems. I don't know why it is so
persistent. Does it all pass locally? Retrigger tests again and see what
happens.

On Tue, Jun 19, 2018, 2:53 AM Petar Zecevic <petar.zece...@gmail.com> wrote:

>
> Thanks, but unfortunately, it died again. Now at pyspark tests:
>
>
> ========================================================================
> Running PySpark tests
> ========================================================================
> Running PySpark tests. Output is in 
> /home/jenkins/workspace/SparkPullRequestBuilder@2/python/unit-tests.log
> Will test against the following Python executables: ['python2.7', 
> 'python3.4', 'pypy']
> Will test the following Python modules: ['pyspark-core', 'pyspark-ml', 
> 'pyspark-mllib', 'pyspark-sql', 'pyspark-streaming']
> Will skip PyArrow related features against Python executable 'python2.7' in 
> 'pyspark-sql' module. PyArrow >= 0.8.0 is required; however, PyArrow was not 
> found.
> Will skip Pandas related features against Python executable 'python2.7' in 
> 'pyspark-sql' module. Pandas >= 0.19.2 is required; however, Pandas 0.16.0 
> was found.
> Will test PyArrow related features against Python executable 'python3.4' in 
> 'pyspark-sql' module.
> Will test Pandas related features against Python executable 'python3.4' in 
> 'pyspark-sql' module.
> Will skip PyArrow related features against Python executable 'pypy' in 
> 'pyspark-sql' module. PyArrow >= 0.8.0 is required; however, PyArrow was not 
> found.
> Will skip Pandas related features against Python executable 'pypy' in 
> 'pyspark-sql' module. Pandas >= 0.19.2 is required; however, Pandas was not 
> found.
> Starting test(python2.7): pyspark.mllib.tests
> Starting test(pypy): pyspark.sql.tests
> Starting test(pypy): pyspark.streaming.tests
> Starting test(pypy): pyspark.tests
> Setting default log level to "WARN".
> To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use 
> setLogLevel(newLevel).
> ...
> [Stage 0:>                                                          (0 + 1) / 
> 1]
>
> ..
> [Stage 0:>                                                          (0 + 4) / 
> 4]
>
> .
> [Stage 0:>                                                          (0 + 4) / 
> 4]
>
> ..
> [Stage 0:>                                                          (0 + 4) / 
> 4]
>
>
> [Stage 0:>                                                          (0 + 4) / 
> 4]
>
>
> [Stage 0:>                                                          (0 + 4) / 
> 4]
>
>
> [Stage 0:>                                                          (0 + 4) / 
> 4]
>
> ....
> [Stage 0:>                                                        (0 + 32) / 
> 32]...........
> [Stage 10:>                                                         (0 + 1) / 
> 1]
>
> .........
> [Stage 0:>                                                          (0 + 4) / 
> 4]
>
> .....s
> [Stage 0:>                                                          (0 + 1) / 
> 1]
>
> .........
> [Stage 0:>                                                          (0 + 4) / 
> 4]
> [Stage 0:==============>                                            (1 + 3) / 
> 4]
>
> .
> [Stage 0:>                                                          (0 + 4) / 
> 4]
>
> ..
> [Stage 0:>                                                          (0 + 2) / 
> 2]
>
> .........
> [Stage 29:===========================================>              (3 + 1) / 
> 4]
>
> ......
> [Stage 79:>                                                         (0 + 1) / 
> 1]
>
> ..
> [Stage 83:>                                                        (0 + 4) / 
> 10]
> [Stage 83:======================>                                  (4 + 4) / 
> 10]
> [Stage 83:=============================================>           (8 + 2) / 
> 10]
>
> ......cc: cc: no input filesno input files
>
> cc: no input files
> cc: no input files
> Exception in thread Thread-1:
> Traceback (most recent call last):
>   File "/usr/lib64/pypy-2.5.1/lib-python/2.7/threading.py", line 806, in 
> __bootstrap_inner
>     self.run()
>   File "/usr/lib64/pypy-2.5.1/lib-python/2.7/threading.py", line 759, in run
>     self.__target(*self.__args, **self.__kwargs)
>   File 
> "/home/jenkins/workspace/SparkPullRequestBuilder@2/python/pyspark/rdd.py" 
> </home/jenkins/workspace/SparkPullRequestBuilder@2/python/pyspark/rdd.py>, 
> line 771, in pipe_objs
>     out.write(s.encode('utf-8'))
> IOError: [Errno 32] Broken pipe: '<fdopen>'
>
> cc: no input files
> cc: no input files
> cc: no input files
> Exception in thread Thread-1:
> Traceback (most recent call last):
>   File "/usr/lib64/pypy-2.5.1/lib-python/2.7/threading.py", line 806, in 
> __bootstrap_inner
>     self.run()
>   File "/usr/lib64/pypy-2.5.1/lib-python/2.7/threading.py", line 759, in run
>     self.__target(*self.__args, **self.__kwargs)
>   File 
> "/home/jenkins/workspace/SparkPullRequestBuilder@2/python/pyspark/rdd.py" 
> </home/jenkins/workspace/SparkPullRequestBuilder@2/python/pyspark/rdd.py>, 
> line 771, in pipe_objs
>     out.write(s.encode('utf-8'))
> IOError: [Errno 32] Broken pipe: '<fdopen>'
>
> Exception in thread Thread-1:
> Traceback (most recent call last):
>   File "/usr/lib64/pypy-2.5.1/lib-python/2.7/threading.py", line 806, in 
> __bootstrap_inner
>     self.run()
>   File "/usr/lib64/pypy-2.5.1/lib-python/2.7/threading.py", line 759, in run
>     self.__target(*self.__args, **self.__kwargs)
>   File 
> "/home/jenkins/workspace/SparkPullRequestBuilder@2/python/pyspark/rdd.py" 
> </home/jenkins/workspace/SparkPullRequestBuilder@2/python/pyspark/rdd.py>, 
> line 771, in pipe_objs
>     out.write(s.encode('utf-8'))
> IOError: [Errno 32] Broken pipe: '<fdopen>'
>
> cc: no input files
> Exception in thread Thread-1:
> Traceback (most recent call last):
>   File "/usr/lib64/pypy-2.5.1/lib-python/2.7/threading.py", line 806, in 
> __bootstrap_inner
>     self.run()
>   File "/usr/lib64/pypy-2.5.1/lib-python/2.7/threading.py", line 759, in run
>     self.__target(*self.__args, **self.__kwargs)
>   File 
> "/home/jenkins/workspace/SparkPullRequestBuilder@2/python/pyspark/rdd.py" 
> </home/jenkins/workspace/SparkPullRequestBuilder@2/python/pyspark/rdd.py>, 
> line 771, in pipe_objs
>     out.write(s.encode('utf-8'))
> IOError: [Errno 32] Broken pipe: '<fdopen>'
>
> .......................s..Traceback (most recent call last):
>   File 
> "/home/jenkins/workspace/SparkPullRequestBuilder@2/python/pyspark/serializers.py"
>  
> </home/jenkins/workspace/SparkPullRequestBuilder@2/python/pyspark/serializers.py>,
>  line 574, in dumps
>     return cloudpickle.dumps(obj, 2)
>   File 
> "/home/jenkins/workspace/SparkPullRequestBuilder@2/python/pyspark/cloudpickle.py"
>  
> </home/jenkins/workspace/SparkPullRequestBuilder@2/python/pyspark/cloudpickle.py>,
>  line 858, in dumps
>     cp.dump(obj)
>   File 
> "/home/jenkins/workspace/SparkPullRequestBuilder@2/python/pyspark/cloudpickle.py"
>  
> </home/jenkins/workspace/SparkPullRequestBuilder@2/python/pyspark/cloudpickle.py>,
>  line 260, in dump
>     return Pickler.dump(self, obj)
>   File "/usr/lib64/pypy-2.5.1/lib-python/2.7/pickle.py", line 224, in dump
>     self.save(obj)
>   File "/usr/lib64/pypy-2.5.1/lib-python/2.7/pickle.py", line 306, in save
>     rv = reduce(self.proto)
>   File 
> "/home/jenkins/workspace/SparkPullRequestBuilder@2/python/pyspark/tests.py" 
> </home/jenkins/workspace/SparkPullRequestBuilder@2/python/pyspark/tests.py>, 
> line 275, in __reduce__
>     raise Exception("not picklable")
> Exception: not picklable
> .......
> [Stage 0:>                                                          (0 + 4) / 
> 4]
>
> ....file:/tmp/tmpgO7AIY added as a remote repository with the name: repo-1
> Ivy Default Cache set to: /home/jenkins/.ivy2/cache
> The jars for the packages stored in: /home/jenkins/.ivy2/jars
> :: loading settings :: url = 
> jar:file:/home/jenkins/workspace/SparkPullRequestBuilder@2/assembly/target/scala-2.11/jars/ivy-2.4.0.jar!/org/apache/ivy/core/settings/ivysettings.xml
> a#mylib added as a dependency
> :: resolving dependencies :: org.apache.spark#spark-submit-parent;1.0
>       confs: [default]
> :: resolution report :: resolve 1990ms :: artifacts dl 0ms
>       :: modules in use:
>       ---------------------------------------------------------------------
>       |                  |            modules            ||   artifacts   |
>       |       conf       | number| search|dwnlded|evicted|| number|dwnlded|
>       ---------------------------------------------------------------------
>       |      default     |   1   |   0   |   0   |   0   ||   0   |   0   |
>       ---------------------------------------------------------------------
>
> :: problems summary ::
> :::: WARNINGS
>       io problem while parsing ivy file: 
> file:/tmp/tmpgO7AIY/a/mylib/0.1/mylib-0.1.pom (java.io.FileNotFoundException: 
> /home/jenkins/.ivy2/cache/a/mylib/ivy-0.1.xml.original (No such file or 
> directory))
>
>               module not found: a#mylib;0.1
>
>       ==== local-m2-cache: tried
>
>         
> file:/home/jenkins/workspace/SparkPullRequestBuilder@2/dummy/.m2/repository/a/mylib/0.1/mylib-0.1.pom
>
>         -- artifact a#mylib;0.1!mylib.jar:
>
>         
> file:/home/jenkins/workspace/SparkPullRequestBuilder@2/dummy/.m2/repository/a/mylib/0.1/mylib-0.1.jar
>
>       ==== local-ivy-cache: tried
>
>         /home/jenkins/.ivy2/local/a/mylib/0.1/ivys/ivy.xml
>
>         -- artifact a#mylib;0.1!mylib.jar:
>
>         /home/jenkins/.ivy2/local/a/mylib/0.1/jars/mylib.jar
>
>       ==== central: tried
>
>         https://repo1.maven.org/maven2/a/mylib/0.1/mylib-0.1.pom
>
>         -- artifact a#mylib;0.1!mylib.jar:
>
>         https://repo1.maven.org/maven2/a/mylib/0.1/mylib-0.1.jar
>
>       ==== spark-packages: tried
>
>         http://dl.bintray.com/spark-packages/maven/a/mylib/0.1/mylib-0.1.pom
>
>         -- artifact a#mylib;0.1!mylib.jar:
>
>         http://dl.bintray.com/spark-packages/maven/a/mylib/0.1/mylib-0.1.jar
>
>       ==== repo-1: tried
>
>         file:/tmp/tmpgO7AIY/a/mylib/0.1/mylib-0.1.pom
>
>               ::::::::::::::::::::::::::::::::::::::::::::::
>
>               ::          UNRESOLVED DEPENDENCIES         ::
>
>               ::::::::::::::::::::::::::::::::::::::::::::::
>
>               :: a#mylib;0.1: not found
>
>               ::::::::::::::::::::::::::::::::::::::::::::::
>
>
>
> :: USE VERBOSE OR DEBUG MESSAGE LEVEL FOR MORE DETAILS
> Exception in thread "main" java.lang.RuntimeException: [unresolved 
> dependency: a#mylib;0.1: not found]
>       at 
> org.apache.spark.deploy.SparkSubmitUtils$.resolveMavenCoordinates(SparkSubmit.scala:1268)
>       at 
> org.apache.spark.deploy.DependencyUtils$.resolveMavenDependencies(DependencyUtils.scala:49)
>       at 
> org.apache.spark.deploy.SparkSubmit$.prepareSubmitEnvironment(SparkSubmit.scala:348)
>       at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:170)
>       at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:136)
>       at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> Ffile:/tmp/tmpwtN2z_ added as a remote repository with the name: repo-1
> Ivy Default Cache set to: /home/jenkins/.ivy2/cache
> The jars for the packages stored in: /home/jenkins/.ivy2/jars
> :: loading settings :: url = 
> jar:file:/home/jenkins/workspace/SparkPullRequestBuilder@2/assembly/target/scala-2.11/jars/ivy-2.4.0.jar!/org/apache/ivy/core/settings/ivysettings.xml
> a#mylib added as a dependency
> :: resolving dependencies :: org.apache.spark#spark-submit-parent;1.0
>       confs: [default]
>       found a#mylib;0.1 in repo-1
> :: resolution report :: resolve 1378ms :: artifacts dl 4ms
>       :: modules in use:
>       a#mylib;0.1 from repo-1 in [default]
>       ---------------------------------------------------------------------
>       |                  |            modules            ||   artifacts   |
>       |       conf       | number| search|dwnlded|evicted|| number|dwnlded|
>       ---------------------------------------------------------------------
>       |      default     |   1   |   1   |   1   |   0   ||   1   |   0   |
>       ---------------------------------------------------------------------
> :: retrieving :: org.apache.spark#spark-submit-parent
>       confs: [default]
>       0 artifacts copied, 1 already retrieved (0kB/8ms)
> .....
> [Stage 0:>                                                          (0 + 4) / 
> 4]
>
> .
> [Stage 0:>                                                          (0 + 1) / 
> 1]
>
> .
> [Stage 0:>                                                          (0 + 1) / 
> 1]
>
> ...
> [Stage 0:>                                                         (0 + 4) / 
> 20]
> [Stage 0:=================>                                        (6 + 4) / 
> 20]
>
> ......
> ======================================================================
> FAIL: test_package_dependency (pyspark.tests.SparkSubmitTests)
> Submit and test a script with a dependency on a Spark Package
> ----------------------------------------------------------------------
> Traceback (most recent call last):
>   File 
> "/home/jenkins/workspace/SparkPullRequestBuilder@2/python/pyspark/tests.py" 
> </home/jenkins/workspace/SparkPullRequestBuilder@2/python/pyspark/tests.py>, 
> line 2093, in test_package_dependency
>     self.assertEqual(0, proc.returncode)
> AssertionError: 0 != 1
>
> ----------------------------------------------------------------------
> Ran 127 tests in 205.547s
>
> FAILED (failures=1, skipped=2)
> NOTE: Skipping SciPy tests as it does not seem to be installed
> NOTE: Skipping NumPy tests as it does not seem to be installed
>    Random listing order was used
>
> Had test failures in pyspark.tests with pypy; see logs.
> [error] running 
> /home/jenkins/workspace/SparkPullRequestBuilder@2/python/run-tests 
> --parallelism=4 ; received return code 255
> Attempting to post to Github...
>  > Post successful.
> Build step 'Execute shell' marked build as failure
> Archiving artifacts
> Recording test results
> Test FAILed.
> Refer to this link for build results (access rights to CI server needed): 
> https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/92038/
> Test FAILed.
> Finished: FAILURE
>
>
> Le 6/18/2018 à 8:05 PM, shane knapp a écrit :
>
> i triggered another build against your PR, so let's see if this happens
> again or was a transient failure.
>
> https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/92038/
>
> shane
>
> On Mon, Jun 18, 2018 at 5:30 AM, Petar Zecevic <petar.zece...@gmail.com>
> wrote:
>
>> Hi,
>> Jenkins build for my PR (https://github.com/apache/spark/pull/21109 ;
>> https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/92023/testReport/org.apache.spark.sql.hive/HiveExternalCatalogVersionsSuite/_It_is_not_a_test_it_is_a_sbt_testing_SuiteSelector_/)
>> keeps failing. First it couldn't download Spark v.2.2.0 (indeed, it wasn't
>> available at the mirror it selected), now it's failing with this exception
>> below.
>>
>> Can someone explain these errors for me? Is anybody else experiencing
>> similar problems?
>>
>> Thanks,
>> Petar
>>
>>
>> Error Message
>>
>> java.io.IOException: Cannot run program "./bin/spark-submit" (in
>> directory "/tmp/test-spark/spark-2.2.1"): error=2, No such file or
>> directory
>>
>> Stacktrace
>>
>> sbt.ForkMain$ForkError: java.io.IOException: Cannot run program
>> "./bin/spark-submit" (in directory "/tmp/test-spark/spark-2.2.1"): error=2,
>> No such file or directory
>>     at java.lang.ProcessBuilder.start(ProcessBuilder.java:1048)
>>     at
>> org.apache.spark.sql.hive.SparkSubmitTestUtils$class.runSparkSubmit(SparkSubmitTestUtils.scala:73)
>>     at
>> org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.runSparkSubmit(HiveExternalCatalogVersionsSuite.scala:43)
>>     at
>> org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite$$anonfun$beforeAll$1.apply(HiveExternalCatalogVersionsSuite.scala:176)
>>     at
>> org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite$$anonfun$beforeAll$1.apply(HiveExternalCatalogVersionsSuite.scala:161)
>>     at scala.collection.immutable.List.foreach(List.scala:381)
>>     at
>> org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.beforeAll(HiveExternalCatalogVersionsSuite.scala:161)
>>     at
>> org.scalatest.BeforeAndAfterAll$class.liftedTree1$1(BeforeAndAfterAll.scala:212)
>>     at
>> org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:210)
>>     at org.apache.spark.SparkFunSuite.run(SparkFunSuite.scala:52)
>>     at org.scalatest.tools.Framework.org
>> $scalatest$tools$Framework$$runSuite(Framework.scala:314)
>>     at
>> org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:480)
>>     at sbt.ForkMain$Run$2.call(ForkMain.java:296)
>>     at sbt.ForkMain$Run$2.call(ForkMain.java:286)
>>     at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>>     at
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>>     at
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>>     at java.lang.Thread.run(Thread.java:745)
>> Caused by: sbt.ForkMain$ForkError: java.io.IOException: error=2, No such
>> file or directory
>>     at java.lang.UNIXProcess.forkAndExec(Native Method)
>>     at java.lang.UNIXProcess.<init>(UNIXProcess.java:248)
>>     at java.lang.ProcessImpl.start(ProcessImpl.java:134)
>>     at java.lang.ProcessBuilder.start(ProcessBuilder.java:1029)
>>     ... 17 more
>>
>>
>
>
> --
> Shane Knapp
> UC Berkeley EECS Research / RISELab Staff Technical Lead
> https://rise.cs.berkeley.edu
>
>
>

Reply via email to