Re: Time for 2.1.3

2018-06-19 Thread Marcelo Vanzin
Quick update for everybody: I was trying to deal with the release
scripts to get them to work with 2.1; there were some fixes needed,
and on top of that Maven Central changed something over the weekend
which made Java 7 unhappy.

I actually was able to create an RC1 after many tries and tweaking,
but there's also currently at least one blocker that I think we should
pick up. So right now I'm waiting for that.

So heads up that RC2 will be the first one put up for a vote (I'll
explain that in the vote e-mail so people who missed this one don't
scratch their heads for too long).


On Tue, Jun 12, 2018 at 4:27 PM, Marcelo Vanzin  wrote:
> Hey all,
>
> There are some fixes that went into 2.1.3 recently that probably
> deserve a release. So as usual, please take a look if there's anything
> else you'd like on that release, otherwise I'd like to start with the
> process by early next week.
>
> I'll go through jira to see what's the status of things targeted at
> that release, but last I checked there wasn't anything on the radar.
>
> Thanks!
>
> --
> Marcelo



-- 
Marcelo

-
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org



Re: Jenkins build errors

2018-06-19 Thread Sean Owen
Those still appear to be env problems. I don't know why it is so
persistent. Does it all pass locally? Retrigger tests again and see what
happens.

On Tue, Jun 19, 2018, 2:53 AM Petar Zecevic  wrote:

>
> Thanks, but unfortunately, it died again. Now at pyspark tests:
>
>
> 
> Running PySpark tests
> 
> Running PySpark tests. Output is in 
> /home/jenkins/workspace/SparkPullRequestBuilder@2/python/unit-tests.log
> Will test against the following Python executables: ['python2.7', 
> 'python3.4', 'pypy']
> Will test the following Python modules: ['pyspark-core', 'pyspark-ml', 
> 'pyspark-mllib', 'pyspark-sql', 'pyspark-streaming']
> Will skip PyArrow related features against Python executable 'python2.7' in 
> 'pyspark-sql' module. PyArrow >= 0.8.0 is required; however, PyArrow was not 
> found.
> Will skip Pandas related features against Python executable 'python2.7' in 
> 'pyspark-sql' module. Pandas >= 0.19.2 is required; however, Pandas 0.16.0 
> was found.
> Will test PyArrow related features against Python executable 'python3.4' in 
> 'pyspark-sql' module.
> Will test Pandas related features against Python executable 'python3.4' in 
> 'pyspark-sql' module.
> Will skip PyArrow related features against Python executable 'pypy' in 
> 'pyspark-sql' module. PyArrow >= 0.8.0 is required; however, PyArrow was not 
> found.
> Will skip Pandas related features against Python executable 'pypy' in 
> 'pyspark-sql' module. Pandas >= 0.19.2 is required; however, Pandas was not 
> found.
> Starting test(python2.7): pyspark.mllib.tests
> Starting test(pypy): pyspark.sql.tests
> Starting test(pypy): pyspark.streaming.tests
> Starting test(pypy): pyspark.tests
> Setting default log level to "WARN".
> To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use 
> setLogLevel(newLevel).
> ...
> [Stage 0:>  (0 + 1) / 
> 1]
>
> ..
> [Stage 0:>  (0 + 4) / 
> 4]
>
> .
> [Stage 0:>  (0 + 4) / 
> 4]
>
> ..
> [Stage 0:>  (0 + 4) / 
> 4]
>
>
> [Stage 0:>  (0 + 4) / 
> 4]
>
>
> [Stage 0:>  (0 + 4) / 
> 4]
>
>
> [Stage 0:>  (0 + 4) / 
> 4]
>
> 
> [Stage 0:>(0 + 32) / 
> 32]...
> [Stage 10:> (0 + 1) / 
> 1]
>
> .
> [Stage 0:>  (0 + 4) / 
> 4]
>
> .s
> [Stage 0:>  (0 + 1) / 
> 1]
>
> .
> [Stage 0:>  (0 + 4) / 
> 4]
> [Stage 0:==>(1 + 3) / 
> 4]
>
> .
> [Stage 0:>  (0 + 4) / 
> 4]
>
> ..
> [Stage 0:>  (0 + 2) / 
> 2]
>
> .
> [Stage 29:===>  (3 + 1) / 
> 4]
>
> ..
> [Stage 79:> (0 + 1) / 
> 1]
>
> ..
> [Stage 83:>(0 + 4) / 
> 10]
> [Stage 83:==>  (4 + 4) / 
> 10]
> [Stage 83:=>   (8 + 2) / 
> 10]
>
> ..cc: cc: no input filesno input files
>
> cc: no input files
> cc: no input files
> Exception in thread Thread-1:
> Traceback (most recent call last):
>   File "/usr/lib64/pypy-2.5.1/lib-python/2.7/threading.py", line 806, in 
> __bootstrap_inner
> self.run()
>   File "/usr/lib64/pypy-2.5.1/lib-python/2.7/threading.py", line 759, in run
> self.__target(*self.__args, **self.__kwargs)
>   File 
> "/home/jenkins/workspace/SparkPullRequestBuilder@2/python/pyspark/rdd.py" 
> , 
> line 771, in pipe_objs
> out.write(s.encode('utf-8'))
> IOError: [Errno 32] Broken pipe: ''
>
> cc: no input files
> cc: no input files
> cc: no input files
> Exception in thread Thread-1:
> Traceback (most recent call last):
>   File "/usr/lib64/pypy-2.5.1/lib-python/2.7/threading.py", line 806, in 
> __bootstrap_inner
> self.run()
>   File "/usr/lib64/pypy-2.5.1/lib-python/2.7/threading.py", line 759, in run
> self.__target(*self.__args, **self.__kwargs)
>   File 
> "/home/jenkins/workspace/SparkPullRequestBuilder@2/python/pyspark/rdd.py" 
> , 
> line 771, in pipe_objs
> out.write(s.encode('utf-8'))
> IOError: [Errno 32] Broken pipe: ''
>
> Exception in 

Asking for reviewing PRs regarding structured streaming

2018-06-19 Thread Jungtaek Lim
Hi Spark devs,

I have couple of pull requests for structured streaming which are getting
older and fading out from earlier pages in PR pages.

https://github.com/apache/spark/pull/21469
https://github.com/apache/spark/pull/21357
https://github.com/apache/spark/pull/21222

Two of them are in a kind of approval by couple of folks, but no approval
from committers yet.
One of them needs rebase and I would be happy to do it after reviewing or
in progress of reviewing.

Getting reviewed in time would be critical for contributors to be honest,
so I'd like to ask dev mailing list to review my PRs.

Thanks in advance,
Jungtaek Lim (HeartSaVioR)


Re: Jenkins build errors

2018-06-19 Thread Petar Zecevic


Thanks, but unfortunately, it died again. Now at pyspark tests:



Running PySpark tests

Running PySpark tests. Output is in 
/home/jenkins/workspace/SparkPullRequestBuilder@2/python/unit-tests.log
Will test against the following Python executables: ['python2.7', 'python3.4', 
'pypy']
Will test the following Python modules: ['pyspark-core', 'pyspark-ml', 
'pyspark-mllib', 'pyspark-sql', 'pyspark-streaming']
Will skip PyArrow related features against Python executable 'python2.7' in 
'pyspark-sql' module. PyArrow >= 0.8.0 is required; however, PyArrow was not 
found.
Will skip Pandas related features against Python executable 'python2.7' in 
'pyspark-sql' module. Pandas >= 0.19.2 is required; however, Pandas 0.16.0 was 
found.
Will test PyArrow related features against Python executable 'python3.4' in 
'pyspark-sql' module.
Will test Pandas related features against Python executable 'python3.4' in 
'pyspark-sql' module.
Will skip PyArrow related features against Python executable 'pypy' in 
'pyspark-sql' module. PyArrow >= 0.8.0 is required; however, PyArrow was not 
found.
Will skip Pandas related features against Python executable 'pypy' in 
'pyspark-sql' module. Pandas >= 0.19.2 is required; however, Pandas was not 
found.
Starting test(python2.7): pyspark.mllib.tests
Starting test(pypy): pyspark.sql.tests
Starting test(pypy): pyspark.streaming.tests
Starting test(pypy): pyspark.tests
Setting default log level to "WARN".
To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use 
setLogLevel(newLevel).
...
[Stage 0:>  (0 + 1) / 1]

..

[Stage 0:>  (0 + 4) / 4]

.

[Stage 0:>  (0 + 4) / 4]

..

[Stage 0:>  (0 + 4) / 4]



[Stage 0:>  (0 + 4) / 4]



[Stage 0:>  (0 + 4) / 4]



[Stage 0:>  (0 + 4) / 4]



[Stage 0:>(0 + 32) / 
32]...
[Stage 10:> (0 + 1) / 1]

.

[Stage 0:>  (0 + 4) / 4]

.s

[Stage 0:>  (0 + 1) / 1]

.

[Stage 0:>  (0 + 4) / 4]
[Stage 0:==>(1 + 3) / 4]

.

[Stage 0:>  (0 + 4) / 4]

..

[Stage 0:>  (0 + 2) / 2]

.

[Stage 29:===>  (3 + 1) / 4]

..

[Stage 79:> (0 + 1) / 1]

..

[Stage 83:>(0 + 4) / 10]
[Stage 83:==>  (4 + 4) / 10]
[Stage 83:=>   (8 + 2) / 10]

..cc: cc: no input filesno input files


cc: no input files
cc: no input files
Exception in thread Thread-1:
Traceback (most recent call last):
  File "/usr/lib64/pypy-2.5.1/lib-python/2.7/threading.py", line 806, in 
__bootstrap_inner
self.run()
  File