You should try cleaning and then building.  We have recently hit a bug in
the scala compiler that sometimes causes non-clean builds to fail.


On Wed, Jul 16, 2014 at 7:56 PM, Matei Zaharia <matei.zaha...@gmail.com>
wrote:

> Yeah, we try to have a regular 3 month release cycle; see
> https://cwiki.apache.org/confluence/display/SPARK/Wiki+Homepage for the
> current window.
>
> Matei
>
> On Jul 16, 2014, at 4:21 PM, Mark Hamstra <m...@clearstorydata.com> wrote:
>
> You should expect master to compile and run: patches aren't merged unless
> they build and pass tests on Jenkins.
>
> You shouldn't expect new features to be added to stable code in
> maintenance releases (e.g. 1.0.1).
>
> AFAIK, we're still on track with Spark 1.1.0 development, which means that
> it should be released sometime in the second half of next month (or shortly
> thereafter).
>
>
> On Wed, Jul 16, 2014 at 4:03 PM, Paul Wais <pw...@yelp.com> wrote:
>
>> Dear List,
>>
>> The version of pyspark on master has a lot of nice new features, e.g.
>> SequenceFile reading, pickle i/o, etc:
>> https://github.com/apache/spark/blob/master/python/pyspark/context.py#L353
>>
>> I downloaded the recent 1.0.1 release and was surprised to see the
>> distribution did not include these changes in master.  (I've tried pulling
>> master [ 9c249743ea ] and compiling from source, but I get a build failure
>> in TestSQLContext.scala FWIW).
>>
>> Is an updated pyspark scheduled for the next release?  (Also, am I wrong
>> in expecting HEAD on master should probably compile and run?)
>>
>> Best Regards,
>> -Paul Wais
>>
>
>
>

Reply via email to