I will cut the RC today.  Sorry for the delay!

On Mon, Feb 22, 2016 at 5:19 AM, Patrick Woody <patrick.woo...@gmail.com>
wrote:

> Hey Michael,
>
> Any update on a first cut of the RC?
>
> Thanks!
> -Pat
>
> On Mon, Feb 15, 2016 at 6:50 PM, Michael Armbrust <mich...@databricks.com>
> wrote:
>
>> I'm not going to be able to do anything until after the Spark Summit, but
>> I will kick off RC1 after that (end of week).  Get your patches in before
>> then!
>>
>> On Sat, Feb 13, 2016 at 4:57 PM, Jong Wook Kim <ilike...@gmail.com>
>> wrote:
>>
>>> Is 1.6.1 going to be ready this week? I see that the two last unresolved
>>> issues targeting 1.6.1 are fixed
>>> <https://github.com/apache/spark/pull/11131> now
>>> <https://github.com/apache/spark/pull/10539>.
>>>
>>> On 3 February 2016 at 08:16, Daniel Darabos <
>>> daniel.dara...@lynxanalytics.com> wrote:
>>>
>>>>
>>>> On Tue, Feb 2, 2016 at 7:10 PM, Michael Armbrust <
>>>> mich...@databricks.com> wrote:
>>>>
>>>>> What about the memory leak bug?
>>>>>> https://issues.apache.org/jira/browse/SPARK-11293
>>>>>> Even after the memory rewrite in 1.6.0, it still happens in some
>>>>>> cases.
>>>>>> Will it be fixed for 1.6.1?
>>>>>>
>>>>>
>>>>> I think we have enough issues queued up that I would not hold the
>>>>> release for that, but if there is a patch we should try and review it.  We
>>>>> can always do 1.6.2 when more issues have been resolved.  Is this an 
>>>>> actual
>>>>> issue that is affecting a production workload or are we concerned about an
>>>>> edge case?
>>>>>
>>>>
>>>> The way we (Lynx Analytics) use RDDs, this affects almost everything we
>>>> do in production. Thankfully it does not cause any issues, it just logs a
>>>> lot of errors. I think the adverse effect may be that the memory manager
>>>> does not have a fully correct picture. But as long as the leak fits in the
>>>> "other" (unmanaged) memory fraction this will not cause issues. We don't
>>>> see this as an urgent issue. Thanks!
>>>>
>>>
>>>
>>
>

Reply via email to