I actually totally agree that we should make sure it should have no impact
on existing code if the feature is not used.


On Tue, Jul 31, 2018 at 1:18 PM Erik Erlandson <eerla...@redhat.com> wrote:

> I don't have a comprehensive knowledge of the project hydrogen PRs,
> however I've perused them, and they make substantial modifications to
> Spark's core DAG scheduler code.
>
> What I'm wondering is: how high is the confidence level that the
> "traditional" code paths are still stable. Put another way, is it even
> possible to "turn off" or "opt out" of this experimental feature? This
> analogy isn't perfect, but for example the k8s back-end is a major body of
> code, but it has a very small impact on any *core* code paths, and so if
> you opt out of it, it is well understood that you aren't running any
> experimental code.
>
> Looking at the project hydrogen code, I'm less sure the same is true.
> However, maybe there is a clear way to show how it is true.
>
>
> On Tue, Jul 31, 2018 at 12:03 PM, Mark Hamstra <m...@clearstorydata.com>
> wrote:
>
>> No reasonable amount of time is likely going to be sufficient to fully
>> vet the code as a PR. I'm not entirely happy with the design and code as
>> they currently are (and I'm still trying to find the time to more publicly
>> express my thoughts and concerns), but I'm fine with them going into 2.4
>> much as they are as long as they go in with proper stability annotations
>> and are understood not to be cast-in-stone final implementations, but
>> rather as a way to get people using them and generating the feedback that
>> is necessary to get us to something more like a final design and
>> implementation.
>>
>> On Tue, Jul 31, 2018 at 11:54 AM Erik Erlandson <eerla...@redhat.com>
>> wrote:
>>
>>>
>>> Barrier mode seems like a high impact feature on Spark's core code: is
>>> one additional week enough time to properly vet this feature?
>>>
>>> On Tue, Jul 31, 2018 at 7:10 AM, Joseph Torres <
>>> joseph.tor...@databricks.com> wrote:
>>>
>>>> Full continuous processing aggregation support ran into unanticipated
>>>> scalability and scheduling problems. We’re planning to overcome those by
>>>> using some of the barrier execution machinery, but since barrier execution
>>>> itself is still in progress the full support isn’t going to make it into
>>>> 2.4.
>>>>
>>>> Jose
>>>>
>>>> On Tue, Jul 31, 2018 at 6:07 AM Tomasz Gawęda <
>>>> tomasz.gaw...@outlook.com> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> what is the status of Continuous Processing + Aggregations? As far as
>>>>> I
>>>>> remember, Jose Torres said it should  be easy to perform aggregations
>>>>> if
>>>>> coalesce(1) work. IIRC it's already merged to master.
>>>>>
>>>>> Is this work in progress? If yes, it would be great to have full
>>>>> aggregation/join support in Spark 2.4 in CP.
>>>>>
>>>>> Pozdrawiam / Best regards,
>>>>>
>>>>> Tomek
>>>>>
>>>>>
>>>>> On 2018-07-31 10:43, Petar Zečević wrote:
>>>>> > This one is important to us:
>>>>> https://issues.apache.org/jira/browse/SPARK-24020 (Sort-merge join
>>>>> inner range optimization) but I think it could be useful to others too.
>>>>> >
>>>>> > It is finished and is ready to be merged (was ready a month ago at
>>>>> least).
>>>>> >
>>>>> > Do you think you could consider including it in 2.4?
>>>>> >
>>>>> > Petar
>>>>> >
>>>>> >
>>>>> > Wenchen Fan @ 1970-01-01 01:00 CET:
>>>>> >
>>>>> >> I went through the open JIRA tickets and here is a list that we
>>>>> should consider for Spark 2.4:
>>>>> >>
>>>>> >> High Priority:
>>>>> >> SPARK-24374: Support Barrier Execution Mode in Apache Spark
>>>>> >> This one is critical to the Spark ecosystem for deep learning. It
>>>>> only has a few remaining works and I think we should have it in Spark 2.4.
>>>>> >>
>>>>> >> Middle Priority:
>>>>> >> SPARK-23899: Built-in SQL Function Improvement
>>>>> >> We've already added a lot of built-in functions in this release,
>>>>> but there are a few useful higher-order functions in progress, like
>>>>> `array_except`, `transform`, etc. It would be great if we can get them in
>>>>> Spark 2.4.
>>>>> >>
>>>>> >> SPARK-14220: Build and test Spark against Scala 2.12
>>>>> >> Very close to finishing, great to have it in Spark 2.4.
>>>>> >>
>>>>> >> SPARK-4502: Spark SQL reads unnecessary nested fields from Parquet
>>>>> >> This one is there for years (thanks for your patience Michael!),
>>>>> and is also close to finishing. Great to have it in 2.4.
>>>>> >>
>>>>> >> SPARK-24882: data source v2 API improvement
>>>>> >> This is to improve the data source v2 API based on what we learned
>>>>> during this release. From the migration of existing sources and design of
>>>>> new features, we found some problems in the API and want to address them. 
>>>>> I
>>>>> believe this should be
>>>>> >> the last significant API change to data source v2, so great to have
>>>>> in Spark 2.4. I'll send a discuss email about it later.
>>>>> >>
>>>>> >> SPARK-24252: Add catalog support in Data Source V2
>>>>> >> This is a very important feature for data source v2, and is
>>>>> currently being discussed in the dev list.
>>>>> >>
>>>>> >> SPARK-24768: Have a built-in AVRO data source implementation
>>>>> >> Most of it is done, but date/timestamp support is still missing.
>>>>> Great to have in 2.4.
>>>>> >>
>>>>> >> SPARK-23243: Shuffle+Repartition on an RDD could lead to incorrect
>>>>> answers
>>>>> >> This is a long-standing correctness bug, great to have in 2.4.
>>>>> >>
>>>>> >> There are some other important features like the adaptive
>>>>> execution, streaming SQL, etc., not in the list, since I think we are not
>>>>> able to finish them before 2.4.
>>>>> >>
>>>>> >> Feel free to add more things if you think they are important to
>>>>> Spark 2.4 by replying to this email.
>>>>> >>
>>>>> >> Thanks,
>>>>> >> Wenchen
>>>>> >>
>>>>> >> On Mon, Jul 30, 2018 at 11:00 PM Sean Owen <sro...@apache.org>
>>>>> wrote:
>>>>> >>
>>>>> >>   In theory releases happen on a time-based cadence, so it's pretty
>>>>> much wrap up what's ready by the code freeze and ship it. In practice, the
>>>>> cadence slips frequently, and it's very much a negotiation about what
>>>>> features should push the
>>>>> >>   code freeze out a few weeks every time. So, kind of a hybrid
>>>>> approach here that works OK.
>>>>> >>
>>>>> >>   Certainly speak up if you think there's something that really
>>>>> needs to get into 2.4. This is that discuss thread.
>>>>> >>
>>>>> >>   (BTW I updated the page you mention just yesterday, to reflect
>>>>> the plan suggested in this thread.)
>>>>> >>
>>>>> >>   On Mon, Jul 30, 2018 at 9:51 AM Tom Graves
>>>>> <tgraves...@yahoo.com.invalid> wrote:
>>>>> >>
>>>>> >>   Shouldn't this be a discuss thread?
>>>>> >>
>>>>> >>   I'm also happy to see more release managers and agree the time is
>>>>> getting close, but we should see what features are in progress and see how
>>>>> close things are and propose a date based on that.  Cutting a branch to
>>>>> soon just creates
>>>>> >>   more work for committers to push to more branches.
>>>>> >>
>>>>> >>    http://spark.apache.org/versioning-policy.html mentioned the
>>>>> code freeze and release branch cut mid-august.
>>>>> >>
>>>>> >>   Tom
>>>>> >
>>>>> > ---------------------------------------------------------------------
>>>>> > To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>> >
>>>>>
>>>>>
>>>
>

Reply via email to