-1 (non-binding)

I also agree with using NEVER_INFER for 2.1.1. The migration cost is
unexpected for a point release.

On Mon, Apr 24, 2017 at 11:08 AM Holden Karau <hol...@pigscanfly.ca> wrote:

> Whoops, sorry finger slipped on that last message.
> It sounds like whatever we do is going to break some existing users
> (either with the tables by case sensitivity or with the unexpected scan).
>
> Personally I agree with Michael Allman on this, I believe we should
> use INFER_NEVER for 2.1.1.
>
> On Mon, Apr 24, 2017 at 11:01 AM, Holden Karau <hol...@pigscanfly.ca>
> wrote:
>
>> It
>>
>> On Mon, Apr 24, 2017 at 10:33 AM, Michael Allman <mich...@videoamp.com>
>> wrote:
>>
>>> The trouble we ran into is that this upgrade was blocking access to our
>>> tables, and we didn't know why. This sounds like a kind of migration
>>> operation, but it was not apparent that this was the case. It took an
>>> expert examining a stack trace and source code to figure this out. Would a
>>> more naive end user be able to debug this issue? Maybe we're an unusual
>>> case, but our particular experience was pretty bad. I have my doubts that
>>> the schema inference on our largest tables would ever complete without
>>> throwing some kind of timeout (which we were in fact receiving) or the end
>>> user just giving up and killing our job. We ended up doing a rollback while
>>> we investigated the source of the issue. In our case, INFER_NEVER is
>>> clearly the best configuration. We're going to add that to our default
>>> configuration files.
>>>
>>> My expectation is that a minor point release is a pretty safe bug fix
>>> release. We were a bit hasty in not doing better due diligence pre-upgrade.
>>>
>>> One suggestion the Spark team might consider is releasing 2.1.1 with
>>> INVER_NEVER and 2.2.0 with INFER_AND_SAVE. Clearly some kind of
>>> up-front migration notes would help in identifying this new behavior in 2.2.
>>>
>>> Thanks,
>>>
>>> Michael
>>>
>>>
>>> On Apr 24, 2017, at 2:09 AM, Wenchen Fan <wenc...@databricks.com> wrote:
>>>
>>> see https://issues.apache.org/jira/browse/SPARK-19611
>>>
>>> On Mon, Apr 24, 2017 at 2:22 PM, Holden Karau <hol...@pigscanfly.ca>
>>> wrote:
>>>
>>>> Whats the regression this fixed in 2.1 from 2.0?
>>>>
>>>> On Fri, Apr 21, 2017 at 7:45 PM, Wenchen Fan <wenc...@databricks.com>
>>>> wrote:
>>>>
>>>>> IIRC, the new "spark.sql.hive.caseSensitiveInferenceMode" stuff will
>>>>> only scan all table files only once, and write back the inferred schema to
>>>>> metastore so that we don't need to do the schema inference again.
>>>>>
>>>>> So technically this will introduce a performance regression for the
>>>>> first query, but compared to branch-2.0, it's not performance regression.
>>>>> And this patch fixed a regression in branch-2.1, which can run in
>>>>> branch-2.0. Personally, I think we should keep INFER_AND_SAVE as the
>>>>> default mode.
>>>>>
>>>>> + [Eric], what do you think?
>>>>>
>>>>> On Sat, Apr 22, 2017 at 1:37 AM, Michael Armbrust <
>>>>> mich...@databricks.com> wrote:
>>>>>
>>>>>> Thanks for pointing this out, Michael.  Based on the conversation on
>>>>>> the PR
>>>>>> <https://github.com/apache/spark/pull/16944#issuecomment-285529275>
>>>>>> this seems like a risky change to include in a release branch with a
>>>>>> default other than NEVER_INFER.
>>>>>>
>>>>>> +Wenchen?  What do you think?
>>>>>>
>>>>>> On Thu, Apr 20, 2017 at 4:14 PM, Michael Allman <mich...@videoamp.com
>>>>>> > wrote:
>>>>>>
>>>>>>> We've identified the cause of the change in behavior. It is related
>>>>>>> to the SQL conf key "spark.sql.hive.caseSensitiveInferenceMode". This 
>>>>>>> key
>>>>>>> and its related functionality was absent from our previous build. The
>>>>>>> default setting in the current build was causing Spark to attempt to 
>>>>>>> scan
>>>>>>> all table files during query analysis. Changing this setting to 
>>>>>>> NEVER_INFER
>>>>>>> disabled this operation and resolved the issue we had.
>>>>>>>
>>>>>>> Michael
>>>>>>>
>>>>>>>
>>>>>>> On Apr 20, 2017, at 3:42 PM, Michael Allman <mich...@videoamp.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>> I want to caution that in testing a build from this morning's
>>>>>>> branch-2.1 we found that Hive partition pruning was not working. We 
>>>>>>> found
>>>>>>> that Spark SQL was fetching all Hive table partitions for a very simple
>>>>>>> query whereas in a build from several weeks ago it was fetching only the
>>>>>>> required partitions. I cannot currently think of a reason for the
>>>>>>> regression outside of some difference between branch-2.1 from our 
>>>>>>> previous
>>>>>>> build and branch-2.1 from this morning.
>>>>>>>
>>>>>>> That's all I know right now. We are actively investigating to find
>>>>>>> the root cause of this problem, and specifically whether this is a 
>>>>>>> problem
>>>>>>> in the Spark codebase or not. I will report back when I have an answer 
>>>>>>> to
>>>>>>> that question.
>>>>>>>
>>>>>>> Michael
>>>>>>>
>>>>>>>
>>>>>>> On Apr 18, 2017, at 11:59 AM, Michael Armbrust <
>>>>>>> mich...@databricks.com> wrote:
>>>>>>>
>>>>>>> Please vote on releasing the following candidate as Apache Spark
>>>>>>> version 2.1.1. The vote is open until Fri, April 21st, 2018 at
>>>>>>> 13:00 PST and passes if a majority of at least 3 +1 PMC votes are
>>>>>>> cast.
>>>>>>>
>>>>>>> [ ] +1 Release this package as Apache Spark 2.1.1
>>>>>>> [ ] -1 Do not release this package because ...
>>>>>>>
>>>>>>>
>>>>>>> To learn more about Apache Spark, please see
>>>>>>> http://spark.apache.org/
>>>>>>>
>>>>>>> The tag to be voted on is v2.1.1-rc3
>>>>>>> <https://github.com/apache/spark/tree/v2.1.1-rc3> (
>>>>>>> 2ed19cff2f6ab79a718526e5d16633412d8c4dd4)
>>>>>>>
>>>>>>> List of JIRA tickets resolved can be found with this filter
>>>>>>> <https://issues.apache.org/jira/browse/SPARK-20134?jql=project%20%3D%20SPARK%20AND%20fixVersion%20%3D%202.1.1>
>>>>>>> .
>>>>>>>
>>>>>>> The release files, including signatures, digests, etc. can be found
>>>>>>> at:
>>>>>>> http://home.apache.org/~pwendell/spark-releases/spark-2.1.1-rc3-bin/
>>>>>>>
>>>>>>> Release artifacts are signed with the following key:
>>>>>>> https://people.apache.org/keys/committer/pwendell.asc
>>>>>>>
>>>>>>> The staging repository for this release can be found at:
>>>>>>>
>>>>>>> https://repository.apache.org/content/repositories/orgapachespark-1230/
>>>>>>>
>>>>>>> The documentation corresponding to this release can be found at:
>>>>>>>
>>>>>>> http://people.apache.org/~pwendell/spark-releases/spark-2.1.1-rc3-docs/
>>>>>>>
>>>>>>>
>>>>>>> *FAQ*
>>>>>>>
>>>>>>> *How can I help test this release?*
>>>>>>>
>>>>>>> If you are a Spark user, you can help us test this release by taking
>>>>>>> an existing Spark workload and running on this release candidate, then
>>>>>>> reporting any regressions.
>>>>>>>
>>>>>>> *What should happen to JIRA tickets still targeting 2.1.1?*
>>>>>>>
>>>>>>> Committers should look at those and triage. Extremely important bug
>>>>>>> fixes, documentation, and API tweaks that impact compatibility should be
>>>>>>> worked on immediately. Everything else please retarget to 2.1.2 or 
>>>>>>> 2.2.0.
>>>>>>>
>>>>>>> *But my bug isn't fixed!??!*
>>>>>>>
>>>>>>> In order to make timely releases, we will typically not hold the
>>>>>>> release unless the bug in question is a regression from 2.1.0.
>>>>>>>
>>>>>>> *What happened to RC1?*
>>>>>>>
>>>>>>> There were issues with the release packaging and as a result was
>>>>>>> skipped.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>>
>>>> --
>>>> Cell : 425-233-8271 <(425)%20233-8271>
>>>> Twitter: https://twitter.com/holdenkarau
>>>>
>>>
>>>
>>>
>>
>>
>> --
>> Cell : 425-233-8271 <(425)%20233-8271>
>> Twitter: https://twitter.com/holdenkarau
>>
>
>
>
> --
> Cell : 425-233-8271 <(425)%20233-8271>
> Twitter: https://twitter.com/holdenkarau
>

Reply via email to