+1 to fail fast. Thanks for reporting this, Jungtaek!

On Mon, Oct 26, 2020 at 8:36 AM Jungtaek Lim <kabhwan.opensou...@gmail.com>
wrote:

> Yeah I'm in favor of fast-fail if things are not working out as end users
> intended. Spark should only fail back when it doesn't make any difference
> but only some sort of performance. (like whole stage codegen) This fail
> back brings behavioral differences, which should be considered as a bug.
>
> I'll file an issue and raise a PR sooner. Thanks for providing voices!
>
> On Sat, Oct 24, 2020 at 2:03 AM Ryan Blue <rb...@netflix.com> wrote:
>
>> I agree. If the user configures an invalid catalog, it should fail and
>> propagate the exception. Running with a catalog other than the one the user
>> requested is incorrect.
>>
>> On Fri, Oct 23, 2020 at 5:24 AM Russell Spitzer <
>> russell.spit...@gmail.com> wrote:
>>
>>> I was convinced that we should probably just fail, but if that is too
>>> much of a change, then logging the exception is also acceptable.
>>>
>>> On Thu, Oct 22, 2020, 10:32 PM Jungtaek Lim <
>>> kabhwan.opensou...@gmail.com> wrote:
>>>
>>>> Hi devs,
>>>>
>>>> I got another report regarding configuring v2 session catalog, when
>>>> Spark fails to instantiate the configured catalog. For now, it just simply
>>>> logs error message without exception information, and silently uses the
>>>> default session catalog.
>>>>
>>>>
>>>> https://github.com/apache/spark/blob/3819d39607392aa968595e3d97b84fedf83d08d9/sql/catalyst/src/main/scala/org/apache/spark/sql/connector/catalog/CatalogManager.scala#L75-L95
>>>>
>>>> IMO, as the user intentionally provides the session catalog, it
>>>> shouldn't fail back and just throw the exception. Otherwise (if we still
>>>> want to do the failback), we need to add the exception information in the
>>>> error log message at least.
>>>>
>>>> Would like to hear the voices.
>>>>
>>>> Thanks,
>>>> Jungtaek Lim (HeartSaVioR)
>>>>
>>>
>>
>> --
>> Ryan Blue
>> Software Engineer
>> Netflix
>>
>

Reply via email to