+1 (non-binding)

On Tue, Sep 23, 2025 at 8:04 PM Kent Yao <[email protected]> wrote:

> +1
>
> Rishab Joshi <[email protected]> 于2025年9月24日周三 10:57写道:
>
>> +1 (non-binding)
>>
>> On Tue, Sep 23, 2025 at 7:23 PM Yuming Wang <[email protected]> wrote:
>>
>>> +1
>>>
>>> On Wed, Sep 24, 2025 at 7:03 AM Denny Lee <[email protected]> wrote:
>>>
>>>> +1 (non-binding)
>>>>
>>>> Sent via Superhuman <https://sprh.mn/[email protected]>
>>>>
>>>>
>>>> On Tue, Sep 23, 2025 at 3:33 PM, Kousuke Saruta <[email protected]>
>>>> wrote:
>>>>
>>>>> +1
>>>>>
>>>>> 2025年9月24日(水) 7:19 Mich Talebzadeh <[email protected]>:
>>>>>
>>>>>> well. I see Nimrod has a valid point here. one can hide “Spark
>>>>>> Connect vs classic Spark SQL (HiveServer2/Beeline/Thrift) from users by
>>>>>> putting a thin abstraction in front.
>>>>>> Once that abstraction layer is there, then the hand over becomes
>>>>>> transparent. Your app talks to a Spark Connect endpoint, the Spark driver
>>>>>> runs in the cluster. Works with DataFrame API and SQL.
>>>>>> In essence you design a small library (or service) that exposes one
>>>>>> API to users and chooses the backend under the bonnet.
>>>>>>
>>>>>> HTH
>>>>>>
>>>>>> Dr Mich Talebzadeh,
>>>>>> Architect | Data Science | Financial Crime | Forensic Analysis | GDPR
>>>>>>
>>>>>>    view my Linkedin profile
>>>>>> <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Tue, 23 Sept 2025 at 21:40, Nimrod Ofek <[email protected]>
>>>>>> wrote:
>>>>>>
>>>>>>> Hi,
>>>>>>>
>>>>>>> That's the thing - I don't expect the users to know if they are
>>>>>>> connecting to Spark or Spark connect.
>>>>>>> That means I would expect the driver to support both Spark Connect -
>>>>>>> and the current Hive/ Beeline /Thrift server.
>>>>>>> Maybe even some regular Spark API (even a simple "jar" that just
>>>>>>> runs the query and returns the results somehow, haven't really thought
>>>>>>> about it in depth).
>>>>>>>
>>>>>>> The point I'm trying to make is that the user that wants to run a
>>>>>>> SQL command using JDBC - doesn't care if it's Spark or Spark connect or
>>>>>>> whatever - it's like a Database for the user...
>>>>>>>
>>>>>>> Regards,
>>>>>>> Nimrod
>>>>>>>
>>>>>>>
>>>>>>> On Mon, Sep 22, 2025 at 5:24 PM Cheng Pan <[email protected]> wrote:
>>>>>>>
>>>>>>>> Hi Nimrod,
>>>>>>>>
>>>>>>>> I'm not sure I get your question. Maybe the name should be 'JDBC
>>>>>>>> Driver for Spark Connect Server'?
>>>>>>>>
>>>>>>>> From the user's perspective, they simply use a JDBC driver to
>>>>>>>> connect
>>>>>>>> to the Connect Server to run SQL
>>>>>>>> and retrieve the results, without having to worry about whether
>>>>>>>> Spark
>>>>>>>> is running in classic or Connect mode.
>>>>>>>>
>>>>>>>> Thanks,
>>>>>>>> Cheng Pan
>>>>>>>>
>>>>>>>> On Mon, Sep 22, 2025 at 10:17 PM Nimrod Ofek <[email protected]>
>>>>>>>> wrote:
>>>>>>>> >
>>>>>>>> > I'll raise an issue with this- I don't think the user that uses
>>>>>>>> jdbc to Spark should know if he is working with Spark connect or 
>>>>>>>> regular
>>>>>>>> Spark....
>>>>>>>> > The jdbc driver should know how to work with connect with
>>>>>>>> fallback maybe, but the user doesn't care if he is getting Spark 
>>>>>>>> connect or
>>>>>>>> not...
>>>>>>>> >
>>>>>>>> > Regards,
>>>>>>>> > Nimrod
>>>>>>>> >
>>>>>>>> > בתאריך יום ב׳, 22 בספט׳ 2025, 16:04, מאת 杨杰 ‏<
>>>>>>>> [email protected]>:
>>>>>>>> >>
>>>>>>>> >> Hi Spark devs,
>>>>>>>> >>
>>>>>>>> >> I would like to start a vote on the SPIP: JDBC Driver for Spark
>>>>>>>> Connect
>>>>>>>> >>
>>>>>>>> >> Discussion thread:
>>>>>>>> >> https://lists.apache.org/thread/rx5pqh01c86slpqv9161hqwgm5lwxxzq
>>>>>>>> >> SPIP:
>>>>>>>> >>
>>>>>>>> https://docs.google.com/document/d/1Ahk4C16o1Jj1TbLg5ylzgHjvu2Ic2zTrcMuvLjqSoAQ/edit?tab=t.0#heading=h.1gf0bimgty0t
>>>>>>>> >> JIRA: https://issues.apache.org/jira/browse/SPARK-53484
>>>>>>>> >>
>>>>>>>> >> Please vote on the SPIP for the next 72 hours:
>>>>>>>> >>
>>>>>>>> >> [ ] +1: Accept the proposal as an official SPIP
>>>>>>>> >> [ ] +0
>>>>>>>> >> [ ] -1: I don’t think this is a good idea because
>>>>>>>
>>>>>>>
>>>>
>>
>> --
>> Regards
>> Rishab Joshi
>>
>

Reply via email to