Ahhh Thanks much! I miss my sparkConf.setJars function instead of this
hacky comma separated jar names.

On Thu, Apr 27, 2017 at 8:01 AM, Yanbo Liang <yblia...@gmail.com> wrote:

> Could you try the following way?
>
> val spark = 
> SparkSession.builder.appName("my-application").config("spark.jars", "a.jar, 
> b.jar").getOrCreate()
>
>
> Thanks
>
> Yanbo
>
>
> On Thu, Apr 27, 2017 at 9:21 AM, kant kodali <kanth...@gmail.com> wrote:
>
>> I am using Spark 2.1 BTW.
>>
>> On Wed, Apr 26, 2017 at 3:22 PM, kant kodali <kanth...@gmail.com> wrote:
>>
>>> Hi All,
>>>
>>> I am wondering how to create SparkSession using SparkConf object?
>>> Although I can see that most of the key value pairs we set in SparkConf we
>>> can also set in SparkSession or  SparkSession.Builder however I don't see
>>> sparkConf.setJars which is required right? Because we want the driver jar
>>> to be distributed across the cluster whether we run it in client mode or
>>> cluster mode. so I am wondering how is this possible?
>>>
>>> Thanks!
>>>
>>>
>>
>

Reply via email to