Could you try the following way? val spark = SparkSession.builder.appName("my-application").config("spark.jars", "a.jar, b.jar").getOrCreate()
Thanks Yanbo On Thu, Apr 27, 2017 at 9:21 AM, kant kodali <kanth...@gmail.com> wrote: > I am using Spark 2.1 BTW. > > On Wed, Apr 26, 2017 at 3:22 PM, kant kodali <kanth...@gmail.com> wrote: > >> Hi All, >> >> I am wondering how to create SparkSession using SparkConf object? >> Although I can see that most of the key value pairs we set in SparkConf we >> can also set in SparkSession or SparkSession.Builder however I don't see >> sparkConf.setJars which is required right? Because we want the driver jar >> to be distributed across the cluster whether we run it in client mode or >> cluster mode. so I am wondering how is this possible? >> >> Thanks! >> >> >