That was really helpful. Thanks! I actually solved my problem using by creating a venv and using the venv flags. Wondering now how to submit the data as an archive? Any idea?
On Mon, Jan 27, 2020, 9:25 PM Chris Teoh <chris.t...@gmail.com> wrote: > Use --py-files > > See > https://spark.apache.org/docs/latest/submitting-applications.html#bundling-your-applications-dependencies > > I hope that helps. > > On Tue, 28 Jan 2020, 9:46 am Tharindu Mathew, <tharindu.mat...@gmail.com> > wrote: > >> Hi, >> >> Newbie to pyspark/spark here. >> >> I'm trying to submit a job to pyspark with a dependency. Spark DL in this >> case. While the local environment has this the pyspark does not see it. How >> do I correctly start pyspark so that it sees this dependency? >> >> Using Spark 2.3.0 in a cloudera setup. >> >> -- >> Regards, >> Tharindu Mathew >> http://tharindumathew.com >> >