on spark-shell this will work

$SPARK_HOME/bin/spark-shell *--packages *com.databricks:spark-csv_2.11:1.3.0

HTH

Dr Mich Talebzadeh



LinkedIn * 
https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
<https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*



http://talebzadehmich.wordpress.com



On 21 April 2016 at 15:13, Marco Mistroni <mmistr...@gmail.com> wrote:

> HI all
>  i need to use spark-csv in my spark instance, and i want to avoid
> launching spark-shell
> by passing the package name every time
> I seem to remember that i need to amend a file in the /conf directory to
> inlcude e,g
> spark.packages  com.databricks:spark-csv_2.11:1.4.0 ....
>
> but i cannot find any docs tell ing me which config file  i have to modify
>
> anyone can assist ?
> kr
>  marco
>

Reply via email to