Thank mich but I seem to remember to modify a config file so that I don't need to specify the --packages option every time I start the shell Kr On 21 Apr 2016 3:20 pm, "Mich Talebzadeh" <mich.talebza...@gmail.com> wrote:
> on spark-shell this will work > > $SPARK_HOME/bin/spark-shell *--packages * > com.databricks:spark-csv_2.11:1.3.0 > > HTH > > Dr Mich Talebzadeh > > > > LinkedIn * > https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw > <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* > > > > http://talebzadehmich.wordpress.com > > > > On 21 April 2016 at 15:13, Marco Mistroni <mmistr...@gmail.com> wrote: > >> HI all >> i need to use spark-csv in my spark instance, and i want to avoid >> launching spark-shell >> by passing the package name every time >> I seem to remember that i need to amend a file in the /conf directory to >> inlcude e,g >> spark.packages com.databricks:spark-csv_2.11:1.4.0 .... >> >> but i cannot find any docs tell ing me which config file i have to modify >> >> anyone can assist ? >> kr >> marco >> > >