on spark-shell this will work $SPARK_HOME/bin/spark-shell *--packages *com.databricks:spark-csv_2.11:1.3.0
HTH Dr Mich Talebzadeh LinkedIn * https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* http://talebzadehmich.wordpress.com On 21 April 2016 at 15:13, Marco Mistroni <mmistr...@gmail.com> wrote: > HI all > i need to use spark-csv in my spark instance, and i want to avoid > launching spark-shell > by passing the package name every time > I seem to remember that i need to amend a file in the /conf directory to > inlcude e,g > spark.packages com.databricks:spark-csv_2.11:1.4.0 .... > > but i cannot find any docs tell ing me which config file i have to modify > > anyone can assist ? > kr > marco >