It shouldn't? This one com.databricks.spark.csv.util.TextFile has hadoop imports.
I figured out that the answer to my question is just to add libraryDependencies += "org.apache.hadoop" % "hadoop-client" % "2.6.0". But i still wonder where is this 2.2.0 default comes from. From: Mohit Jaggi <mohitja...@gmail.com> To: Gil Vernik/Haifa/IBM@IBMIL Cc: Dev <dev@spark.apache.org> Date: 19/08/2015 21:47 Subject: Re: [spark-csv] how to build with Hadoop 2.6.0? spark-csv should not depend on hadoop On Sun, Aug 16, 2015 at 9:05 AM, Gil Vernik <g...@il.ibm.com> wrote: I would like to build spark-csv with Hadoop 2.6.0 I noticed that when i build it with sbt/sbt ++2.10.4 package it build it with Hadoop 2.2.0 ( at least this is what i saw in the .ivy2 repository). How to define 2.6.0 during spark-csv build? By the way, is it possible to build spark-csv using maven repository? Thanks, Gil.