And which version does your Spark cluster use?
> On 6. Apr 2017, at 16:11, nayan sharma <nayansharm...@gmail.com> wrote: > > scalaVersion := “2.10.5" > > > > >> On 06-Apr-2017, at 7:35 PM, Jörn Franke <jornfra...@gmail.com> wrote: >> >> Maybe your Spark is based on scala 2.11, but you compile it for 2.10 or the >> other way around? >> >>> On 6. Apr 2017, at 15:54, nayan sharma <nayansharm...@gmail.com> wrote: >>> >>> In addition I am using spark version 1.6.2 >>> Is there any chance of error coming because of Scala version or >>> dependencies are not matching.?I just guessed. >>> >>> Thanks, >>> Nayan >>> >>> >>>> On 06-Apr-2017, at 7:16 PM, nayan sharma <nayansharm...@gmail.com> wrote: >>>> >>>> Hi Jorn, >>>> Thanks for replying. >>>> >>>> jar -tf catalyst-data-prepration-assembly-1.0.jar | grep csv >>>> >>>> after doing this I have found a lot of classes under >>>> com/databricks/spark/csv/ >>>> >>>> do I need to check for any specific class ?? >>>> >>>> Regards, >>>> Nayan >>>>> On 06-Apr-2017, at 6:42 PM, Jörn Franke <jornfra...@gmail.com> wrote: >>>>> >>>>> Is the library in your assembly jar? >>>>> >>>>>> On 6. Apr 2017, at 15:06, nayan sharma <nayansharm...@gmail.com> wrote: >>>>>> >>>>>> Hi All, >>>>>> I am getting error while loading CSV file. >>>>>> >>>>>> val >>>>>> datacsv=sqlContext.read.format("com.databricks.spark.csv").option("header", >>>>>> "true").load("timeline.csv") >>>>>> java.lang.NoSuchMethodError: >>>>>> org.apache.commons.csv.CSVFormat.withQuote(Ljava/lang/Character;)Lorg/apache/commons/csv/CSVFormat; >>>>>> >>>>>> >>>>>> I have added the dependencies in sbt file >>>>>> // Spark Additional Library - CSV Read as DF >>>>>> libraryDependencies += "com.databricks" %% "spark-csv" % “1.5.0" >>>>>> and starting the spark-shell with command >>>>>> >>>>>> spark-shell --master yarn-client --jars >>>>>> /opt/packages/xxxx-data-prepration/target/scala-2.10/xxxx-data-prepration-assembly-1.0.jar >>>>>> --name nayan >>>>>> >>>>>> >>>>>> >>>>>> Thanks for any help!! >>>>>> >>>>>> >>>>>> Thanks, >>>>>> Nayan >>>> >>> >