[ 
https://issues.apache.org/jira/browse/SPARK-3761?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Igor Tkachenko updated SPARK-3761:
----------------------------------
    Description: 
I have Scala code:

val master = "spark://<server address>:7077"

    val sc = new SparkContext(new SparkConf()
      .setMaster(master)
      .setAppName("SparkQueryDemo 01")
      .set("spark.executor.memory", "512m"))

val count2 = sc         .textFile("hdfs://<server 
address>:8020/tmp/data/risk/account.txt")
      .filter(line  => line.contains("Word"))
      .count()

I've got such an error:
[error] (run-main-0) org.apache.spark.SparkException: Job aborted due to stage 
failure: Task 0.0:0 failed 4 times, most
recent failure: Exception failure in TID 6 on host <server address>: 
java.lang.ClassNotFoundExcept
ion: SimpleApp$$anonfun$1

My dependencies :

object Version {
  val spark        = "1.0.0-cdh5.1.0"
}

object Library {
  val sparkCore      = "org.apache.spark"  % "spark-assembly_2.10"  % 
Version.spark
}

My OS is Win 7

  was:
I have Scala code:

val master = "spark://<server address>:7077"

    val sc = new SparkContext(new SparkConf()
      .setMaster(master)
      .setAppName("SparkQueryDemo 01")
      .set("spark.executor.memory", "512m"))

val count2 = sc         .textFile("hdfs://<server 
address>:8020/tmp/data/risk/account.txt")
      .filter(line  => line.contains("Word"))
      .count()

I've got such an error:
[error] (run-main-0) org.apache.spark.SparkException: Job aborted due to stage 
failure: Task 0.0:0 failed 4 times, most
recent failure: Exception failure in TID 6 on host <server address>: 
java.lang.ClassNotFoundExcept
ion: SimpleApp$$anonfun$1

My dependencies :

object Version {
  val spark        = "1.0.0-cdh5.1.0"
  val hadoop       = "2.4.1"
  val slf4j        = "1.7.6"
  val logback      = "1.1.1"
  val scalaTest    = "2.1.0"
  val mockito      = "1.9.5"
}

object Library {
  val sparkCore      = "org.apache.spark"  %% "spark-assembly"  % Version.spark
  val hadoopClient   = "org.apache.hadoop" %  "hadoop-client"   % Version.hadoop
  val slf4jApi       = "org.slf4j"         %  "slf4j-api"       % Version.slf4j
  val logbackClassic = "ch.qos.logback"    %  "logback-classic" % 
Version.logback
  val scalaTest      = "org.scalatest"     %% "scalatest"       % 
Version.scalaTest
  val mockitoAll     = "org.mockito"       %  "mockito-all"     % 
Version.mockito
}

My OS is Win 7


> Class not found exception / sbt 13.5 / Scala 2.10.4
> ---------------------------------------------------
>
>                 Key: SPARK-3761
>                 URL: https://issues.apache.org/jira/browse/SPARK-3761
>             Project: Spark
>          Issue Type: Bug
>    Affects Versions: 1.0.0
>            Reporter: Igor Tkachenko
>
> I have Scala code:
> val master = "spark://<server address>:7077"
>     val sc = new SparkContext(new SparkConf()
>       .setMaster(master)
>       .setAppName("SparkQueryDemo 01")
>       .set("spark.executor.memory", "512m"))
> val count2 = sc         .textFile("hdfs://<server 
> address>:8020/tmp/data/risk/account.txt")
>       .filter(line  => line.contains("Word"))
>       .count()
> I've got such an error:
> [error] (run-main-0) org.apache.spark.SparkException: Job aborted due to 
> stage failure: Task 0.0:0 failed 4 times, most
> recent failure: Exception failure in TID 6 on host <server address>: 
> java.lang.ClassNotFoundExcept
> ion: SimpleApp$$anonfun$1
> My dependencies :
> object Version {
>   val spark        = "1.0.0-cdh5.1.0"
> }
> object Library {
>   val sparkCore      = "org.apache.spark"  % "spark-assembly_2.10"  % 
> Version.spark
> }
> My OS is Win 7



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to