Things I tried and the errors are :
String path =
"/home/ubuntu/spark-0.9.1/SimpleApp/target/simple-project-1.0-allinone.jar";..
.set(path)
$mvn package[ERROR] Failed to execute goal
org.apache.maven.plugins:maven-compiler-plugin:2.0.2:compile (default-compile)
on project simple-project: Compilation failure[ERROR]
/home/ubuntu/spark-0.9.1/SimpleApp/src/main/java/SimpleApp.java:[14,23] error:
method set in class SparkConf cannot be applied to given types;[ERROR] -> [Help
1]
String path =
"/home/ubuntu/spark-0.9.1/SimpleApp/target/simple-project-1.0-allinone.jar";..
.setJars(path)
$mvn package[ERROR] Failed to execute goal
org.apache.maven.plugins:maven-compiler-plugin:2.0.2:compile (default-compile)
on project simple-project: Compilation failure[ERROR]
/home/ubuntu/spark-0.9.1/SimpleApp/src/main/java/SimpleApp.java:[14,23] error:
no suitable method found for setJars(String)
content of my pom.xml file (maybe the problem arises here)<project> <build>
<plugins> <plugin> <groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-compiler-plugin</artifactId> <configuration>
<source>1.7</source> <target>1.7</target> </configuration> </plugin>
<plugin> <groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-shade-plugin</artifactId> <version>1.5</version>
<executions> <execution> <phase>package</phase> <goals>
<goal>shade</goal> </goals> <configuration>
<shadedArtifactAttached>true</shadedArtifactAttached>
<shadedClassifierName>allinone</shadedClassifierName> <artifactSet>
<includes> <include>*:*</include> </includes>
</artifactSet> <filters> <filter>
<artifact>*:*</artifact> <excludes>
<exclude>META-INF/*.SF</exclude>
<exclude>META-INF/*.DSA</exclude>
<exclude>META-INF/*.RSA</exclude> </excludes> </filter>
</filters> <transformers> <transformer
implementation="org.apache.maven.plugins.shade.resource.AppendingTransformer">
<resource>reference.conf</resource> </transformer>
<transformer
implementation="org.apache.maven.plugins.shade.resource.AppendingTransformer">
<resource>META-INF/spring.handlers</resource>
</transformer> <transformer
implementation="org.apache.maven.plugins.shade.resource.AppendingTransformer">
<resource>META-INF/spring.schemas</resource>
</transformer> <transformer
implementation="org.apache.maven.plugins.shade.resource.ManifestResourceTransformer">
<manifestEntries>
<Main-Class>com.echoed.chamber.Main</Main-Class>
</manifestEntries> </transformer> </transformers>
</configuration> </execution> </executions></plugin></plugins>
</build> <groupId>edu.berkeley</groupId>
<artifactId>simple-project</artifactId> <modelVersion>4.0.0</modelVersion>
<name>Simple Project</name> <packaging>jar</packaging> <version>1.0</version>
<repositories> <repository> <id>Akka repository</id>
<url>http://repo.akka.io/releases</url> </repository> </repositories>
<dependencies> <dependency> <!-- Spark dependency -->
<groupId>org.apache.spark</groupId>
<artifactId>spark-core_2.10</artifactId> <version>0.9.1</version>
</dependency> </dependencies></project>
Date: Thu, 1 May 2014 20:52:59 -0700
From: [email protected]
To: [email protected]
Subject: Re: java.lang.ClassNotFoundException
Hi, You should include the jar file of your project. for example:
conf.set("yourjarfilepath.jar")
Joe On Friday, May 2, 2014 7:39 AM, proofmoore [via Apache Spark User List]
<[hidden email]> wrote:
HelIo. I followed "A Standalone App in Java" part of the tutorial
https://spark.apache.org/docs/0.8.1/quick-start.html
Spark standalone cluster looks it's running without a problem :
http://i.stack.imgur.com/7bFv8.png
I have built a fat jar for running this JavaApp on the cluster. Before maven
package: find . ./pom.xml ./src ./src/main
./src/main/java ./src/main/java/SimpleApp.java
content of SimpleApp.java is :
import org.apache.spark.api.java.*; import
org.apache.spark.api.java.function.Function; import
org.apache.spark.SparkConf; import org.apache.spark.SparkContext;
public class SimpleApp { public static
void main(String[] args) {
SparkConf conf = new SparkConf()
.setMaster("spark://10.35.23.13:7077") .setAppName("My
app") .set("spark.executor.memory", "1g");
JavaSparkContext sc = new JavaSparkContext (conf); String logFile =
"/home/ubuntu/spark-0.9.1/test_data"; JavaRDD<String> logData =
sc.textFile(logFile).cache();
long numAs = logData.filter(new Function<String, Boolean>() { public
Boolean call(String s) { return s.contains("a");
} }).count();
System.out.println("Lines with a: " + numAs); } } This program
only works when master is set as setMaster("local"). Otherwise I get this error
: http://i.stack.imgur.com/doRSn.png
Thanks,Ibrahim
If you reply to this email, your message will be added to the
discussion below:
http://apache-spark-user-list.1001560.n3.nabble.com/java-lang-ClassNotFoundException-tp5191.html
To start a new topic under Apache Spark User List, email
[hidden email]
To unsubscribe from Apache Spark User List, click here.
NAML
View this message in context: Re: java.lang.ClassNotFoundException
Sent from the Apache Spark User List mailing list archive at Nabble.com.