Are you building / running with Java 6? I imagine your .jar files has
more than 65536 files, and Java 6 has various issues with jars this
large. If possible, use Java 7 everywhere.

https://issues.apache.org/jira/browse/SPARK-1520

On Sat, Jul 19, 2014 at 2:30 PM, boci <boci.b...@gmail.com> wrote:
> Hi Guys,
>
> I try to create spark uber jar with sbt but I have a lot of problem... I
> want to use the following:
> - Spark streaming
> - Kafka
> - Elsaticsearch
> - HBase
>
> the current jar size is cca 60M and it's not working.
> - When I deploy with spark-submit: It's running and exit without any error
> - When I try to start with local[*]  mode, it's say:
>  Exception in thread "main" java.lang.NoClassDefFoundError:
> org/apache/spark/Logging
> => but I start with java -cp /.../spark-assembly-1.0.1-hadoop2.2.0.jar -jar
> my.jar
>
> Any idea how can solve this? (which lib required to set provided wich
> required for run... later I want to run this jar in yarn cluster)
>
> b0c1
> ----------------------------------------------------------------------------------------------------------------------------------
> Skype: boci13, Hangout: boci.b...@gmail.com

Reply via email to