What is the instance type ?. Use an instance with atleast 4Gb+ RAM.  I
don't think it is possible to build on less than that. Other option would
be to use prebuilt binary.


On Wed, Nov 20, 2013 at 8:56 PM, Pankhuri Gupta <[email protected]> wrote:

> Hi,
> I am new to Spark and Scala. As a part of one of my projects, I am trying
> to build and locally publish spark-0.8.0-incubating on an Amazon ec2
> cluster.
>  After setting up all the java class paths and options, when I run :
> ** sbt/sbt compile , OR
>  **  sbt/sbt assembly, OR
> **  sbt/sbt publish-local
>  The command runs for some time (approx 10 mins) and after that the java
> command simply gets killed. No error message is thrown out. Here is a
> small snapshot of the messages:
>
> [info] Updating {file:/home/ec2-user/spark-0.8.0-incubating/}bagel...
> [info] Resolving cglib#cglib-nodep;2.2.2 …
> [info] Done updating.
>  [info] Compiling 258 Scala sources and 16 Java sources to
> /home/ec2-user/spark-0.8.0-incubating/core/target/scala-2.9.3/classes... 
> *sbt/sbt:
> line 30: 21454 Killed java* -Xmx1200m -XX:MaxPermSize=350m
> -XX:ReservedCodeCacheSize=256m $EXTRA_ARGS $SBT_OPTS -jar
> "$SPARK_HOME"/sbt/sbt-launch-*.jar "$@"
>
> When I run the *jstat*, I get the following output:
> Timestamp S0 S1 E O P YGC YGCT FGC FGCT GCT LGCC GCC
>   257.3 100.00 0.00 50.66 82.97 99.59 156 5.163 10 12.076 17.239 unknown
> GCCause No GC
>  365.2 0.00 100.00 53.13 88.96 99.94 157 7.563 10 12.076 19.639 unknown
> GCCause No GC
>   386.6 0.00 0.00 1.97 60.00 99.51 157 7.563 11 25.281 32.844 *Permanent
> Generation Full *No GC
>  407.9 0.00 0.00 29.53 60.00 99.97 157 7.563 11 25.281 32.844 *Permanent
> Generation Full* No GC
>   578.8 64.82 0.00 41.90 77.75 99.68 162 10.896 11 25.281 36.178 unknown
> GCCause No GC
>  600.1 64.82 0.00 91.90 77.75 99.72 162 10.896 11 25.281 36.178 unknown
> GCCause No GC
>  664.2 77.92 70.32 100.00 99.94 99.71 168 12.451 11 25.281 37.732 unknown
> GCCause *Allocation Failure*
>
> I changed the memory limits as :  -XX:MaxPermSize=720m and
> -XX:ReservedCodeCacheSize=512m, but still the problem persists.
>  I am not able to figure out the reason why the command is getting
> killed. Please let me know if I need to do some other checks. I read
> through many links on google and spark site as well but was not able to get
> any insight into this problem.
>
> I am using the following-
> Java version: 6
> Jvm :  1.6.0-openjdk.x86_64
>  Scala Version : 2.9.3 installed
>
> Any help would be deeply appreciated.
>
> Thanks,
>



-- 
s

Reply via email to