Thanks for the help.
I will try deploying spark on a larger instance and then get back.
Best,
Pankhuri
On Nov 21, 2013, at 2:30 AM, Prashant Sharma <[email protected]> wrote:
> You mean t1.micro, The ram is less than a GB (615 MB) on those instances. It
> will not build. The size you are referring to is probably the storage size
> and not RAM. It might not be worth trying out spark on such instances.
> However if you plan to upgrade, chose atleast m1.large instances and then
> probably build on just one node and deploy to every other.
>
> HTH
>
>
> On Thu, Nov 21, 2013 at 12:49 PM, Pankhuri Gupta <[email protected]> wrote:
> The instance type is "ti.micro" with size as 7.9GB out of which 4.3GB is
> still available.
> For running spark (and later on hadoop), should i use a Storage Optimized
> instance or it can work on this as well?
>
> On Nov 20, 2013, at 11:39 AM, Prashant Sharma <[email protected]> wrote:
>
>> What is the instance type ?. Use an instance with atleast 4Gb+ RAM. I don't
>> think it is possible to build on less than that. Other option would be to
>> use prebuilt binary.
>>
>>
>> On Wed, Nov 20, 2013 at 8:56 PM, Pankhuri Gupta <[email protected]> wrote:
>> Hi,
>> I am new to Spark and Scala. As a part of one of my projects, I am
>> trying to build and locally publish spark-0.8.0-incubating on an Amazon ec2
>> cluster.
>> After setting up all the java class paths and options, when I run :
>> ** sbt/sbt compile , OR
>> ** sbt/sbt assembly, OR
>> ** sbt/sbt publish-local
>> The command runs for some time (approx 10 mins) and after that the java
>> command simply gets killed. No error message is thrown out. Here is a
>> small snapshot of the messages:
>>
>> [info] Updating {file:/home/ec2-user/spark-0.8.0-incubating/}bagel...
>> [info] Resolving cglib#cglib-nodep;2.2.2 …
>> [info] Done updating.
>> [info] Compiling 258 Scala sources and 16 Java sources to
>> /home/ec2-user/spark-0.8.0-incubating/core/target/scala-2.9.3/classes...
>> sbt/sbt: line 30: 21454 Killed java -Xmx1200m -XX:MaxPermSize=350m
>> -XX:ReservedCodeCacheSize=256m $EXTRA_ARGS $SBT_OPTS -jar
>> "$SPARK_HOME"/sbt/sbt-launch-*.jar "$@"
>>
>> When I run the jstat, I get the following output:
>> Timestamp S0 S1 E O P YGC YGCT FGC FGCT GCT LGCC GCC
>> 257.3 100.00 0.00 50.66 82.97 99.59 156 5.163 10 12.076 17.239 unknown
>> GCCause No GC
>> 365.2 0.00 100.00 53.13 88.96 99.94 157 7.563 10 12.076 19.639 unknown
>> GCCause No GC
>> 386.6 0.00 0.00 1.97 60.00 99.51 157 7.563 11 25.281 32.844 Permanent
>> Generation Full No GC
>> 407.9 0.00 0.00 29.53 60.00 99.97 157 7.563 11 25.281 32.844 Permanent
>> Generation Full No GC
>> 578.8 64.82 0.00 41.90 77.75 99.68 162 10.896 11 25.281 36.178 unknown
>> GCCause No GC
>> 600.1 64.82 0.00 91.90 77.75 99.72 162 10.896 11 25.281 36.178 unknown
>> GCCause No GC
>> 664.2 77.92 70.32 100.00 99.94 99.71 168 12.451 11 25.281 37.732
>> unknown GCCause Allocation Failure
>>
>> I changed the memory limits as : -XX:MaxPermSize=720m and
>> -XX:ReservedCodeCacheSize=512m, but still the problem persists.
>> I am not able to figure out the reason why the command is getting
>> killed. Please let me know if I need to do some other checks. I read through
>> many links on google and spark site as well but was not able to get any
>> insight into this problem.
>>
>> I am using the following-
>> Java version: 6
>> Jvm : 1.6.0-openjdk.x86_64
>> Scala Version : 2.9.3 installed
>>
>> Any help would be deeply appreciated.
>>
>> Thanks,
>>
>>
>>
>> --
>> s
>
>
>
>
> --
> s