df -i  # on a slave

Filesystem            Inodes   IUsed   IFree IUse% Mounted on
/dev/xvda1            524288  277701  246587   53% /
tmpfs                1917974       1 1917973    1% /dev/shm


On Tue, Jul 15, 2014 at 11:39 PM, Xiangrui Meng <men...@gmail.com> wrote:

> Check the number of inodes (df -i). The assembly build may create many
> small files. -Xiangrui
>
> On Tue, Jul 15, 2014 at 11:35 PM, Chris DuBois <chris.dub...@gmail.com>
> wrote:
> > Hi all,
> >
> > I am encountering the following error:
> >
> > INFO scheduler.TaskSetManager: Loss was due to java.io.IOException: No
> space
> > left on device [duplicate 4]
> >
> > For each slave, df -h looks roughtly like this, which makes the above
> error
> > surprising.
> >
> > Filesystem            Size  Used Avail Use% Mounted on
> > /dev/xvda1            7.9G  4.4G  3.5G  57% /
> > tmpfs                 7.4G  4.0K  7.4G   1% /dev/shm
> > /dev/xvdb              37G  3.3G   32G  10% /mnt
> > /dev/xvdf              37G  2.0G   34G   6% /mnt2
> > /dev/xvdv             500G   33M  500G   1% /vol
> >
> > I'm on an EC2 cluster (c3.xlarge + 5 x m3) that I launched using the
> > spark-ec2 scripts and a clone of spark from today. The job I am running
> > closely resembles the collaborative filtering example. This issue happens
> > with the 1M version as well as the 10 million rating version of the
> > MovieLens dataset.
> >
> > I have seen previous questions, but they haven't helped yet. For
> example, I
> > tried setting the Spark tmp directory to the EBS volume at /vol/, both by
> > editing the spark conf file (and copy-dir'ing it to the slaves) as well
> as
> > through the SparkConf. Yet I still get the above error. Here is my
> current
> > Spark config below. Note that I'm launching via ~/spark/bin/spark-submit.
> >
> > conf = SparkConf()
> > conf.setAppName("RecommendALS").set("spark.local.dir",
> > "/vol/").set("spark.executor.memory", "7g").set("spark.akka.frameSize",
> > "100").setExecutorEnv("SPARK_JAVA_OPTS", " -Dspark.akka.frameSize=100")
> > sc = SparkContext(conf=conf)
> >
> > Thanks for any advice,
> > Chris
> >
>

Reply via email to