Can you try decreasing the level of parallelism that you are giving for those functions? I had this issue when i gave a value > 500 and it was gone when i dropped it to 200.
Thanks Best Regards On Wed, Oct 8, 2014 at 9:28 AM, Andrew Ash <and...@andrewash.com> wrote: > Hi Meethu, > > I believe you may be hitting a regression in > https://issues.apache.org/jira/browse/SPARK-3633 > > If you are able, could you please try running a patched version of Spark > 1.1.0 that has commit 4fde28c reverted and see if the errors go away? > Posting your results on that bug would be useful, especially if the issues > disappear after the revert. > > Thanks! > Andrew > > On Tue, Oct 7, 2014 at 8:27 AM, MEETHU MATHEW <meethu2...@yahoo.co.in> > wrote: > >> Hi all, >> >> My code was working fine in spark 1.0.2 ,but after upgrading to 1.1.0, >> its throwing exceptions and tasks are getting failed. >> >> The code contains some map and filter transformations followed by >> groupByKey (reduceByKey in another code ). What I could find out is that >> the code works fine until groupByKey or reduceByKey in both >> versions.But after that the following errors show up in Spark 1.1.0 >> >> java.io.FileNotFoundException: >> /tmp/spark-local-20141006173014-4178/35/shuffle_6_0_5161 (Too many open >> files) >> java.io.FileOutputStream.openAppend(Native Method) >> java.io.FileOutputStream.<init>(FileOutputStream.java:210) >> >> org.apache.spark.storage.DiskBlockObjectWriter.open(BlockObjectWriter.scala:123) >> >> org.apache.spark.storage.DiskBlockObjectWriter.write(BlockObjectWriter.scala:192) >> >> org.apache.spark.shuffle.hash.HashShuffleWriter$$anonfun$write$1.apply(HashShuffleWriter.scala:67) >> >> org.apache.spark.shuffle.hash.HashShuffleWriter$$anonfun$write$1.apply(HashShuffleWriter.scala:65) >> scala.collection.Iterator$class.foreach(Iterator.scala:727) >> scala.collection.AbstractIterator.foreach(Iterator.scala:1157) >> >> org.apache.spark.shuffle.hash.HashShuffleWriter.write(HashShuffleWriter.scala:65) >> >> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68) >> >> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41) >> org.apache.spark.scheduler.Task.run(Task.scala:54) >> >> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:177) >> >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1146) >> >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >> java.lang.Thread.run(Thread.java:701) >> >> I cleaned my /tmp directory,changed my local directory to another folder >> ; but nothing helped. >> >> Can anyone say what could be the reason .? >> >> Thanks & Regards, >> Meethu M >> > >